Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
18,897
2,616,011,931
IssuesEvent
2015-03-02 00:54:56
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
Create functionality that can disable starcraft rendering / GUI
auto-migrated Priority-Low Type-Enhancement
``` Create a function that can disable/enable the starcraft GUI and rendering code. This will be useful for speeding up starcraft games. ``` Original issue reported on code.google.com by `lowerlo...@gmail.com` on 16 Oct 2010 at 3:40
1.0
Create functionality that can disable starcraft rendering / GUI - ``` Create a function that can disable/enable the starcraft GUI and rendering code. This will be useful for speeding up starcraft games. ``` Original issue reported on code.google.com by `lowerlo...@gmail.com` on 16 Oct 2010 at 3:40
non_process
create functionality that can disable starcraft rendering gui create a function that can disable enable the starcraft gui and rendering code this will be useful for speeding up starcraft games original issue reported on code google com by lowerlo gmail com on oct at
0
74,460
20,167,167,746
IssuesEvent
2022-02-10 06:27:48
buildingSMART/IFC4.3.x-development
https://api.github.com/repos/buildingSMART/IFC4.3.x-development
opened
Is ParameterTakesPrecedence for IfcWindowType and IfcDoorType still a thing?
allocated-core proposal allocated-building
Originally brought up [here](https://forums.buildingsmart.org/t/using-parametertakesprecedence-in-ifcwindowtype-and-ifcdoortype/2574). I personally have never seen ParameterTakesPrecedence used. In the implementer agreement https://standards.buildingsmart.org/documents/Implementation/IFC_Implementation_Agreements/CV-2x3-181.html it basically states that it should always be false. I propose to mark it as deprecated.
1.0
Is ParameterTakesPrecedence for IfcWindowType and IfcDoorType still a thing? - Originally brought up [here](https://forums.buildingsmart.org/t/using-parametertakesprecedence-in-ifcwindowtype-and-ifcdoortype/2574). I personally have never seen ParameterTakesPrecedence used. In the implementer agreement https://standards.buildingsmart.org/documents/Implementation/IFC_Implementation_Agreements/CV-2x3-181.html it basically states that it should always be false. I propose to mark it as deprecated.
non_process
is parametertakesprecedence for ifcwindowtype and ifcdoortype still a thing originally brought up i personally have never seen parametertakesprecedence used in the implementer agreement it basically states that it should always be false i propose to mark it as deprecated
0
542,222
15,857,304,261
IssuesEvent
2021-04-08 04:28:10
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
[Bug] Notification service crash the app if it runs during sync
2.15 Bug Priority-High
###### Reproduction Steps 0. Turn on notification 1. Reinstall AnkiDroid 2. Start ankidroid 3. Sync ###### Expected Result Syncing. A notification. Not crashing ###### Actual Result Crash with the stacktrace: > 2021-03-18 04:53:05.640 22370-22370/com.ichi2.anki E/AndroidRuntime: FATAL EXCEPTION: main Process: com.ichi2.anki, PID: 22370 java.lang.RuntimeException: Unable to start receiver com.ichi2.anki.services.ReminderService: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference at android.app.ActivityThread.handleReceiver(ActivityThread.java:4207) at android.app.ActivityThread.access$1600(ActivityThread.java:273) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107) at android.os.Handler.dispatchMessage(Handler.java:107) at android.os.Looper.loop(Looper.java:237) at android.app.ActivityThread.main(ActivityThread.java:8107) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100) Caused by: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference at com.ichi2.libanki.sched.SchedV2._lrnForDeck(SchedV2.java:1478) at com.ichi2.libanki.sched.SchedV2.deckDueList(SchedV2.java:552) at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:592) at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:587) at com.ichi2.anki.services.ReminderService.getDeckOptionDue(ReminderService.java:172) at com.ichi2.anki.services.ReminderService.onReceive(ReminderService.java:110) at android.app.ActivityThread.handleReceiver(ActivityThread.java:4198) at android.app.ActivityThread.access$1600(ActivityThread.java:273)Β  at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)Β  at android.os.Handler.dispatchMessage(Handler.java:107)Β  at android.os.Looper.loop(Looper.java:237)Β  at android.app.ActivityThread.main(ActivityThread.java:8107)Β  at java.lang.reflect.Method.invoke(Native Method)Β  at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)Β  at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)Β  My understanding is that the notification service started running. However, sync closed the database, which means that notification suddenly can't access the database anymore One way to deal with it would be to try/catch the NullPointerException here and state that notification is not so important that it should crash the program. Or that sync interact with notification, to pause notification until database is reoladed
1.0
[Bug] Notification service crash the app if it runs during sync - ###### Reproduction Steps 0. Turn on notification 1. Reinstall AnkiDroid 2. Start ankidroid 3. Sync ###### Expected Result Syncing. A notification. Not crashing ###### Actual Result Crash with the stacktrace: > 2021-03-18 04:53:05.640 22370-22370/com.ichi2.anki E/AndroidRuntime: FATAL EXCEPTION: main Process: com.ichi2.anki, PID: 22370 java.lang.RuntimeException: Unable to start receiver com.ichi2.anki.services.ReminderService: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference at android.app.ActivityThread.handleReceiver(ActivityThread.java:4207) at android.app.ActivityThread.access$1600(ActivityThread.java:273) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107) at android.os.Handler.dispatchMessage(Handler.java:107) at android.os.Looper.loop(Looper.java:237) at android.app.ActivityThread.main(ActivityThread.java:8107) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100) Caused by: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference at com.ichi2.libanki.sched.SchedV2._lrnForDeck(SchedV2.java:1478) at com.ichi2.libanki.sched.SchedV2.deckDueList(SchedV2.java:552) at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:592) at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:587) at com.ichi2.anki.services.ReminderService.getDeckOptionDue(ReminderService.java:172) at com.ichi2.anki.services.ReminderService.onReceive(ReminderService.java:110) at android.app.ActivityThread.handleReceiver(ActivityThread.java:4198) at android.app.ActivityThread.access$1600(ActivityThread.java:273)Β  at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)Β  at android.os.Handler.dispatchMessage(Handler.java:107)Β  at android.os.Looper.loop(Looper.java:237)Β  at android.app.ActivityThread.main(ActivityThread.java:8107)Β  at java.lang.reflect.Method.invoke(Native Method)Β  at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)Β  at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)Β  My understanding is that the notification service started running. However, sync closed the database, which means that notification suddenly can't access the database anymore One way to deal with it would be to try/catch the NullPointerException here and state that notification is not so important that it should crash the program. Or that sync interact with notification, to pause notification until database is reoladed
non_process
notification service crash the app if it runs during sync reproduction steps turn on notification reinstall ankidroid start ankidroid sync expected result syncing a notification not crashing actual result crash with the stacktrace com anki e androidruntime fatal exception main process com anki pid java lang runtimeexception unable to start receiver com anki services reminderservice java lang nullpointerexception attempt to invoke virtual method int com libanki db queryscalar java lang string java lang object on a null object reference at android app activitythread handlereceiver activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java caused by java lang nullpointerexception attempt to invoke virtual method int com libanki db queryscalar java lang string java lang object on a null object reference at com libanki sched lrnfordeck java at com libanki sched deckduelist java at com libanki sched deckduetree java at com libanki sched deckduetree java at com anki services reminderservice getdeckoptiondue reminderservice java at com anki services reminderservice onreceive reminderservice java at android app activitythread handlereceiver activitythread java at android app activitythread access activitythread java Β  at android app activitythread h handlemessage activitythread java Β  at android os handler dispatchmessage handler java Β  at android os looper loop looper java Β  at android app activitythread main activitythread java Β  at java lang reflect method invoke native method Β  at com android internal os runtimeinit methodandargscaller run runtimeinit java Β  at com android internal os zygoteinit main zygoteinit java Β  my understanding is that the notification service started running however sync closed the database which means that notification suddenly can t access the database anymore one way to deal with it would be to try catch the nullpointerexception here and state that notification is not so important that it should crash the program or that sync interact with notification to pause notification until database is reoladed
0
145,039
19,319,018,374
IssuesEvent
2021-12-14 01:50:26
jozseftiborcz/sast-eval-springboot1
https://api.github.com/repos/jozseftiborcz/sast-eval-springboot1
opened
CVE-2021-22096 (Medium) detected in spring-webmvc-5.2.2.RELEASE.jar, spring-web-5.2.2.RELEASE.jar
security vulnerability
## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-webmvc-5.2.2.RELEASE.jar</b>, <b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-webmvc-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: sast-eval-springboot1/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library) - :x: **spring-webmvc-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: sast-eval-springboot1/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library) - :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22096 (Medium) detected in spring-webmvc-5.2.2.RELEASE.jar, spring-web-5.2.2.RELEASE.jar - ## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-webmvc-5.2.2.RELEASE.jar</b>, <b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-webmvc-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: sast-eval-springboot1/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library) - :x: **spring-webmvc-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: sast-eval-springboot1/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library) - :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in spring webmvc release jar spring web release jar cve medium severity vulnerability vulnerable libraries spring webmvc release jar spring web release jar spring webmvc release jar spring web mvc library home page a href path to dependency file sast eval build gradle path to vulnerable library home wss scanner repository org springframework spring webmvc release spring webmvc release jar home wss scanner repository org springframework spring webmvc release spring webmvc release jar dependency hierarchy spring boot starter web release jar root library x spring webmvc release jar vulnerable library spring web release jar spring web library home page a href path to dependency file sast eval pom xml path to vulnerable library home wss scanner repository org springframework spring web release spring web release jar home wss scanner repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in base branch master vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release org springframework spring web release org springframework spring webmvc release org springframework spring webflux release step up your open source security game with whitesource
0
143,971
5,533,498,813
IssuesEvent
2017-03-21 13:30:52
djoproject/pyshell
https://api.github.com/repos/djoproject/pyshell
closed
EVERYWHERE/make code compliant with python3
enhancement Middle priority To reCheck
### Description Just execute pytest with a python3 interpreter and fix the issues until the test pass, then try to execute the whole software. ### Stuff to check/improve - encoding #99 - decorator (pyshell.arg.decorator) - ...
1.0
EVERYWHERE/make code compliant with python3 - ### Description Just execute pytest with a python3 interpreter and fix the issues until the test pass, then try to execute the whole software. ### Stuff to check/improve - encoding #99 - decorator (pyshell.arg.decorator) - ...
non_process
everywhere make code compliant with description just execute pytest with a interpreter and fix the issues until the test pass then try to execute the whole software stuff to check improve encoding decorator pyshell arg decorator
0
53,615
13,185,437,429
IssuesEvent
2020-08-12 21:24:26
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
FNX2-15613 ⁃ Rename mozilla_ui_publicsuffixlist to mozilla_lib_publicsuffixlist
eng:build good first issue help wanted
`Deps.mozilla_ui_publicsuffixlist` should be renamed to `Deps.mozilla_lib`_publicsuffixlist and updated in the build.gradle file. The corresponding dependency is named "lib-publicsuffixlist", so the current name is incorrect.
1.0
FNX2-15613 ⁃ Rename mozilla_ui_publicsuffixlist to mozilla_lib_publicsuffixlist - `Deps.mozilla_ui_publicsuffixlist` should be renamed to `Deps.mozilla_lib`_publicsuffixlist and updated in the build.gradle file. The corresponding dependency is named "lib-publicsuffixlist", so the current name is incorrect.
non_process
⁃ rename mozilla ui publicsuffixlist to mozilla lib publicsuffixlist deps mozilla ui publicsuffixlist should be renamed to deps mozilla lib publicsuffixlist and updated in the build gradle file the corresponding dependency is named lib publicsuffixlist so the current name is incorrect
0
73,018
7,320,547,729
IssuesEvent
2018-03-02 07:53:03
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
closed
FeedbackSessionQuestionTypeTest.json: fix malformed question metadata
a-Testing d.Contributors e.1 p.Low
Discovered while looking into https://github.com/TEAMMATES/teammates/issues/8415#issuecomment-364003271. The test data of one question is malformed (it's an MSQ question with `msqChoices` but `numOfMcqChoices` instead of `numOfMsqChoices` in the metadata): https://github.com/TEAMMATES/teammates/blob/85faa7829f4edf19383b033e2d59825652b4b835/src/test/resources/data/FeedbackSessionQuestionTypeTest.json#L711 To investigate whether there are other similar cases, and why the deserialisation does not fail.
1.0
FeedbackSessionQuestionTypeTest.json: fix malformed question metadata - Discovered while looking into https://github.com/TEAMMATES/teammates/issues/8415#issuecomment-364003271. The test data of one question is malformed (it's an MSQ question with `msqChoices` but `numOfMcqChoices` instead of `numOfMsqChoices` in the metadata): https://github.com/TEAMMATES/teammates/blob/85faa7829f4edf19383b033e2d59825652b4b835/src/test/resources/data/FeedbackSessionQuestionTypeTest.json#L711 To investigate whether there are other similar cases, and why the deserialisation does not fail.
non_process
feedbacksessionquestiontypetest json fix malformed question metadata discovered while looking into the test data of one question is malformed it s an msq question with msqchoices but numofmcqchoices instead of numofmsqchoices in the metadata to investigate whether there are other similar cases and why the deserialisation does not fail
0
17,679
23,512,656,480
IssuesEvent
2022-08-18 18:08:23
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
Port code samples of GAE Python bundled APIs from GCS
type: process api: appengine samples
## In which file did you encounter the issue? N/A. I'm about to add a set of code samples. ### Did you change the file? If so, how? Yes. `gsutil cp gs://python3-wormhole-api-samples/python3-wormhole-samples.zip .` ## Describe the issue GAE Python bundled APIs went GA in April. The code samples for these APIs are still in a GCS bucket. https://storage.googleapis.com/python3-wormhole-api-samples/python3-wormhole-samples.zip This issue is to publish these samples to [python-docs-samples](https://github.com/GoogleCloudPlatform/python-docs-samples)
1.0
Port code samples of GAE Python bundled APIs from GCS - ## In which file did you encounter the issue? N/A. I'm about to add a set of code samples. ### Did you change the file? If so, how? Yes. `gsutil cp gs://python3-wormhole-api-samples/python3-wormhole-samples.zip .` ## Describe the issue GAE Python bundled APIs went GA in April. The code samples for these APIs are still in a GCS bucket. https://storage.googleapis.com/python3-wormhole-api-samples/python3-wormhole-samples.zip This issue is to publish these samples to [python-docs-samples](https://github.com/GoogleCloudPlatform/python-docs-samples)
process
port code samples of gae python bundled apis from gcs in which file did you encounter the issue n a i m about to add a set of code samples did you change the file if so how yes gsutil cp gs wormhole api samples wormhole samples zip describe the issue gae python bundled apis went ga in april the code samples for these apis are still in a gcs bucket this issue is to publish these samples to
1
319,271
23,763,092,604
IssuesEvent
2022-09-01 10:30:30
celestiaorg/docs
https://api.github.com/repos/celestiaorg/docs
closed
Document Celestia's Reserved Namespaces
documentation
It's important for developers to know what Celestia's Reserved Namespaces are
1.0
Document Celestia's Reserved Namespaces - It's important for developers to know what Celestia's Reserved Namespaces are
non_process
document celestia s reserved namespaces it s important for developers to know what celestia s reserved namespaces are
0
1,316
3,867,055,676
IssuesEvent
2016-04-09 01:37:57
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
closed
Properly initialize servo values
Necessary Addition Processing
The servos all start at 0 but they should start at the "center" value for the camera and open for the claw (make these constants because we will have to change them when we get the mechanical specs from @ddworken or @vishnuramasawmy) Also, create variables (probably arrays) for the range of values that the servos operate on and replace those in the code.
1.0
Properly initialize servo values - The servos all start at 0 but they should start at the "center" value for the camera and open for the claw (make these constants because we will have to change them when we get the mechanical specs from @ddworken or @vishnuramasawmy) Also, create variables (probably arrays) for the range of values that the servos operate on and replace those in the code.
process
properly initialize servo values the servos all start at but they should start at the center value for the camera and open for the claw make these constants because we will have to change them when we get the mechanical specs from ddworken or vishnuramasawmy also create variables probably arrays for the range of values that the servos operate on and replace those in the code
1
278,192
30,702,228,143
IssuesEvent
2023-07-27 01:13:08
nidhi7598/linux-4.1.15_CVE-2019-10220
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220
closed
CVE-2016-4485 (High) detected in linuxlinux-4.4.302 - autoclosed
Mend: dependency security vulnerability
## CVE-2016-4485 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The llc_cmsg_rcv function in net/llc/af_llc.c in the Linux kernel before 4.5.5 does not initialize a certain data structure, which allows attackers to obtain sensitive information from kernel stack memory by reading a message. <p>Publish Date: 2016-05-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4485>CVE-2016-4485</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4485">https://nvd.nist.gov/vuln/detail/CVE-2016-4485</a></p> <p>Release Date: 2016-05-23</p> <p>Fix Resolution: 4.5.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-4485 (High) detected in linuxlinux-4.4.302 - autoclosed - ## CVE-2016-4485 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The llc_cmsg_rcv function in net/llc/af_llc.c in the Linux kernel before 4.5.5 does not initialize a certain data structure, which allows attackers to obtain sensitive information from kernel stack memory by reading a message. <p>Publish Date: 2016-05-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4485>CVE-2016-4485</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4485">https://nvd.nist.gov/vuln/detail/CVE-2016-4485</a></p> <p>Release Date: 2016-05-23</p> <p>Fix Resolution: 4.5.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net llc af llc c net llc af llc c vulnerability details the llc cmsg rcv function in net llc af llc c in the linux kernel before does not initialize a certain data structure which allows attackers to obtain sensitive information from kernel stack memory by reading a message publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
15,952
20,171,691,257
IssuesEvent
2022-02-10 10:59:05
ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed
https://api.github.com/repos/ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed
opened
πŸ›‘ Processing failed: KeyError
process
## Overview `KeyError` found in `processing_task` task during run ended on 2022-02-10T10:59:05.295909. ## Details Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed` Task name: `processing_task` Error type: `KeyError` Error message: 'vadcp_eastward_seawater_velocity_dim_0' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 346, in append_to_zarr dim_indexer, modify_zarr_dims = _validate_dims( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 76, in _validate_dims existing_var = existing_zarr[dim] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 349, in __getitem__ raise KeyError(item) KeyError: 'vadcp_eastward_seawater_velocity_dim_0' ``` </details>
1.0
πŸ›‘ Processing failed: KeyError - ## Overview `KeyError` found in `processing_task` task during run ended on 2022-02-10T10:59:05.295909. ## Details Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed` Task name: `processing_task` Error type: `KeyError` Error message: 'vadcp_eastward_seawater_velocity_dim_0' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 346, in append_to_zarr dim_indexer, modify_zarr_dims = _validate_dims( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 76, in _validate_dims existing_var = existing_zarr[dim] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 349, in __getitem__ raise KeyError(item) KeyError: 'vadcp_eastward_seawater_velocity_dim_0' ``` </details>
process
πŸ›‘ processing failed keyerror overview keyerror found in processing task task during run ended on details flow name streamed vadcp beam parsed task name processing task error type keyerror error message vadcp eastward seawater velocity dim traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing process dataset file srv conda envs notebook lib site packages ooi harvester processor init py line in process dataset append to zarr mod ds store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr dim indexer modify zarr dims validate dims file srv conda envs notebook lib site packages ooi harvester processor utils py line in validate dims existing var existing zarr file srv conda envs notebook lib site packages zarr hierarchy py line in getitem raise keyerror item keyerror vadcp eastward seawater velocity dim
1
695,035
23,841,273,886
IssuesEvent
2022-09-06 10:25:43
RinMinase/anidb
https://api.github.com/repos/RinMinase/anidb
closed
Add checkbox for filtering download list
Type: Feature Request Priority: Normal
- To not include or exclude watched - By default exclude watched titles
1.0
Add checkbox for filtering download list - - To not include or exclude watched - By default exclude watched titles
non_process
add checkbox for filtering download list to not include or exclude watched by default exclude watched titles
0
14,402
17,456,177,591
IssuesEvent
2021-08-06 01:47:13
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Rasterize mesh dataset tool: python error when running it in batch mode
Feedback stale Processing Bug Mesh
When running the said tool in batch mode (Right click>Execute as Batch Process) with several hundred processes (100s of lines), the tool crashes intermittently after successfully generating several, but not all of the output rasters. The batch processes are populated by a JSON file via the Open (Batch) button in the tool dialog. I am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully. Here's a sample of the JSON batch file: `[{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5038000, 5039000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5038.tif"}}, {"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5039000, 5040000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5039.tif"}}, {"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5040000, 5041000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5040.tif"}}]` I am also able to run the same entire batch process successfully with the old Crayfish plugin version of the tool on QGIS 3.10 without errors/crashes. Error message: `An error has occurred while executing Python code: Exception: unknown Traceback (most recent call last): File "C:/QGIS_L~1/apps/qgis/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 157, in runAlgorithm results, ok = self.algorithm().run(parameters, context, multi_feedback) Exception: unknown Python version: 3.7.0 (v3.7.0:1bf9cc5093, Jun 27 2018, 04:59:51) [MSC v.1914 64 bit (AMD64)] QGIS version: 3.18.0-ZΓΌrich ZΓΌrich, bdef9fb328 Python Path: C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins\LAStools C:/QGIS_L~1/apps/qgis/./python C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins C:/QGIS_L~1/apps/qgis/./python/plugins C:\QGIS_L~1\apps\Python37 C:\QGIS_L~1\apps\Python37\Scripts C:\QGIS_latest\bin\python37.zip C:\QGIS_L~1\apps\Python37\DLLs C:\QGIS_L~1\apps\Python37\lib C:\QGIS_latest\bin C:\QGIS_L~1\apps\Python37\lib\site-packages C:\QGIS_L~1\apps\Python37\lib\site-packages\win32 C:\QGIS_L~1\apps\Python37\lib\site-packages\win32\lib C:\QGIS_L~1\apps\Python37\lib\site-packages\Pythonwin C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\crayfish\gui\..\pyqtgraph-0.10.0-py2.py3-none-any.whl C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\QuickMultiAttributeEdit3/forms` QGIS version info: ` QGIS version 3.18.0-ZΓΌrich QGIS code revision bdef9fb328 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins changeDataSource; crayfish; ImportPhotos; LAStools; latlontools; LecoS; Mergin; profiletool; qgis-maptiler-plugin; QPackage; QuickMultiAttributeEdit3; quick_map_services; splitmultipart; valuetool; db_manager; processing `
1.0
Rasterize mesh dataset tool: python error when running it in batch mode - When running the said tool in batch mode (Right click>Execute as Batch Process) with several hundred processes (100s of lines), the tool crashes intermittently after successfully generating several, but not all of the output rasters. The batch processes are populated by a JSON file via the Open (Batch) button in the tool dialog. I am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully. Here's a sample of the JSON batch file: `[{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5038000, 5039000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5038.tif"}}, {"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5039000, 5040000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5039.tif"}}, {"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5040000, 5041000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5040.tif"}}]` I am also able to run the same entire batch process successfully with the old Crayfish plugin version of the tool on QGIS 3.10 without errors/crashes. Error message: `An error has occurred while executing Python code: Exception: unknown Traceback (most recent call last): File "C:/QGIS_L~1/apps/qgis/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 157, in runAlgorithm results, ok = self.algorithm().run(parameters, context, multi_feedback) Exception: unknown Python version: 3.7.0 (v3.7.0:1bf9cc5093, Jun 27 2018, 04:59:51) [MSC v.1914 64 bit (AMD64)] QGIS version: 3.18.0-ZΓΌrich ZΓΌrich, bdef9fb328 Python Path: C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins\LAStools C:/QGIS_L~1/apps/qgis/./python C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins C:/QGIS_L~1/apps/qgis/./python/plugins C:\QGIS_L~1\apps\Python37 C:\QGIS_L~1\apps\Python37\Scripts C:\QGIS_latest\bin\python37.zip C:\QGIS_L~1\apps\Python37\DLLs C:\QGIS_L~1\apps\Python37\lib C:\QGIS_latest\bin C:\QGIS_L~1\apps\Python37\lib\site-packages C:\QGIS_L~1\apps\Python37\lib\site-packages\win32 C:\QGIS_L~1\apps\Python37\lib\site-packages\win32\lib C:\QGIS_L~1\apps\Python37\lib\site-packages\Pythonwin C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\crayfish\gui\..\pyqtgraph-0.10.0-py2.py3-none-any.whl C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\QuickMultiAttributeEdit3/forms` QGIS version info: ` QGIS version 3.18.0-ZΓΌrich QGIS code revision bdef9fb328 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins changeDataSource; crayfish; ImportPhotos; LAStools; latlontools; LecoS; Mergin; profiletool; qgis-maptiler-plugin; QPackage; QuickMultiAttributeEdit3; quick_map_services; splitmultipart; valuetool; db_manager; processing `
process
rasterize mesh dataset tool python error when running it in batch mode when running the said tool in batch mode right click execute as batch process with several hundred processes of lines the tool crashes intermittently after successfully generating several but not all of the output rasters the batch processes are populated by a json file via the open batch button in the tool dialog i am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully here s a sample of the json batch file dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif parameters input path to my mesh dataset groups dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif parameters input path to my mesh dataset groups dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif i am also able to run the same entire batch process successfully with the old crayfish plugin version of the tool on qgis without errors crashes error message an error has occurred while executing python code exception unknown traceback most recent call last file c qgis l apps qgis python plugins processing gui batchalgorithmdialog py line in runalgorithm results ok self algorithm run parameters context multi feedback exception unknown python version jun qgis version zΓΌrich zΓΌrich python path c users mywinusername appdata roaming qgis profiles default python plugins lastools c qgis l apps qgis python c users mywinusername appdata roaming qgis profiles default python c users mywinusername appdata roaming qgis profiles default python plugins c qgis l apps qgis python plugins c qgis l apps c qgis l apps scripts c qgis latest bin zip c qgis l apps dlls c qgis l apps lib c qgis latest bin c qgis l apps lib site packages c qgis l apps lib site packages c qgis l apps lib site packages lib c qgis l apps lib site packages pythonwin c users mywinusername appdata roaming qgis profiles default python c users mywinusername appdata roaming qgis profiles default python plugins crayfish gui pyqtgraph none any whl c users mywinusername appdata roaming qgis profiles default python plugins forms qgis version info qgis version zΓΌrich qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins changedatasource crayfish importphotos lastools latlontools lecos mergin profiletool qgis maptiler plugin qpackage quick map services splitmultipart valuetool db manager processing
1
20,656
27,329,693,305
IssuesEvent
2023-02-25 13:13:25
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
closed
Connect to UB Servers and create a static "Hello World" website.
Processing Task Sprint 1
Test 1. User will go to this website: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/ 2. Click on the hello.php file 3. Confirm the user can see the text "Hello World".
1.0
Connect to UB Servers and create a static "Hello World" website. - Test 1. User will go to this website: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/ 2. Click on the hello.php file 3. Confirm the user can see the text "Hello World".
process
connect to ub servers and create a static hello world website test user will go to this website click on the hello php file confirm the user can see the text hello world
1
45,821
24,228,833,655
IssuesEvent
2022-09-26 16:24:56
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
closed
Consider prefetching API data on page load
type: enhancement category: performance category: engineering
At the cost of skipping the loading placeholder view, we could **speed up loading time on initial page load** by pre<s>fetch</s>loading the data that will be needed by the view, in order to avoid round-trips on what I'd expect to be the most latency-prone leg of the stack, browser to site. Currently, descriptors of what data is needed to populate the view are trapped deep inside the JS bundle, which must first be received and evaluated by the browser, making it inaccessible to the PHP server rendering piece. However, we can write an adapter to derive required resources from the route and query params – see [prototype](https://github.com/Automattic/woocommerce-payments/commit/1c527064a43c4db1a866c8998579d6a72d1c1d2c) implementing one possible prefetching scheme for a few routes. In many cases, the query params can be mapped directed to what the REST API handler expects (as is already done in client-side hooks [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/hooks.js#L33)] and resolvers [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/resolvers.js#L47)]). It looks like this will be good enough for all current routes except possibly when advanced filters are present, which I haven't attempted to prefetch due to the complexity of the mapping. (It would be nice to have framework-level support for this but the routing logic [[code](https://github.com/woocommerce/woocommerce-admin/blob/d2fab15cb2b6e8cce8fd258e0d07580241334e9f/src/PageController.php#L116-L122)] doesn't currently even support patterns so we cannot do `/dispute/:dispute_id` even though the bundled client-side routing library fully supports it.) Other considerations: - One advantage of fetching the data from JS rather than PHP is that multiple requests are made in parallel – is there a way to do this from PHP without adding too much complexity? - Since some of the same mapping logic would be implemented in the plugin PHP as in the browser code, we'd want to be sure they stayed in sync rather than drifting apart by accident – bidirectional references in code comments, perhaps - Are there any good alternative ways of determining which requests will be needed by the view, or having the same specification be ingested in both contexts? - Should the page query params be made to match the API query params more closely, to simplify the mapping? AFAIK the only hard limitation here is that other query params (`page`, in particular) are reserved, but we could make explicit exceptions for mapping those (just that one?) - Would be worth doing some timing comparisons, primary consideration being the _time to load all data_, with _time to first UI render_ as another concern. (Some casual testing with a stopwatch _seems_ to indicate a quicker loading experience even with the deposit details screen preloading 3 resources in series)
True
Consider prefetching API data on page load - At the cost of skipping the loading placeholder view, we could **speed up loading time on initial page load** by pre<s>fetch</s>loading the data that will be needed by the view, in order to avoid round-trips on what I'd expect to be the most latency-prone leg of the stack, browser to site. Currently, descriptors of what data is needed to populate the view are trapped deep inside the JS bundle, which must first be received and evaluated by the browser, making it inaccessible to the PHP server rendering piece. However, we can write an adapter to derive required resources from the route and query params – see [prototype](https://github.com/Automattic/woocommerce-payments/commit/1c527064a43c4db1a866c8998579d6a72d1c1d2c) implementing one possible prefetching scheme for a few routes. In many cases, the query params can be mapped directed to what the REST API handler expects (as is already done in client-side hooks [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/hooks.js#L33)] and resolvers [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/resolvers.js#L47)]). It looks like this will be good enough for all current routes except possibly when advanced filters are present, which I haven't attempted to prefetch due to the complexity of the mapping. (It would be nice to have framework-level support for this but the routing logic [[code](https://github.com/woocommerce/woocommerce-admin/blob/d2fab15cb2b6e8cce8fd258e0d07580241334e9f/src/PageController.php#L116-L122)] doesn't currently even support patterns so we cannot do `/dispute/:dispute_id` even though the bundled client-side routing library fully supports it.) Other considerations: - One advantage of fetching the data from JS rather than PHP is that multiple requests are made in parallel – is there a way to do this from PHP without adding too much complexity? - Since some of the same mapping logic would be implemented in the plugin PHP as in the browser code, we'd want to be sure they stayed in sync rather than drifting apart by accident – bidirectional references in code comments, perhaps - Are there any good alternative ways of determining which requests will be needed by the view, or having the same specification be ingested in both contexts? - Should the page query params be made to match the API query params more closely, to simplify the mapping? AFAIK the only hard limitation here is that other query params (`page`, in particular) are reserved, but we could make explicit exceptions for mapping those (just that one?) - Would be worth doing some timing comparisons, primary consideration being the _time to load all data_, with _time to first UI render_ as another concern. (Some casual testing with a stopwatch _seems_ to indicate a quicker loading experience even with the deposit details screen preloading 3 resources in series)
non_process
consider prefetching api data on page load at the cost of skipping the loading placeholder view we could speed up loading time on initial page load by pre fetch loading the data that will be needed by the view in order to avoid round trips on what i d expect to be the most latency prone leg of the stack browser to site currently descriptors of what data is needed to populate the view are trapped deep inside the js bundle which must first be received and evaluated by the browser making it inaccessible to the php server rendering piece however we can write an adapter to derive required resources from the route and query params – see implementing one possible prefetching scheme for a few routes in many cases the query params can be mapped directed to what the rest api handler expects as is already done in client side hooks and resolvers it looks like this will be good enough for all current routes except possibly when advanced filters are present which i haven t attempted to prefetch due to the complexity of the mapping it would be nice to have framework level support for this but the routing logic doesn t currently even support patterns so we cannot do dispute dispute id even though the bundled client side routing library fully supports it other considerations one advantage of fetching the data from js rather than php is that multiple requests are made in parallel – is there a way to do this from php without adding too much complexity since some of the same mapping logic would be implemented in the plugin php as in the browser code we d want to be sure they stayed in sync rather than drifting apart by accident – bidirectional references in code comments perhaps are there any good alternative ways of determining which requests will be needed by the view or having the same specification be ingested in both contexts should the page query params be made to match the api query params more closely to simplify the mapping afaik the only hard limitation here is that other query params page in particular are reserved but we could make explicit exceptions for mapping those just that one would be worth doing some timing comparisons primary consideration being the time to load all data with time to first ui render as another concern some casual testing with a stopwatch seems to indicate a quicker loading experience even with the deposit details screen preloading resources in series
0
485,330
13,964,181,530
IssuesEvent
2020-10-25 17:06:17
AY2021S1-CS2103T-T13-2/tp
https://api.github.com/repos/AY2021S1-CS2103T-T13-2/tp
closed
As an expert user, I want to be able to see the statistics of my overall performance of a certain tag
priority.High type.Story
Statistics aggregated by `tag`.
1.0
As an expert user, I want to be able to see the statistics of my overall performance of a certain tag - Statistics aggregated by `tag`.
non_process
as an expert user i want to be able to see the statistics of my overall performance of a certain tag statistics aggregated by tag
0
342,111
10,312,180,807
IssuesEvent
2019-08-29 19:12:28
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
Default portal is not created by default
Priority: Low Type: Bug
I see that in the logs: Aug 29 08:40:28 unicorn packetfence: WARN pfperl-api(11281): Can't stat /usr/local/pf/html/captive-portal/profile-templates/default: No such file or directory at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/ConnectionProfiles.pm line 393. I wonder if we should create by default ?
1.0
Default portal is not created by default - I see that in the logs: Aug 29 08:40:28 unicorn packetfence: WARN pfperl-api(11281): Can't stat /usr/local/pf/html/captive-portal/profile-templates/default: No such file or directory at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/ConnectionProfiles.pm line 393. I wonder if we should create by default ?
non_process
default portal is not created by default i see that in the logs aug unicorn packetfence warn pfperl api can t stat usr local pf html captive portal profile templates default no such file or directory at usr local pf lib pf unifiedapi controller config connectionprofiles pm line i wonder if we should create by default
0
40,255
10,478,844,346
IssuesEvent
2019-09-24 01:46:10
filecoin-project/go-filecoin
https://api.github.com/repos/filecoin-project/go-filecoin
closed
[bug] go run ./build/ lint warning messages after libp2p upgrade
A-build C-bug
**Describe the bug** We updated the version of `github.com/golangci/golangci-lint` (v1.17.0) we use when upgrading go-libp2p. This update results in the following warning logs being returned when the linter is ran: ``` $ go run ./build/ lint Linting ./... ... go run github.com/golangci/golangci-lint/cmd/golangci-lint run level=warning msg="[runner/golint] Golint: can't lint 3 files: no file name for file &{Doc:<nil> Package:6844383 Name:bls Decls:[0xc008dc4780 0xc008dc47c0 0xc008dc4800 0xc008dc48c0 0xc008dc4900 0xc008dc4940 0xc008dc4980 0xc008dc49c0 0xc008dc4a40] Scope:scope 0xc003ec3000 {\n\tconst SignatureBytes\n\tconst PublicKeyBytes\n\ttype PublicKey\n\ttype PrivateKey\n\ttype Message\n\ttype Digest\n\tconst PrivateKeyBytes\n\tconst DigestBytes\n\ttype Signature\n}\n Imports:[] Unresolved:[byte byte byte byte byte] Comments:[0xc007b9e2e0 0xc007b9e360 0xc007b9e3e0 0xc007b9e4a0 0xc007b9e520 0xc007b9e5c0 0xc007b9e660 0xc007b9e720 0xc007b9e7a0]}" ``` This was introduced in a1771a863590be1432979fea9dc98d5bc99e4d0e **To Reproduce** run: `go run ./build/ lint` **Expected behavior** No warning logs are returned when the command is ran. **Version information** <!-- If you are having issue building go-filecoin please include the following: --> - Go: go1.12.1 - Commit: a1771a863590be1432979fea9dc98d5bc99e4d0e
1.0
[bug] go run ./build/ lint warning messages after libp2p upgrade - **Describe the bug** We updated the version of `github.com/golangci/golangci-lint` (v1.17.0) we use when upgrading go-libp2p. This update results in the following warning logs being returned when the linter is ran: ``` $ go run ./build/ lint Linting ./... ... go run github.com/golangci/golangci-lint/cmd/golangci-lint run level=warning msg="[runner/golint] Golint: can't lint 3 files: no file name for file &{Doc:<nil> Package:6844383 Name:bls Decls:[0xc008dc4780 0xc008dc47c0 0xc008dc4800 0xc008dc48c0 0xc008dc4900 0xc008dc4940 0xc008dc4980 0xc008dc49c0 0xc008dc4a40] Scope:scope 0xc003ec3000 {\n\tconst SignatureBytes\n\tconst PublicKeyBytes\n\ttype PublicKey\n\ttype PrivateKey\n\ttype Message\n\ttype Digest\n\tconst PrivateKeyBytes\n\tconst DigestBytes\n\ttype Signature\n}\n Imports:[] Unresolved:[byte byte byte byte byte] Comments:[0xc007b9e2e0 0xc007b9e360 0xc007b9e3e0 0xc007b9e4a0 0xc007b9e520 0xc007b9e5c0 0xc007b9e660 0xc007b9e720 0xc007b9e7a0]}" ``` This was introduced in a1771a863590be1432979fea9dc98d5bc99e4d0e **To Reproduce** run: `go run ./build/ lint` **Expected behavior** No warning logs are returned when the command is ran. **Version information** <!-- If you are having issue building go-filecoin please include the following: --> - Go: go1.12.1 - Commit: a1771a863590be1432979fea9dc98d5bc99e4d0e
non_process
go run build lint warning messages after upgrade describe the bug we updated the version of github com golangci golangci lint we use when upgrading go this update results in the following warning logs being returned when the linter is ran go run build lint linting go run github com golangci golangci lint cmd golangci lint run level warning msg golint can t lint files no file name for file doc package name bls decls scope scope n tconst signaturebytes n tconst publickeybytes n ttype publickey n ttype privatekey n ttype message n ttype digest n tconst privatekeybytes n tconst digestbytes n ttype signature n n imports unresolved comments this was introduced in to reproduce run go run build lint expected behavior no warning logs are returned when the command is ran version information go commit
0
21,614
30,017,402,656
IssuesEvent
2023-06-26 19:54:19
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Measure and optimize pty host reconnection
feature-request verified perf terminal-process
Reconnecting/restoring terminals becomes quite slow when you have many terminals. We should measure what's happening with the pty host trace logs and try optimize.
1.0
Measure and optimize pty host reconnection - Reconnecting/restoring terminals becomes quite slow when you have many terminals. We should measure what's happening with the pty host trace logs and try optimize.
process
measure and optimize pty host reconnection reconnecting restoring terminals becomes quite slow when you have many terminals we should measure what s happening with the pty host trace logs and try optimize
1
17,879
23,832,320,577
IssuesEvent
2022-09-05 23:22:04
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Out of date Tentacle and Calamari warnings are no longer useful
kind/enhancement feature/usability size/small feature/ops-processes p3 area/core team/fire-and-motion
Currently, users of Octopus Server using Machine Policy to automatically upgrade tentacles during deployments can see warnings regarding out of date Tentacle and Calamari instances. [Relevant documentation here](https://octopus.com/docs/infrastructure/machine-policies#MachinePolicies-ConfigurehowCalamariandTentacleareupdated) As evidenced by customer queries, we are finding that these warnings do not hold much value anymore, now that the default behaviour is 'on demand' upgrades. For e.g. ``` As a user I want to be able to upgrade my Octopus server without negatively impacting deployment schedules nor do I want to cause undue consternation among the operations support folks. ``` The 'Healthy with warnings' icons shown in the deployment targets view (and other locations) are also causing some confusion. ## Proposed solution We think that removing these warnings entirely will encourage use of the default behaviour, which carries better performance characteristics overall, and not make unnecessary alarm bells sound off for customers. ## Before ![before](https://user-images.githubusercontent.com/579657/60071244-9e48d400-9758-11e9-8c61-2582a91a7f99.png) 'Healthy with warnings' icons are displayed for such deployment targets. ## After ![after](https://user-images.githubusercontent.com/579657/60071260-a43eb500-9758-11e9-9fb0-52beba41531a.png) Such targets would not be labelled as 'Healthy with warnings' ## Workarounds At this stage, ignore the warnings, or move to the 'Always up to date' model. The trade-off for the latter is that in large ecosystems, tentacle upgrades can take a significant amount of time due to a CPU bound limit on how many upgrades can be done at once. ## Links [public support ticket](https://help.octopus.com/t/increase-number-of-concurrent-tentacle-updates/23368/6) [internal discussion thread](https://octopusdeploy.slack.com/archives/C033W4273/p1561418888036100) [related issue](https://github.com/OctopusDeploy/Issues/issues/4374) # Prerequisites - [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested - [x] I have written a descriptive issue title - [x] I have linked the original source of this feature request - [x] I have tagged the issue appropriately (area/*, kind/enhancement)
1.0
Out of date Tentacle and Calamari warnings are no longer useful - Currently, users of Octopus Server using Machine Policy to automatically upgrade tentacles during deployments can see warnings regarding out of date Tentacle and Calamari instances. [Relevant documentation here](https://octopus.com/docs/infrastructure/machine-policies#MachinePolicies-ConfigurehowCalamariandTentacleareupdated) As evidenced by customer queries, we are finding that these warnings do not hold much value anymore, now that the default behaviour is 'on demand' upgrades. For e.g. ``` As a user I want to be able to upgrade my Octopus server without negatively impacting deployment schedules nor do I want to cause undue consternation among the operations support folks. ``` The 'Healthy with warnings' icons shown in the deployment targets view (and other locations) are also causing some confusion. ## Proposed solution We think that removing these warnings entirely will encourage use of the default behaviour, which carries better performance characteristics overall, and not make unnecessary alarm bells sound off for customers. ## Before ![before](https://user-images.githubusercontent.com/579657/60071244-9e48d400-9758-11e9-8c61-2582a91a7f99.png) 'Healthy with warnings' icons are displayed for such deployment targets. ## After ![after](https://user-images.githubusercontent.com/579657/60071260-a43eb500-9758-11e9-9fb0-52beba41531a.png) Such targets would not be labelled as 'Healthy with warnings' ## Workarounds At this stage, ignore the warnings, or move to the 'Always up to date' model. The trade-off for the latter is that in large ecosystems, tentacle upgrades can take a significant amount of time due to a CPU bound limit on how many upgrades can be done at once. ## Links [public support ticket](https://help.octopus.com/t/increase-number-of-concurrent-tentacle-updates/23368/6) [internal discussion thread](https://octopusdeploy.slack.com/archives/C033W4273/p1561418888036100) [related issue](https://github.com/OctopusDeploy/Issues/issues/4374) # Prerequisites - [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested - [x] I have written a descriptive issue title - [x] I have linked the original source of this feature request - [x] I have tagged the issue appropriately (area/*, kind/enhancement)
process
out of date tentacle and calamari warnings are no longer useful currently users of octopus server using machine policy to automatically upgrade tentacles during deployments can see warnings regarding out of date tentacle and calamari instances as evidenced by customer queries we are finding that these warnings do not hold much value anymore now that the default behaviour is on demand upgrades for e g as a user i want to be able to upgrade my octopus server without negatively impacting deployment schedules nor do i want to cause undue consternation among the operations support folks the healthy with warnings icons shown in the deployment targets view and other locations are also causing some confusion proposed solution we think that removing these warnings entirely will encourage use of the default behaviour which carries better performance characteristics overall and not make unnecessary alarm bells sound off for customers before healthy with warnings icons are displayed for such deployment targets after such targets would not be labelled as healthy with warnings workarounds at this stage ignore the warnings or move to the always up to date model the trade off for the latter is that in large ecosystems tentacle upgrades can take a significant amount of time due to a cpu bound limit on how many upgrades can be done at once links prerequisites i have searched and issues to make sure it isn t already requested i have written a descriptive issue title i have linked the original source of this feature request i have tagged the issue appropriately area kind enhancement
1
350,819
31,932,351,254
IssuesEvent
2023-09-19 08:17:42
masters2023-project-06-second-hand/be-a
https://api.github.com/repos/masters2023-project-06-second-hand/be-a
opened
[BE] μ±„νŒ…λ°© 생성, μ±„νŒ… 참여정보 μ—”ν‹°ν‹° 생성
🚦test ✨ feature
## ✨ ν•΄λ‹Ή κΈ°λŠ₯을 κ΅¬ν˜„ν•˜κΈ° μœ„ν•΄ ν•  일이 λ¬΄μ—‡μΈκ°€μš”? - [ ] μ±„νŒ…λ°© 쑰회/생성 - [ ] μ±„νŒ… 참여정보 μ—”ν‹°ν‹° - [ ] ## βœ… Test Case - [ ] - [ ]
1.0
[BE] μ±„νŒ…λ°© 생성, μ±„νŒ… 참여정보 μ—”ν‹°ν‹° 생성 - ## ✨ ν•΄λ‹Ή κΈ°λŠ₯을 κ΅¬ν˜„ν•˜κΈ° μœ„ν•΄ ν•  일이 λ¬΄μ—‡μΈκ°€μš”? - [ ] μ±„νŒ…λ°© 쑰회/생성 - [ ] μ±„νŒ… 참여정보 μ—”ν‹°ν‹° - [ ] ## βœ… Test Case - [ ] - [ ]
non_process
μ±„νŒ…λ°© 생성 μ±„νŒ… 참여정보 μ—”ν‹°ν‹° 생성 ✨ ν•΄λ‹Ή κΈ°λŠ₯을 κ΅¬ν˜„ν•˜κΈ° μœ„ν•΄ ν•  일이 λ¬΄μ—‡μΈκ°€μš” μ±„νŒ…λ°© 쑰회 생성 μ±„νŒ… 참여정보 μ—”ν‹°ν‹° βœ… test case
0
10,090
13,044,162,044
IssuesEvent
2020-07-29 03:47:28
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `SubDatetimeAndDuration` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `SubDatetimeAndDuration` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `SubDatetimeAndDuration` from TiDB - ## Description Port the scalar function `SubDatetimeAndDuration` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function subdatetimeandduration from tidb description port the scalar function subdatetimeandduration from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
1
1,873
4,699,340,790
IssuesEvent
2016-10-12 15:25:00
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
49 Kay Street
processed regression
Is it more important for the house number to match in the result or the street name? We're seeing `49 De Kay Street` show up higher than `3`,`4`, and `5 KAY STREET`. Separate issue here is that there are OA records with all caps coming through now, due to the fact that OA stopped normalizing their data for capitalization.
1.0
49 Kay Street - Is it more important for the house number to match in the result or the street name? We're seeing `49 De Kay Street` show up higher than `3`,`4`, and `5 KAY STREET`. Separate issue here is that there are OA records with all caps coming through now, due to the fact that OA stopped normalizing their data for capitalization.
process
kay street is it more important for the house number to match in the result or the street name we re seeing de kay street show up higher than and kay street separate issue here is that there are oa records with all caps coming through now due to the fact that oa stopped normalizing their data for capitalization
1
330,307
10,038,306,541
IssuesEvent
2019-07-18 14:54:13
Citykleta/web-app
https://api.github.com/repos/Citykleta/web-app
closed
Route exploration
component:Itinerary enhancement priority:medium
At the moment, a user can quickly find a route going from A to B going through various intermediate points. It would be however important to improve this part of the application. - [ ] change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown - [ ] propose alternatives - [ ] give a way to the user to preview the instructions
1.0
Route exploration - At the moment, a user can quickly find a route going from A to B going through various intermediate points. It would be however important to improve this part of the application. - [ ] change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown - [ ] propose alternatives - [ ] give a way to the user to preview the instructions
non_process
route exploration at the moment a user can quickly find a route going from a to b going through various intermediate points it would be however important to improve this part of the application change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown propose alternatives give a way to the user to preview the instructions
0
135,988
5,267,789,695
IssuesEvent
2017-02-05 02:40:20
TrinityCore/TrinityCore
https://api.github.com/repos/TrinityCore/TrinityCore
opened
[3.3.5] Extra lashtails in Barrens
Branch-3.3.5a Comp-Database Priority-VeryLow Sub-Spawns
[//]: # (***************************************************************) [//]: # (** DON'T DELETE THIS TEMPLATE ELSE YOUR ISSUE WILL BE CLOSED **) [//]: # (***************************************************************) **Description:** While looking at some videos of the game, on the subject of pooling. This one in particular at around this point: https://youtu.be/xyhkTcJuOoE?t=168 There are some extra lashtails in the Barrens that don't belong there, for sure you would really not see them spawned like this in reality. As seen here: ![image](https://cloud.githubusercontent.com/assets/173349/22623241/00f124ca-eb4b-11e6-850d-09ff5aaedb6a.png) The one that belongs can be seen to be roaming, the other 3 are static. The one that belongs is spawnId: 19816. Those that don't are: 40493, 49495, 49496. There is also another static one added 40494: ![image](https://cloud.githubusercontent.com/assets/173349/22623250/904540b6-eb4b-11e6-9c44-cb6f26565d5a.png) I can't say whether that one belongs or not. For sure it shouldn't be static. Soliciting opinion before I remove them/enable movement on the last one. **Current behaviour:** (Tell us what happens.) Too many lashtails. **Expected behaviour:** (Tell us what should happen instead.) Not so many lashtails. **Steps to reproduce the problem:** 1. .go creature 40493 2. Observe extra lashtails 3. ????? 4. Profit **Branch(es):** 3.3.5 / master (Tell us which branch(es) this issue affects.) 3.3.5 **TC rev. hash/commit:** e991a15dd7cbf5ba374f8942b7ea88f3a4d7b365 **TDB version:** TDB_full_world_335.62_2016_10_17 + as per hash above. **Operating system:** Linux [//]: # (This template is for problem reports. For other types of report, edit it accordingly.) [//]: # (If this is a crash report, include the crashlog with https://gist.github.com/) [//]: # (For fixes containing C++, create a Pull Request.)
1.0
[3.3.5] Extra lashtails in Barrens - [//]: # (***************************************************************) [//]: # (** DON'T DELETE THIS TEMPLATE ELSE YOUR ISSUE WILL BE CLOSED **) [//]: # (***************************************************************) **Description:** While looking at some videos of the game, on the subject of pooling. This one in particular at around this point: https://youtu.be/xyhkTcJuOoE?t=168 There are some extra lashtails in the Barrens that don't belong there, for sure you would really not see them spawned like this in reality. As seen here: ![image](https://cloud.githubusercontent.com/assets/173349/22623241/00f124ca-eb4b-11e6-850d-09ff5aaedb6a.png) The one that belongs can be seen to be roaming, the other 3 are static. The one that belongs is spawnId: 19816. Those that don't are: 40493, 49495, 49496. There is also another static one added 40494: ![image](https://cloud.githubusercontent.com/assets/173349/22623250/904540b6-eb4b-11e6-9c44-cb6f26565d5a.png) I can't say whether that one belongs or not. For sure it shouldn't be static. Soliciting opinion before I remove them/enable movement on the last one. **Current behaviour:** (Tell us what happens.) Too many lashtails. **Expected behaviour:** (Tell us what should happen instead.) Not so many lashtails. **Steps to reproduce the problem:** 1. .go creature 40493 2. Observe extra lashtails 3. ????? 4. Profit **Branch(es):** 3.3.5 / master (Tell us which branch(es) this issue affects.) 3.3.5 **TC rev. hash/commit:** e991a15dd7cbf5ba374f8942b7ea88f3a4d7b365 **TDB version:** TDB_full_world_335.62_2016_10_17 + as per hash above. **Operating system:** Linux [//]: # (This template is for problem reports. For other types of report, edit it accordingly.) [//]: # (If this is a crash report, include the crashlog with https://gist.github.com/) [//]: # (For fixes containing C++, create a Pull Request.)
non_process
extra lashtails in barrens don t delete this template else your issue will be closed description while looking at some videos of the game on the subject of pooling this one in particular at around this point there are some extra lashtails in the barrens that don t belong there for sure you would really not see them spawned like this in reality as seen here the one that belongs can be seen to be roaming the other are static the one that belongs is spawnid those that don t are there is also another static one added i can t say whether that one belongs or not for sure it shouldn t be static soliciting opinion before i remove them enable movement on the last one current behaviour tell us what happens too many lashtails expected behaviour tell us what should happen instead not so many lashtails steps to reproduce the problem go creature observe extra lashtails profit branch es master tell us which branch es this issue affects tc rev hash commit tdb version tdb full world as per hash above operating system linux this template is for problem reports for other types of report edit it accordingly if this is a crash report include the crashlog with for fixes containing c create a pull request
0
16,515
21,527,132,701
IssuesEvent
2022-04-28 19:40:10
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Something went wrong error for card, no 'dataset_query' returned by API/card
Type:Bug Priority:P2 Querying/Processor .Backend .Regression
**Describe the bug** When loading a user-created question, the generic "Something went wrong" (+flaming server) error page is shown. It is not clear why this happened or what the exact repro steps are. Some notes: - User interface default language is set to German - Instance was upgraded from v0.35.4 to v0.39.2, question created under v0.35.4 - The GET /api/card endpoint does not include a "dataset_query" object, unlike a normal response - The question has a join to a saved question as per #12928 / #16353 - The question has filters on: - an expression in the saved question - a date field, where the filter includes a time - The expression is an integer field divided by the 'count' metric calculated in the question If it's not able to be reproduced, I would like a pointer on how to fix the question in the backend. Though I'm not sure if the "dataset_query" object looks correct or what should be changed if it's wrong. **Logs** The browser (Firefox 88) console says: ``` Error: TypeError: c.dataset_query is null ``` The Metabase database entry for the affected card (note leading space). ``` > select dataset_query from report_card where id = 123; {"type":"query","query":{"source-table":"card__122","joins":[{"fields":"all","source-table":555,"condition":["=",["field-literal","person_id","type/Text"],["joined-field","persons",["field-id",16666]]],"alias":"persons"}],"filter":["and",["<",["datetime-field",["field-literal","date_seen","type/Date"],"minute"],"2021-05-01T12:30:00"],["=",["field-literal","keep_entry","type/Float"],1]]},"database":12} ``` When viewing the question, the Metabase logs say: ``` qtp432197459-6257 | 2021-06-08 07:51:13,041 | ERROR | metabase.models.interface | Unable to normalize: {:type "query", :query {:source-table "card__122", :joins [{:fields "all", :source-table 555, :condition ["=" ["field-literal" "person_id" "type/Text"] ["joined-field" "persons" ["field-id" 16666]]], :alias "persons"}], :filter ["and" ["<" ["datetime-field" ["field-literal" "date_seen" "type/Date"] "minute"] "2021-05-01T12:30:00"] ["=" ["field-literal" "keep_entry" "type/Float"] 1]]}, :database 12} clojure.lang.ExceptionInfo: Error normalizing query at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:795) ~[metabase.jar:?] at metabase.models.interface$maybe_normalize.invokeStatic(interface.clj:68) ~[metabase.jar:?] at metabase.models.interface$maybe_normalize.invoke(interface.clj:67) ~[metabase.jar:?] at metabase.models.interface$catch_normalization_exceptions$fn__24676.invoke(interface.clj:78) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at toucan.models$apply_type_fns$iter__24558__24562$fn__24563.invoke(models.clj:304) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?] at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?] at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?] at toucan.models$apply_type_fns.invokeStatic(models.clj:302) ~[metabase.jar:?] at toucan.models$apply_type_fns.invoke(models.clj:299) ~[metabase.jar:?] at toucan.models$do_post_select.invokeStatic(models.clj:349) ~[metabase.jar:?] at toucan.models$do_post_select.invoke(models.clj:344) ~[metabase.jar:?] at toucan.db$do_post_select$iter__28532__28536$fn__28537.invoke(db.clj:373) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.lang.LazilyPersistentVector.create(LazilyPersistentVector.java:44) ~[metabase.jar:?] at clojure.core$vec.invokeStatic(core.clj:377) ~[metabase.jar:?] at clojure.core$vec.invoke(core.clj:367) ~[metabase.jar:?] at toucan.db$do_post_select.invokeStatic(db.clj:372) ~[metabase.jar:?] at toucan.db$do_post_select.invoke(db.clj:363) ~[metabase.jar:?] at toucan.db$simple_select.invokeStatic(db.clj:394) ~[metabase.jar:?] at toucan.db$simple_select.invoke(db.clj:383) ~[metabase.jar:?] at toucan.db$simple_select_one.invokeStatic(db.clj:420) ~[metabase.jar:?] at toucan.db$simple_select_one.invoke(db.clj:409) ~[metabase.jar:?] at toucan.db$select_one.invokeStatic(db.clj:627) ~[metabase.jar:?] at toucan.db$select_one.doInvoke(db.clj:620) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:139) ~[metabase.jar:?] at clojure.lang.Var.applyTo(Var.java:705) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:671) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at toucan.models$invoke_model.invokeStatic(models.clj:380) ~[metabase.jar:?] at toucan.models$invoke_model.doInvoke(models.clj:368) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:445) ~[metabase.jar:?] at toucan.models$invoke_model.invokeStatic(models.clj:378) ~[metabase.jar:?] at toucan.models$invoke_model.invoke(models.clj:368) ~[metabase.jar:?] at clojure.lang.AFn.applyToHelper(AFn.java:156) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:132) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:667) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at toucan.models$invoke_model_or_instance.invokeStatic(models.clj:395) ~[metabase.jar:?] at toucan.models$invoke_model_or_instance.doInvoke(models.clj:391) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:423) ~[metabase.jar:?] at metabase.models.card.CardInstance.invoke(card.clj:25) ~[metabase.jar:?] at metabase.api.card$fn__65802.invokeStatic(card.clj:162) ~[metabase.jar:?] at metabase.api.card$fn__65802.invoke(card.clj:159) ~[metabase.jar:?] at compojure.core$wrap_response$fn__12050.invoke(core.clj:160) ~[metabase.jar:?] at compojure.core$wrap_route_middleware$fn__12034.invoke(core.clj:132) ~[metabase.jar:?] at compojure.core$wrap_route_info$fn__12039.invoke(core.clj:139) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:151) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at metabase.server.middleware.auth$enforce_authentication$fn__56336.invoke(auth.clj:14) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at clojure.lang.AFn.applyToHelper(AFn.java:160) ~[metabase.jar:?] at clojure.lang.AFn.applyTo(AFn.java:144) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:665) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at metabase.server.routes$fn__83020$fn__83021.doInvoke(routes.clj:57) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:436) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__81060.invoke(exceptions.clj:98) ~[metabase.jar:?] at metabase.server.middleware.exceptions$catch_api_exceptions$fn__81057.invoke(exceptions.clj:86) ~[metabase.jar:?] at metabase.server.middleware.log$log_api_call$fn__82936$fn__82937.invoke(log.clj:195) ~[metabase.jar:?] at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216) ~[metabase.jar:?] at toucan.db$_do_with_call_counting.invoke(db.clj:209) ~[metabase.jar:?] at metabase.server.middleware.log$log_api_call$fn__82936.invoke(log.clj:189) ~[metabase.jar:?] at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__82572.invoke(browser_cookie.clj:30) ~[metabase.jar:?] at metabase.server.middleware.security$add_security_headers$fn__81022.invoke(security.clj:142) ~[metabase.jar:?] at metabase.server.middleware.json$wrap_json_body$fn__82717.invoke(json.clj:62) ~[metabase.jar:?] at metabase.server.middleware.json$wrap_streamed_json_response$fn__82735.invoke(json.clj:98) ~[metabase.jar:?] at ring.middleware.keyword_params$wrap_keyword_params$fn__83277.invoke(keyword_params.clj:55) ~[metabase.jar:?] at ring.middleware.params$wrap_params$fn__83293.invoke(params.clj:69) ~[metabase.jar:?] at metabase.server.middleware.misc$maybe_set_site_url$fn__35809.invoke(misc.clj:58) ~[metabase.jar:?] at metabase.server.middleware.session$bind_current_user$fn__43391$fn__43392.invoke(session.clj:248) ~[metabase.jar:?] at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:229) ~[metabase.jar:?] at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:221) ~[metabase.jar:?] at metabase.server.middleware.session$bind_current_user$fn__43391.invoke(session.clj:247) ~[metabase.jar:?] at metabase.server.middleware.session$wrap_current_user_info$fn__43378.invoke(session.clj:207) ~[metabase.jar:?] at metabase.server.middleware.session$wrap_session_id$fn__43364.invoke(session.clj:153) ~[metabase.jar:?] at metabase.server.middleware.auth$wrap_api_key$fn__56344.invoke(auth.clj:27) ~[metabase.jar:?] at ring.middleware.cookies$wrap_cookies$fn__83197.invoke(cookies.clj:216) ~[metabase.jar:?] at metabase.server.middleware.misc$add_content_type$fn__35792.invoke(misc.clj:27) ~[metabase.jar:?] at metabase.server.middleware.misc$disable_streaming_buffering$fn__35817.invoke(misc.clj:75) ~[metabase.jar:?] at ring.middleware.gzip$wrap_gzip$fn__83239.invoke(gzip.clj:86) ~[metabase.jar:?] at metabase.server.middleware.misc$bind_request$fn__35820.invoke(misc.clj:92) ~[metabase.jar:?] at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__82953.invoke(ssl.clj:48) ~[metabase.jar:?] at metabase.server$async_proxy_handler$fn__82516.invoke(server.clj:71) ~[metabase.jar:?] at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) ~[metabase.jar:?] at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) ~[metabase.jar:?] at org.eclipse.jetty.server.Server.handle(Server.java:516) ~[metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:383) ~[metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:556) [metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:375) [metabase.jar:?] at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:273) [metabase.jar:?] at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311) [metabase.jar:?] at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) [metabase.jar:?] at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:336) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:313) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:171) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:129) [metabase.jar:?] at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:375) [metabase.jar:?] at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:773) [metabase.jar:?] at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:905) [metabase.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: clojure.lang.ExceptionInfo: Error canonicalizing query at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:677) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more Caused by: clojure.lang.ExceptionInfo: Invalid MBQL clause at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:555) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?] at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?] at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?] at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?] at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?] at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?] at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?] at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?] at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?] at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?] at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?] at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more Caused by: clojure.lang.ExceptionInfo: Output of update-field-options does not match schema: [nil nil (named (not ("Invalid :temporal-unit for the specified :base-type." a-clojure.lang.PersistentArrayMap)) "options")] at metabase.mbql.util$fn__20573$update_field_options__20578.doInvoke(util.cljc:581) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:142) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:669) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at metabase.mbql.util$assoc_field_options.invokeStatic(util.cljc:590) ~[metabase.jar:?] at metabase.mbql.util$assoc_field_options.doInvoke(util.cljc:587) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:439) ~[metabase.jar:?] at metabase.mbql.util$with_temporal_unit.invokeStatic(util.cljc:595) ~[metabase.jar:?] at metabase.mbql.util$with_temporal_unit.invoke(util.cljc:592) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20861.invokeStatic(normalize.cljc:404) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20861.invoke(normalize.cljc:398) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_implicit_field_id.invokeStatic(normalize.cljc:363) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_implicit_field_id.invoke(normalize.cljc:360) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_simple_filter_clause.invokeStatic(normalize.cljc:455) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_simple_filter_clause.invoke(normalize.cljc:452) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20918$fn__20919.invoke(normalize.cljc:464) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at clojure.core$map$fn__5866.invoke(core.clj:2753) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?] at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?] at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_compound_filter_clause.invokeStatic(normalize.cljc:425) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_compound_filter_clause.invoke(normalize.cljc:423) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20888$fn__20889.invoke(normalize.cljc:433) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:552) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?] at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?] at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?] at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?] at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?] at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?] at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?] at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?] at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?] at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?] at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?] at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more ``` **To Reproduce** Not sure - have not found exactly how to reproduce it with the Sample Database. A similar question complains about having a time filter on a date field, but no "something went wrong" ``` Unsupported temporal bucketing: You can't bucket a :type/Date Field by :minute. ``` Maybe it is about the combination of metric + expression + saved question + filters? **Expected behavior** The question should load without error, or at least an error that the user can rectify. The "something went wrong" page does not show any of the normal Question UI controls. **Information about your Metabase Installation:** You can get this information by going to Admin -> Troubleshooting. - Your browser and the version: (e.x. Chrome 52.1, Firefox 48.0, Safari 11.1, …) Firefox 88 - Your operating system: (e.x. OS X 10.10, Windows 10.1809, Ubuntu 16.04, …) CentOS 7 - Your databases: (e.x. MySQL, Postgres, MongoDB, …) PostgreSQL 10 - Metabase version: (e.x. 0.19.3) Metabase v0.39.2 - Metabase hosting environment: (e.x. Mac app, Elastic Beanstalk, Docker, Heroku, Jar-file on Ubuntu, …) CentOS 7 - Metabase internal database: (e.x. H2 (default), Postgres or MySQL) PostgreSQL 10 **Severity** Blocking with respect to not being able to view or fix this particular question
1.0
Something went wrong error for card, no 'dataset_query' returned by API/card - **Describe the bug** When loading a user-created question, the generic "Something went wrong" (+flaming server) error page is shown. It is not clear why this happened or what the exact repro steps are. Some notes: - User interface default language is set to German - Instance was upgraded from v0.35.4 to v0.39.2, question created under v0.35.4 - The GET /api/card endpoint does not include a "dataset_query" object, unlike a normal response - The question has a join to a saved question as per #12928 / #16353 - The question has filters on: - an expression in the saved question - a date field, where the filter includes a time - The expression is an integer field divided by the 'count' metric calculated in the question If it's not able to be reproduced, I would like a pointer on how to fix the question in the backend. Though I'm not sure if the "dataset_query" object looks correct or what should be changed if it's wrong. **Logs** The browser (Firefox 88) console says: ``` Error: TypeError: c.dataset_query is null ``` The Metabase database entry for the affected card (note leading space). ``` > select dataset_query from report_card where id = 123; {"type":"query","query":{"source-table":"card__122","joins":[{"fields":"all","source-table":555,"condition":["=",["field-literal","person_id","type/Text"],["joined-field","persons",["field-id",16666]]],"alias":"persons"}],"filter":["and",["<",["datetime-field",["field-literal","date_seen","type/Date"],"minute"],"2021-05-01T12:30:00"],["=",["field-literal","keep_entry","type/Float"],1]]},"database":12} ``` When viewing the question, the Metabase logs say: ``` qtp432197459-6257 | 2021-06-08 07:51:13,041 | ERROR | metabase.models.interface | Unable to normalize: {:type "query", :query {:source-table "card__122", :joins [{:fields "all", :source-table 555, :condition ["=" ["field-literal" "person_id" "type/Text"] ["joined-field" "persons" ["field-id" 16666]]], :alias "persons"}], :filter ["and" ["<" ["datetime-field" ["field-literal" "date_seen" "type/Date"] "minute"] "2021-05-01T12:30:00"] ["=" ["field-literal" "keep_entry" "type/Float"] 1]]}, :database 12} clojure.lang.ExceptionInfo: Error normalizing query at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:795) ~[metabase.jar:?] at metabase.models.interface$maybe_normalize.invokeStatic(interface.clj:68) ~[metabase.jar:?] at metabase.models.interface$maybe_normalize.invoke(interface.clj:67) ~[metabase.jar:?] at metabase.models.interface$catch_normalization_exceptions$fn__24676.invoke(interface.clj:78) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at toucan.models$apply_type_fns$iter__24558__24562$fn__24563.invoke(models.clj:304) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?] at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?] at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?] at toucan.models$apply_type_fns.invokeStatic(models.clj:302) ~[metabase.jar:?] at toucan.models$apply_type_fns.invoke(models.clj:299) ~[metabase.jar:?] at toucan.models$do_post_select.invokeStatic(models.clj:349) ~[metabase.jar:?] at toucan.models$do_post_select.invoke(models.clj:344) ~[metabase.jar:?] at toucan.db$do_post_select$iter__28532__28536$fn__28537.invoke(db.clj:373) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.lang.LazilyPersistentVector.create(LazilyPersistentVector.java:44) ~[metabase.jar:?] at clojure.core$vec.invokeStatic(core.clj:377) ~[metabase.jar:?] at clojure.core$vec.invoke(core.clj:367) ~[metabase.jar:?] at toucan.db$do_post_select.invokeStatic(db.clj:372) ~[metabase.jar:?] at toucan.db$do_post_select.invoke(db.clj:363) ~[metabase.jar:?] at toucan.db$simple_select.invokeStatic(db.clj:394) ~[metabase.jar:?] at toucan.db$simple_select.invoke(db.clj:383) ~[metabase.jar:?] at toucan.db$simple_select_one.invokeStatic(db.clj:420) ~[metabase.jar:?] at toucan.db$simple_select_one.invoke(db.clj:409) ~[metabase.jar:?] at toucan.db$select_one.invokeStatic(db.clj:627) ~[metabase.jar:?] at toucan.db$select_one.doInvoke(db.clj:620) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:139) ~[metabase.jar:?] at clojure.lang.Var.applyTo(Var.java:705) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:671) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at toucan.models$invoke_model.invokeStatic(models.clj:380) ~[metabase.jar:?] at toucan.models$invoke_model.doInvoke(models.clj:368) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:445) ~[metabase.jar:?] at toucan.models$invoke_model.invokeStatic(models.clj:378) ~[metabase.jar:?] at toucan.models$invoke_model.invoke(models.clj:368) ~[metabase.jar:?] at clojure.lang.AFn.applyToHelper(AFn.java:156) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:132) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:667) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at toucan.models$invoke_model_or_instance.invokeStatic(models.clj:395) ~[metabase.jar:?] at toucan.models$invoke_model_or_instance.doInvoke(models.clj:391) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:423) ~[metabase.jar:?] at metabase.models.card.CardInstance.invoke(card.clj:25) ~[metabase.jar:?] at metabase.api.card$fn__65802.invokeStatic(card.clj:162) ~[metabase.jar:?] at metabase.api.card$fn__65802.invoke(card.clj:159) ~[metabase.jar:?] at compojure.core$wrap_response$fn__12050.invoke(core.clj:160) ~[metabase.jar:?] at compojure.core$wrap_route_middleware$fn__12034.invoke(core.clj:132) ~[metabase.jar:?] at compojure.core$wrap_route_info$fn__12039.invoke(core.clj:139) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:151) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at metabase.server.middleware.auth$enforce_authentication$fn__56336.invoke(auth.clj:14) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at clojure.lang.AFn.applyToHelper(AFn.java:160) ~[metabase.jar:?] at clojure.lang.AFn.applyTo(AFn.java:144) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:665) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at metabase.server.routes$fn__83020$fn__83021.doInvoke(routes.clj:57) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:436) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?] at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?] at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?] at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__81060.invoke(exceptions.clj:98) ~[metabase.jar:?] at metabase.server.middleware.exceptions$catch_api_exceptions$fn__81057.invoke(exceptions.clj:86) ~[metabase.jar:?] at metabase.server.middleware.log$log_api_call$fn__82936$fn__82937.invoke(log.clj:195) ~[metabase.jar:?] at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216) ~[metabase.jar:?] at toucan.db$_do_with_call_counting.invoke(db.clj:209) ~[metabase.jar:?] at metabase.server.middleware.log$log_api_call$fn__82936.invoke(log.clj:189) ~[metabase.jar:?] at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__82572.invoke(browser_cookie.clj:30) ~[metabase.jar:?] at metabase.server.middleware.security$add_security_headers$fn__81022.invoke(security.clj:142) ~[metabase.jar:?] at metabase.server.middleware.json$wrap_json_body$fn__82717.invoke(json.clj:62) ~[metabase.jar:?] at metabase.server.middleware.json$wrap_streamed_json_response$fn__82735.invoke(json.clj:98) ~[metabase.jar:?] at ring.middleware.keyword_params$wrap_keyword_params$fn__83277.invoke(keyword_params.clj:55) ~[metabase.jar:?] at ring.middleware.params$wrap_params$fn__83293.invoke(params.clj:69) ~[metabase.jar:?] at metabase.server.middleware.misc$maybe_set_site_url$fn__35809.invoke(misc.clj:58) ~[metabase.jar:?] at metabase.server.middleware.session$bind_current_user$fn__43391$fn__43392.invoke(session.clj:248) ~[metabase.jar:?] at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:229) ~[metabase.jar:?] at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:221) ~[metabase.jar:?] at metabase.server.middleware.session$bind_current_user$fn__43391.invoke(session.clj:247) ~[metabase.jar:?] at metabase.server.middleware.session$wrap_current_user_info$fn__43378.invoke(session.clj:207) ~[metabase.jar:?] at metabase.server.middleware.session$wrap_session_id$fn__43364.invoke(session.clj:153) ~[metabase.jar:?] at metabase.server.middleware.auth$wrap_api_key$fn__56344.invoke(auth.clj:27) ~[metabase.jar:?] at ring.middleware.cookies$wrap_cookies$fn__83197.invoke(cookies.clj:216) ~[metabase.jar:?] at metabase.server.middleware.misc$add_content_type$fn__35792.invoke(misc.clj:27) ~[metabase.jar:?] at metabase.server.middleware.misc$disable_streaming_buffering$fn__35817.invoke(misc.clj:75) ~[metabase.jar:?] at ring.middleware.gzip$wrap_gzip$fn__83239.invoke(gzip.clj:86) ~[metabase.jar:?] at metabase.server.middleware.misc$bind_request$fn__35820.invoke(misc.clj:92) ~[metabase.jar:?] at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__82953.invoke(ssl.clj:48) ~[metabase.jar:?] at metabase.server$async_proxy_handler$fn__82516.invoke(server.clj:71) ~[metabase.jar:?] at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) ~[metabase.jar:?] at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) ~[metabase.jar:?] at org.eclipse.jetty.server.Server.handle(Server.java:516) ~[metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:383) ~[metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:556) [metabase.jar:?] at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:375) [metabase.jar:?] at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:273) [metabase.jar:?] at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311) [metabase.jar:?] at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) [metabase.jar:?] at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:336) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:313) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:171) [metabase.jar:?] at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:129) [metabase.jar:?] at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:375) [metabase.jar:?] at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:773) [metabase.jar:?] at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:905) [metabase.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: clojure.lang.ExceptionInfo: Error canonicalizing query at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:677) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more Caused by: clojure.lang.ExceptionInfo: Invalid MBQL clause at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:555) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?] at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?] at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?] at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?] at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?] at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?] at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?] at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?] at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?] at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?] at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?] at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more Caused by: clojure.lang.ExceptionInfo: Output of update-field-options does not match schema: [nil nil (named (not ("Invalid :temporal-unit for the specified :base-type." a-clojure.lang.PersistentArrayMap)) "options")] at metabase.mbql.util$fn__20573$update_field_options__20578.doInvoke(util.cljc:581) ~[metabase.jar:?] at clojure.lang.RestFn.applyTo(RestFn.java:142) ~[metabase.jar:?] at clojure.core$apply.invokeStatic(core.clj:669) ~[metabase.jar:?] at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?] at metabase.mbql.util$assoc_field_options.invokeStatic(util.cljc:590) ~[metabase.jar:?] at metabase.mbql.util$assoc_field_options.doInvoke(util.cljc:587) ~[metabase.jar:?] at clojure.lang.RestFn.invoke(RestFn.java:439) ~[metabase.jar:?] at metabase.mbql.util$with_temporal_unit.invokeStatic(util.cljc:595) ~[metabase.jar:?] at metabase.mbql.util$with_temporal_unit.invoke(util.cljc:592) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20861.invokeStatic(normalize.cljc:404) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20861.invoke(normalize.cljc:398) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_implicit_field_id.invokeStatic(normalize.cljc:363) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_implicit_field_id.invoke(normalize.cljc:360) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_simple_filter_clause.invokeStatic(normalize.cljc:455) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_simple_filter_clause.invoke(normalize.cljc:452) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20918$fn__20919.invoke(normalize.cljc:464) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at clojure.core$map$fn__5866.invoke(core.clj:2753) ~[metabase.jar:?] at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?] at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?] at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?] at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?] at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?] at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_compound_filter_clause.invokeStatic(normalize.cljc:425) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_compound_filter_clause.invoke(normalize.cljc:423) ~[metabase.jar:?] at metabase.mbql.normalize$fn__20888$fn__20889.invoke(normalize.cljc:433) ~[metabase.jar:?] at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:552) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?] at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?] at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?] at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?] at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?] at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?] at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?] at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?] at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?] at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?] at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?] at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?] at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?] at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?] at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?] at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?] at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?] at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?] at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?] ... 201 more ``` **To Reproduce** Not sure - have not found exactly how to reproduce it with the Sample Database. A similar question complains about having a time filter on a date field, but no "something went wrong" ``` Unsupported temporal bucketing: You can't bucket a :type/Date Field by :minute. ``` Maybe it is about the combination of metric + expression + saved question + filters? **Expected behavior** The question should load without error, or at least an error that the user can rectify. The "something went wrong" page does not show any of the normal Question UI controls. **Information about your Metabase Installation:** You can get this information by going to Admin -> Troubleshooting. - Your browser and the version: (e.x. Chrome 52.1, Firefox 48.0, Safari 11.1, …) Firefox 88 - Your operating system: (e.x. OS X 10.10, Windows 10.1809, Ubuntu 16.04, …) CentOS 7 - Your databases: (e.x. MySQL, Postgres, MongoDB, …) PostgreSQL 10 - Metabase version: (e.x. 0.19.3) Metabase v0.39.2 - Metabase hosting environment: (e.x. Mac app, Elastic Beanstalk, Docker, Heroku, Jar-file on Ubuntu, …) CentOS 7 - Metabase internal database: (e.x. H2 (default), Postgres or MySQL) PostgreSQL 10 **Severity** Blocking with respect to not being able to view or fix this particular question
process
something went wrong error for card no dataset query returned by api card describe the bug when loading a user created question the generic something went wrong flaming server error page is shown it is not clear why this happened or what the exact repro steps are some notes user interface default language is set to german instance was upgraded from to question created under the get api card endpoint does not include a dataset query object unlike a normal response the question has a join to a saved question as per the question has filters on an expression in the saved question a date field where the filter includes a time the expression is an integer field divided by the count metric calculated in the question if it s not able to be reproduced i would like a pointer on how to fix the question in the backend though i m not sure if the dataset query object looks correct or what should be changed if it s wrong logs the browser firefox console says error typeerror c dataset query is null the metabase database entry for the affected card note leading space select dataset query from report card where id type query query source table card joins alias persons filter minute database when viewing the question the metabase logs say error metabase models interface unable to normalize type query query source table card joins fields all source table condition alias persons filter and minute database clojure lang exceptioninfo error normalizing query at metabase mbql normalize fn fn invoke normalize cljc at metabase models interface maybe normalize invokestatic interface clj at metabase models interface maybe normalize invoke interface clj at metabase models interface catch normalization exceptions fn invoke interface clj at clojure core comp fn invoke core clj at toucan models apply type fns iter fn invoke models clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure core seq invokestatic core clj at clojure core protocols seq reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core into invokestatic core clj at clojure core into invoke core clj at toucan models apply type fns invokestatic models clj at toucan models apply type fns invoke models clj at toucan models do post select invokestatic models clj at toucan models do post select invoke models clj at toucan db do post select iter fn invoke db clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure lang lazilypersistentvector create lazilypersistentvector java at clojure core vec invokestatic core clj at clojure core vec invoke core clj at toucan db do post select invokestatic db clj at toucan db do post select invoke db clj at toucan db simple select invokestatic db clj at toucan db simple select invoke db clj at toucan db simple select one invokestatic db clj at toucan db simple select one invoke db clj at toucan db select one invokestatic db clj at toucan db select one doinvoke db clj at clojure lang restfn applyto restfn java at clojure lang var applyto var java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at toucan models invoke model invokestatic models clj at toucan models invoke model doinvoke models clj at clojure lang restfn invoke restfn java at toucan models invoke model invokestatic models clj at toucan models invoke model invoke models clj at clojure lang afn applytohelper afn java at clojure lang restfn applyto restfn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at toucan models invoke model or instance invokestatic models clj at toucan models invoke model or instance doinvoke models clj at clojure lang restfn invoke restfn java at metabase models card cardinstance invoke card clj at metabase api card fn invokestatic card clj at metabase api card fn invoke card clj at compojure core wrap response fn invoke core clj at compojure core wrap route middleware fn invoke core clj at compojure core wrap route info fn invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware auth enforce authentication fn invoke auth clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at clojure lang afn applytohelper afn java at clojure lang afn applyto afn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at metabase server routes fn fn doinvoke routes clj at clojure lang restfn invoke restfn java at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware exceptions catch uncaught exceptions fn invoke exceptions clj at metabase server middleware exceptions catch api exceptions fn invoke exceptions clj at metabase server middleware log log api call fn fn invoke log clj at toucan db do with call counting invokestatic db clj at toucan db do with call counting invoke db clj at metabase server middleware log log api call fn invoke log clj at metabase server middleware browser cookie ensure browser id cookie fn invoke browser cookie clj at metabase server middleware security add security headers fn invoke security clj at metabase server middleware json wrap json body fn invoke json clj at metabase server middleware json wrap streamed json response fn invoke json clj at ring middleware keyword params wrap keyword params fn invoke keyword params clj at ring middleware params wrap params fn invoke params clj at metabase server middleware misc maybe set site url fn invoke misc clj at metabase server middleware session bind current user fn fn invoke session clj at metabase server middleware session do with current user invokestatic session clj at metabase server middleware session do with current user invoke session clj at metabase server middleware session bind current user fn invoke session clj at metabase server middleware session wrap current user info fn invoke session clj at metabase server middleware session wrap session id fn invoke session clj at metabase server middleware auth wrap api key fn invoke auth clj at ring middleware cookies wrap cookies fn invoke cookies clj at metabase server middleware misc add content type fn invoke misc clj at metabase server middleware misc disable streaming buffering fn invoke misc clj at ring middleware gzip wrap gzip fn invoke gzip clj at metabase server middleware misc bind request fn invoke misc clj at metabase server middleware ssl redirect to https middleware fn invoke ssl clj at metabase server async proxy handler fn invoke server clj at metabase server proxy org eclipse jetty server handler abstracthandler handle unknown source at org eclipse jetty server handler handlerwrapper handle handlerwrapper java at org eclipse jetty server server handle server java at org eclipse jetty server httpchannel lambda handle httpchannel java at org eclipse jetty server httpchannel dispatch httpchannel java at org eclipse jetty server httpchannel handle httpchannel java at org eclipse jetty server httpconnection onfillable httpconnection java at org eclipse jetty io abstractconnection readcallback succeeded abstractconnection java at org eclipse jetty io fillinterest fillable fillinterest java at org eclipse jetty io channelendpoint run channelendpoint java at org eclipse jetty util thread strategy eatwhatyoukill runtask eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill doproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill tryproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill run eatwhatyoukill java at org eclipse jetty util thread reservedthreadexecutor reservedthread run reservedthreadexecutor java at org eclipse jetty util thread queuedthreadpool runjob queuedthreadpool java at org eclipse jetty util thread queuedthreadpool runner run queuedthreadpool java at java lang thread run thread java caused by clojure lang exceptioninfo error canonicalizing query at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more caused by clojure lang exceptioninfo invalid mbql clause at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at medley core map vals fn fn invoke core cljc at clojure core fn fn invoke core clj at clojure core protocols iter reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core fn invokestatic core clj at clojure core fn invoke core clj at clojure core protocols fn g invoke protocols clj at clojure core reduce kv invokestatic core clj at clojure core reduce kv invoke core clj at medley core reduce map invokestatic core cljc at medley core reduce map invoke core cljc at medley core map vals invokestatic core cljc at medley core map vals invoke core cljc at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at clojure core comp fn invoke core clj at clojure core update invokestatic core clj at clojure core update invoke core clj at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more caused by clojure lang exceptioninfo output of update field options does not match schema at metabase mbql util fn update field options doinvoke util cljc at clojure lang restfn applyto restfn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at metabase mbql util assoc field options invokestatic util cljc at metabase mbql util assoc field options doinvoke util cljc at clojure lang restfn invoke restfn java at metabase mbql util with temporal unit invokestatic util cljc at metabase mbql util with temporal unit invoke util cljc at metabase mbql normalize fn invokestatic normalize cljc at metabase mbql normalize fn invoke normalize cljc at clojure lang multifn invoke multifn java at metabase mbql normalize canonicalize implicit field id invokestatic normalize cljc at metabase mbql normalize canonicalize implicit field id invoke normalize cljc at metabase mbql normalize canonicalize simple filter clause invokestatic normalize cljc at metabase mbql normalize canonicalize simple filter clause invoke normalize cljc at metabase mbql normalize fn fn invoke normalize cljc at clojure lang multifn invoke multifn java at clojure core map fn invoke core clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure core seq invokestatic core clj at clojure core protocols seq reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core into invokestatic core clj at clojure core into invoke core clj at metabase mbql normalize canonicalize compound filter clause invokestatic normalize cljc at metabase mbql normalize canonicalize compound filter clause invoke normalize cljc at metabase mbql normalize fn fn invoke normalize cljc at clojure lang multifn invoke multifn java at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at medley core map vals fn fn invoke core cljc at clojure core fn fn invoke core clj at clojure core protocols iter reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core fn invokestatic core clj at clojure core fn invoke core clj at clojure core protocols fn g invoke protocols clj at clojure core reduce kv invokestatic core clj at clojure core reduce kv invoke core clj at medley core reduce map invokestatic core cljc at medley core reduce map invoke core cljc at medley core map vals invokestatic core cljc at medley core map vals invoke core cljc at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at clojure core comp fn invoke core clj at clojure core update invokestatic core clj at clojure core update invoke core clj at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more to reproduce not sure have not found exactly how to reproduce it with the sample database a similar question complains about having a time filter on a date field but no something went wrong unsupported temporal bucketing you can t bucket a type date field by minute maybe it is about the combination of metric expression saved question filters expected behavior the question should load without error or at least an error that the user can rectify the something went wrong page does not show any of the normal question ui controls information about your metabase installation you can get this information by going to admin troubleshooting your browser and the version e x chrome firefox safari … firefox your operating system e x os x windows ubuntu … centos your databases e x mysql postgres mongodb … postgresql metabase version e x metabase metabase hosting environment e x mac app elastic beanstalk docker heroku jar file on ubuntu … centos metabase internal database e x default postgres or mysql postgresql severity blocking with respect to not being able to view or fix this particular question
1
17,253
23,035,290,332
IssuesEvent
2022-07-22 17:59:54
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
[BUG]: Combination of PCA and remove_multicollinearity is causing error
bug preprocessing
### pycaret version checks - [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues). - [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret. - [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master). ### Issue Description Not possible to use remove_multicollinearity and pca in combination in setup. ### Reproducible Example ```python from pycaret.datasets import get_data from pycaret.classification import * _data = get_data('satellite') exp = ClassificationExperiment() exp.setup(data = _data, remove_multicollinearity = True, multicollinearity_threshold = 0.9, pca_components= 1.0, pca = True, target = 'Class') ``` ### Expected Behavior Since the nr of pca components is set to 1.0, you would expect 100% of the columns that are left after all the other feature selection methods to be kept to the same nr. ### Actual Results ```python-traceback Instead, this is throwing the error: 497 raise ValueError( 498 "n_components='mle' is only supported if n_samples >= n_features" 499 ) 500 elif not 0 <= n_components <= min(n_samples, n_features): --> 501 raise ValueError( 502 "n_components=%r must be between 0 and " 503 "min(n_samples, n_features)=%r with " 504 "svd_solver='full'" % (n_components, min(n_samples, n_features)) 505 ) 506 elif n_components >= 1: 507 if not isinstance(n_components, numbers.Integral): ValueError: n_components=36 must be between 0 and min(n_samples, n_features)=4 with svd_solver='full' ``` ### Installed Versions <details> '3.0.0.rc2' </details>
1.0
[BUG]: Combination of PCA and remove_multicollinearity is causing error - ### pycaret version checks - [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues). - [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret. - [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master). ### Issue Description Not possible to use remove_multicollinearity and pca in combination in setup. ### Reproducible Example ```python from pycaret.datasets import get_data from pycaret.classification import * _data = get_data('satellite') exp = ClassificationExperiment() exp.setup(data = _data, remove_multicollinearity = True, multicollinearity_threshold = 0.9, pca_components= 1.0, pca = True, target = 'Class') ``` ### Expected Behavior Since the nr of pca components is set to 1.0, you would expect 100% of the columns that are left after all the other feature selection methods to be kept to the same nr. ### Actual Results ```python-traceback Instead, this is throwing the error: 497 raise ValueError( 498 "n_components='mle' is only supported if n_samples >= n_features" 499 ) 500 elif not 0 <= n_components <= min(n_samples, n_features): --> 501 raise ValueError( 502 "n_components=%r must be between 0 and " 503 "min(n_samples, n_features)=%r with " 504 "svd_solver='full'" % (n_components, min(n_samples, n_features)) 505 ) 506 elif n_components >= 1: 507 if not isinstance(n_components, numbers.Integral): ValueError: n_components=36 must be between 0 and min(n_samples, n_features)=4 with svd_solver='full' ``` ### Installed Versions <details> '3.0.0.rc2' </details>
process
combination of pca and remove multicollinearity is causing error pycaret version checks i have checked that this issue has not already been reported i have confirmed this bug exists on the of pycaret i have confirmed this bug exists on the master branch of pycaret pip install u git issue description not possible to use remove multicollinearity and pca in combination in setup reproducible example python from pycaret datasets import get data from pycaret classification import data get data satellite exp classificationexperiment exp setup data data remove multicollinearity true multicollinearity threshold pca components pca true target class expected behavior since the nr of pca components is set to you would expect of the columns that are left after all the other feature selection methods to be kept to the same nr actual results python traceback instead this is throwing the error raise valueerror n components mle is only supported if n samples n features elif not n components min n samples n features raise valueerror n components r must be between and min n samples n features r with svd solver full n components min n samples n features elif n components if not isinstance n components numbers integral valueerror n components must be between and min n samples n features with svd solver full installed versions
1
17,525
23,336,236,622
IssuesEvent
2022-08-09 10:11:01
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Diagnostics.Process work planned for .NET 7
Epic area-System.Diagnostics.Process Priority:3 Team:Libraries
**This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope.** ## Summary We are not planning any notable investments into the System.Diagnostics.Process area in .NET 7. We will address high-impact issues and resolve test issues. We will consider small community contributions that improve cross-platform compatibility. ## Planned for .NET 7 - [x] #45017 - [x] #63937 - [x] #44453 ## Backlog (in roughly priority order) - [ ] #58492 - [ ] #53095 - [ ] #49107 - [ ] #45685
1.0
System.Diagnostics.Process work planned for .NET 7 - **This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope.** ## Summary We are not planning any notable investments into the System.Diagnostics.Process area in .NET 7. We will address high-impact issues and resolve test issues. We will consider small community contributions that improve cross-platform compatibility. ## Planned for .NET 7 - [x] #45017 - [x] #63937 - [x] #44453 ## Backlog (in roughly priority order) - [ ] #58492 - [ ] #53095 - [ ] #49107 - [ ] #45685
process
system diagnostics process work planned for net this issue captures the planned work for net this list is expected to change throughout the release cycle according to ongoing planning and discussions with possible additions and subtractions to the scope summary we are not planning any notable investments into the system diagnostics process area in net we will address high impact issues and resolve test issues we will consider small community contributions that improve cross platform compatibility planned for net backlog in roughly priority order
1
17,056
22,475,071,639
IssuesEvent
2022-06-22 11:33:56
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Update Go client to use 1.17
kind/toil scope/clients-go team/process-automation scope/zbctl area/maintainability
**Description** Our Go client is now built against 1.15, which is not supported anymore. We should start building against Go 1.17, which is the lowest support version.
1.0
Update Go client to use 1.17 - **Description** Our Go client is now built against 1.15, which is not supported anymore. We should start building against Go 1.17, which is the lowest support version.
process
update go client to use description our go client is now built against which is not supported anymore we should start building against go which is the lowest support version
1
22,096
30,615,780,034
IssuesEvent
2023-07-24 02:53:37
solop-develop/frontend-core
https://api.github.com/repos/solop-develop/frontend-core
closed
[Bug Report] Reporte/Proceso: Parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores
bug (PRC) Processes (RPT) Reports (ULP) UI Lookups (UX) User Experience
<!-- Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed. --> ## Bug report #### Steps to reproduce 1. Abrir el reporte de `Saldos Abiertos`. 2. Mostrar los campos de fecha (todos estΓ‘n marcados como rango). 3. Establecer los valores en las fechas. #### Screenshot or Gif(ζˆͺε›Ύζˆ–εŠ¨ζ€ε›ΎοΌ‰ https://github.com/solop-develop/frontend-core/assets/20288327/ff36d84e-a97f-40c5-9c83-1b8d39e843ac #### Expected behavior Observe como no se puede establecer el valor desde y hasta, solo se establece un valor por campo.
1.0
[Bug Report] Reporte/Proceso: Parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores - <!-- Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed. --> ## Bug report #### Steps to reproduce 1. Abrir el reporte de `Saldos Abiertos`. 2. Mostrar los campos de fecha (todos estΓ‘n marcados como rango). 3. Establecer los valores en las fechas. #### Screenshot or Gif(ζˆͺε›Ύζˆ–εŠ¨ζ€ε›ΎοΌ‰ https://github.com/solop-develop/frontend-core/assets/20288327/ff36d84e-a97f-40c5-9c83-1b8d39e843ac #### Expected behavior Observe como no se puede establecer el valor desde y hasta, solo se establece un valor por campo.
process
reporte proceso parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores note in order to better solve your problem please refer to the template to provide complete information accurately describe the problem and the incomplete information issue will be closed bug report steps to reproduce abrir el reporte de saldos abiertos mostrar los campos de fecha todos estΓ‘n marcados como rango establecer los valores en las fechas screenshot or gif(ζˆͺε›Ύζˆ–εŠ¨ζ€ε›ΎοΌ‰ expected behavior observe como no se puede establecer el valor desde y hasta solo se establece un valor por campo
1
14,623
17,766,788,285
IssuesEvent
2021-08-30 08:33:58
googleapis/nodejs-pubsub
https://api.github.com/repos/googleapis/nodejs-pubsub
closed
Dependency Dashboard
type: process api: pubsub
This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/1123) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/1302) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[chore(deps): update dependency yargs to v17](../pull/1277) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/1273) - [ ] <!-- recreate-branch=renovate/p-defer-4.x -->[fix(deps): update dependency p-defer to v4](../pull/1250) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/1123) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/1302) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[chore(deps): update dependency yargs to v17](../pull/1277) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/1273) - [ ] <!-- recreate-branch=renovate/p-defer-4.x -->[fix(deps): update dependency p-defer to v4](../pull/1250) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull mocha types mocha pull pull pull check this box to trigger a request for renovate to run again on this repository
1
10,437
13,220,071,105
IssuesEvent
2020-08-17 11:43:14
km4ack/pi-build
https://api.github.com/repos/km4ack/pi-build
closed
Check for location of GPS in script
in process
My GPS is located at "ttyUSB0" not the default of the script. Could the script check for the location? ls /dev/*USB* and use that to install the GPS. My GPS is a GlobalSat BU-353-S4 USB
1.0
Check for location of GPS in script - My GPS is located at "ttyUSB0" not the default of the script. Could the script check for the location? ls /dev/*USB* and use that to install the GPS. My GPS is a GlobalSat BU-353-S4 USB
process
check for location of gps in script my gps is located at not the default of the script could the script check for the location ls dev usb and use that to install the gps my gps is a globalsat bu usb
1
683,540
23,385,858,647
IssuesEvent
2022-08-11 13:42:35
TheYellowArchitect/doubledamnation
https://api.github.com/repos/TheYellowArchitect/doubledamnation
opened
Level Skip - Death Dialogue
bug needs more info low priority netcoding dialogue
If host is having a dialogue after death, and client joins right there, client cannot change state, and is on a permanently "Idle" state, regardless of movement. I have not investigated what causes this. Inputs do pass to host (he can magejump properly) but probably disabled locally (as if in cutscene) Regardless, it is low priority because its easily fixed by simply dying/reviving.
1.0
Level Skip - Death Dialogue - If host is having a dialogue after death, and client joins right there, client cannot change state, and is on a permanently "Idle" state, regardless of movement. I have not investigated what causes this. Inputs do pass to host (he can magejump properly) but probably disabled locally (as if in cutscene) Regardless, it is low priority because its easily fixed by simply dying/reviving.
non_process
level skip death dialogue if host is having a dialogue after death and client joins right there client cannot change state and is on a permanently idle state regardless of movement i have not investigated what causes this inputs do pass to host he can magejump properly but probably disabled locally as if in cutscene regardless it is low priority because its easily fixed by simply dying reviving
0
863
3,322,369,443
IssuesEvent
2015-11-09 14:13:40
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
opened
Keyref link breaks if target topic contains keyref
bug preprocess/keyref
`develop` branch. [Sample files](https://github.com/eerohele/dita-ot-issues/tree/master/fixtures/2133). Easier to show than tell: the first link in `topic1.dita` is broken, the second one works. If you remove `<keyword keyref="keyword"/>` from `topic2.dita`, the first link starts working. It looks like the `keyref` stage creates a file called `topic2-1.dita` in `temp`. It is an exact clone of `topic2.dita`. The subsequent preprocessing steps then use `topic2-1.dita` instead of `topic2.dita`. Didn't get further than that yet.
1.0
Keyref link breaks if target topic contains keyref - `develop` branch. [Sample files](https://github.com/eerohele/dita-ot-issues/tree/master/fixtures/2133). Easier to show than tell: the first link in `topic1.dita` is broken, the second one works. If you remove `<keyword keyref="keyword"/>` from `topic2.dita`, the first link starts working. It looks like the `keyref` stage creates a file called `topic2-1.dita` in `temp`. It is an exact clone of `topic2.dita`. The subsequent preprocessing steps then use `topic2-1.dita` instead of `topic2.dita`. Didn't get further than that yet.
process
keyref link breaks if target topic contains keyref develop branch easier to show than tell the first link in dita is broken the second one works if you remove from dita the first link starts working it looks like the keyref stage creates a file called dita in temp it is an exact clone of dita the subsequent preprocessing steps then use dita instead of dita didn t get further than that yet
1
13,282
15,761,497,271
IssuesEvent
2021-03-31 10:02:48
nodejs/node
https://api.github.com/repos/nodejs/node
opened
child_process 'close' always (?) emitted after 'exit'
child_process
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows) Subsystem: if known, please specify affected core module name --> * **Version**: master * **Platform**: Linux 5.4.0-70-generic #78-Ubuntu SMP Fri Mar 19 13:29:52 UTC 2021 x86_64 x86_64 x86_64 GNU/Linuxd * **Subsystem**: child_process ### What steps will reproduce the bug? ```js const child_process = require('child_process'); const proc = child_process.spawn('bash', ['-c', 'exec 0>&- 1>&- 2>&-; sleep 5'], { stdio: ['inherit', 'pipe', 'inherit'] }); proc.stdout.pipe(process.stdout); proc.on('exit', () => console.log('exit')); proc.on('close', () => console.log('close')); ``` ### How often does it reproduce? Is there a required condition? Always ### What is the expected behavior? 'close' should be emitted immediately (all stdio streams are closed), 'exit' after 5 seconds ### What do you see instead? 5 seconds pass, then 'exit' is emitted and then 'closed'
1.0
child_process 'close' always (?) emitted after 'exit' - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows) Subsystem: if known, please specify affected core module name --> * **Version**: master * **Platform**: Linux 5.4.0-70-generic #78-Ubuntu SMP Fri Mar 19 13:29:52 UTC 2021 x86_64 x86_64 x86_64 GNU/Linuxd * **Subsystem**: child_process ### What steps will reproduce the bug? ```js const child_process = require('child_process'); const proc = child_process.spawn('bash', ['-c', 'exec 0>&- 1>&- 2>&-; sleep 5'], { stdio: ['inherit', 'pipe', 'inherit'] }); proc.stdout.pipe(process.stdout); proc.on('exit', () => console.log('exit')); proc.on('close', () => console.log('close')); ``` ### How often does it reproduce? Is there a required condition? Always ### What is the expected behavior? 'close' should be emitted immediately (all stdio streams are closed), 'exit' after 5 seconds ### What do you see instead? 5 seconds pass, then 'exit' is emitted and then 'closed'
process
child process close always emitted after exit thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or output of osversion foreach object versionstring if else in powershell console windows subsystem if known please specify affected core module name version master platform linux generic ubuntu smp fri mar utc gnu linuxd subsystem child process what steps will reproduce the bug js const child process require child process const proc child process spawn bash stdio proc stdout pipe process stdout proc on exit console log exit proc on close console log close how often does it reproduce is there a required condition always what is the expected behavior close should be emitted immediately all stdio streams are closed exit after seconds what do you see instead seconds pass then exit is emitted and then closed
1
406,108
27,551,774,582
IssuesEvent
2023-03-07 15:22:19
MentorMate/node-project-cli
https://api.github.com/repos/MentorMate/node-project-cli
closed
Setup swagger generation for express apps
documentation express
# Description The application should be able to produce a `swagger.json` file documenting its endpoints # Tasks - [x] integrate a `zod-to-openapi` package for generating OpenAPI definitions from zod schemas - [ ] create a script and/or an endpoint that will produce the [OpenAPI document](https://swagger.io/specification/)
1.0
Setup swagger generation for express apps - # Description The application should be able to produce a `swagger.json` file documenting its endpoints # Tasks - [x] integrate a `zod-to-openapi` package for generating OpenAPI definitions from zod schemas - [ ] create a script and/or an endpoint that will produce the [OpenAPI document](https://swagger.io/specification/)
non_process
setup swagger generation for express apps description the application should be able to produce a swagger json file documenting its endpoints tasks integrate a zod to openapi package for generating openapi definitions from zod schemas create a script and or an endpoint that will produce the
0
4,529
7,371,636,803
IssuesEvent
2018-03-13 12:28:26
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Broken Link
cxp doc-bug in-process media-services triaged
The link under https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download That is linked to: http://southworks.com/blog/2015/05/27/reusing-azure-media-services-locators-to-avoid-facing-the-5-shared-access-policy-limitation/ ... is broken. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ca2f89cc-d993-eed3-ebf0-c7bfaa7716e0 * Version Independent ID: fd773e3d-6581-c356-ddd0-f9b609d63d7a * Content: [Get started with delivering content on demand using .NET | Microsoft Docs](https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download) * Content Source: [articles/media-services/media-services-dotnet-get-started.md](https://github.com/Microsoft/azure-docs/blob/master/articles/media-services/media-services-dotnet-get-started.md) * Service: **media-services** * GitHub Login: @Juliako * Microsoft Alias: **juliako**
1.0
Broken Link - The link under https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download That is linked to: http://southworks.com/blog/2015/05/27/reusing-azure-media-services-locators-to-avoid-facing-the-5-shared-access-policy-limitation/ ... is broken. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ca2f89cc-d993-eed3-ebf0-c7bfaa7716e0 * Version Independent ID: fd773e3d-6581-c356-ddd0-f9b609d63d7a * Content: [Get started with delivering content on demand using .NET | Microsoft Docs](https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download) * Content Source: [articles/media-services/media-services-dotnet-get-started.md](https://github.com/Microsoft/azure-docs/blob/master/articles/media-services/media-services-dotnet-get-started.md) * Service: **media-services** * GitHub Login: @Juliako * Microsoft Alias: **juliako**
process
broken link the link under that is linked to is broken document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service media services github login juliako microsoft alias juliako
1
11,382
14,222,882,162
IssuesEvent
2020-11-17 17:26:57
JuliaParallel/Dagger.jl
https://api.github.com/repos/JuliaParallel/Dagger.jl
opened
Multithread and distribute the scheduler
performance processors scheduler
The scheduler is quite close to being able to run multithreaded, with many parts running concurrently already. We're also already pushing some extra work onto the worker nodes (mainly argument fetching and processor selection/load balancing), and it would be beneficial for large DAGs to move more work onto each worker. The main blocker is providing a way to split the DAG into multiple domains, where each domain is handled by a given thread on a given worker. With efficient Thunk serialization, we can then send a subgraph to each worker and let them process their own DAG without conflicts. We'll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data; if possible, we can also have workers broadcast and shard their `Chunk`s onto dependent workers as soon as the data is made available.
1.0
Multithread and distribute the scheduler - The scheduler is quite close to being able to run multithreaded, with many parts running concurrently already. We're also already pushing some extra work onto the worker nodes (mainly argument fetching and processor selection/load balancing), and it would be beneficial for large DAGs to move more work onto each worker. The main blocker is providing a way to split the DAG into multiple domains, where each domain is handled by a given thread on a given worker. With efficient Thunk serialization, we can then send a subgraph to each worker and let them process their own DAG without conflicts. We'll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data; if possible, we can also have workers broadcast and shard their `Chunk`s onto dependent workers as soon as the data is made available.
process
multithread and distribute the scheduler the scheduler is quite close to being able to run multithreaded with many parts running concurrently already we re also already pushing some extra work onto the worker nodes mainly argument fetching and processor selection load balancing and it would be beneficial for large dags to move more work onto each worker the main blocker is providing a way to split the dag into multiple domains where each domain is handled by a given thread on a given worker with efficient thunk serialization we can then send a subgraph to each worker and let them process their own dag without conflicts we ll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data if possible we can also have workers broadcast and shard their chunk s onto dependent workers as soon as the data is made available
1
148,686
13,244,199,643
IssuesEvent
2020-08-19 12:42:26
OraOpenSource/apex-nitro
https://api.github.com/repos/OraOpenSource/apex-nitro
closed
minor: Upate warning message for openApp
documentation enhancement
When running in headless mode the following warning message is shown: `[APEX Nitro] Couldn't open browser (if you are using BrowserSync in a headless environment, you might want to set the open option to false)` The config file has this option as `launch.openApp`. Perhaps update the message with that instead of just `open`.
1.0
minor: Upate warning message for openApp - When running in headless mode the following warning message is shown: `[APEX Nitro] Couldn't open browser (if you are using BrowserSync in a headless environment, you might want to set the open option to false)` The config file has this option as `launch.openApp`. Perhaps update the message with that instead of just `open`.
non_process
minor upate warning message for openapp when running in headless mode the following warning message is shown couldn t open browser if you are using browsersync in a headless environment you might want to set the open option to false the config file has this option as launch openapp perhaps update the message with that instead of just open
0
345,815
10,373,588,218
IssuesEvent
2019-09-09 07:40:07
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
use the Version Stream to manage versions of client side packages l
area/versions lifecycle/stale priority/important-soon
we should lock down client side binaries like `helm, git, kubectl` so that if folks have ancient or mismatched versions we warn/recommend upgrading/downgrading / fail https://jenkins-x.io/architecture/version-stream/
1.0
use the Version Stream to manage versions of client side packages l - we should lock down client side binaries like `helm, git, kubectl` so that if folks have ancient or mismatched versions we warn/recommend upgrading/downgrading / fail https://jenkins-x.io/architecture/version-stream/
non_process
use the version stream to manage versions of client side packages l we should lock down client side binaries like helm git kubectl so that if folks have ancient or mismatched versions we warn recommend upgrading downgrading fail
0
5,480
8,355,650,927
IssuesEvent
2018-10-02 16:13:39
pelias/acceptance-tests
https://api.github.com/repos/pelias/acceptance-tests
closed
Add acceptance tests for Eraser Map usage
processed
- [x] [`/autocomplete + focus.point + text`](https://search.mapzen.com/v1/autocomplete?text=wyt&focus.point.lat=40.744569&focus.point.lon=-73.9902313&api_key=search-BcgTWes) - [ ] [`/search + focus.viewport + text`](https://search.mapzen.com/v1/search?text=wyth&focus.viewport.min_lat=40.73644229295322&focus.viewport.min_lon=-73.99405878124092&focus.viewport.max_lat=40.74684717132519&focus.viewport.max_lon=-73.9863340187591&api_key=search-BcgTWes) - [ ] [`/reverse + point`](https://search.mapzen.com/v1/reverse?point.lat=40.74549540292336&point.lon=-73.98920473153952&api_key=search-BcgTWes) - [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=36.57751244954497&point.lon=-68.0269221649082&api_key=search-BcgTWes) - [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=31.181260064156664&point.lon=-35.36817462744051&api_key=search-BcgTWes) - [ ] test for privacy headers: maybe check that there is a `private` flag sent back as part of the response?
1.0
Add acceptance tests for Eraser Map usage - - [x] [`/autocomplete + focus.point + text`](https://search.mapzen.com/v1/autocomplete?text=wyt&focus.point.lat=40.744569&focus.point.lon=-73.9902313&api_key=search-BcgTWes) - [ ] [`/search + focus.viewport + text`](https://search.mapzen.com/v1/search?text=wyth&focus.viewport.min_lat=40.73644229295322&focus.viewport.min_lon=-73.99405878124092&focus.viewport.max_lat=40.74684717132519&focus.viewport.max_lon=-73.9863340187591&api_key=search-BcgTWes) - [ ] [`/reverse + point`](https://search.mapzen.com/v1/reverse?point.lat=40.74549540292336&point.lon=-73.98920473153952&api_key=search-BcgTWes) - [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=36.57751244954497&point.lon=-68.0269221649082&api_key=search-BcgTWes) - [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=31.181260064156664&point.lon=-35.36817462744051&api_key=search-BcgTWes) - [ ] test for privacy headers: maybe check that there is a `private` flag sent back as part of the response?
process
add acceptance tests for eraser map usage test for privacy headers maybe check that there is a private flag sent back as part of the response
1
91,900
3,863,516,893
IssuesEvent
2016-04-08 09:45:42
iamxavier/elmah
https://api.github.com/repos/iamxavier/elmah
closed
Use of C#-SQLite library
auto-migrated Priority-Medium Type-Enhancement
``` What new or enhanced feature are you proposing? Using the csharp-sqlite library instead of the existing library. C#-Sqlite is written in C# and is 100% compatible with SQLite (according to the unit tests). It's SQLite written in C#. http://code.google.com/p/csharp-sqlite/ I'd be more than happy to make this change. What goal would this enhancement help you achieve? It removes the dependence on a 32-bit or 64-bit SQLite DLL. After all, isn't platform independence one of the reasons we use C#? Dealing with the 32/64 bit issue is a pain when deploying across that barrier. The downside is perhaps the code will go stale. This isn't much of an issue since this is a risk even with the existing library. In fact, the existing one did go dormant until SQLite took over maintenance. There was even talk about it using the C#-SQLite library instead of the SQLite DLLs. ``` Original issue reported on code.google.com by `masil...@gmail.com` on 5 Dec 2011 at 5:44
1.0
Use of C#-SQLite library - ``` What new or enhanced feature are you proposing? Using the csharp-sqlite library instead of the existing library. C#-Sqlite is written in C# and is 100% compatible with SQLite (according to the unit tests). It's SQLite written in C#. http://code.google.com/p/csharp-sqlite/ I'd be more than happy to make this change. What goal would this enhancement help you achieve? It removes the dependence on a 32-bit or 64-bit SQLite DLL. After all, isn't platform independence one of the reasons we use C#? Dealing with the 32/64 bit issue is a pain when deploying across that barrier. The downside is perhaps the code will go stale. This isn't much of an issue since this is a risk even with the existing library. In fact, the existing one did go dormant until SQLite took over maintenance. There was even talk about it using the C#-SQLite library instead of the SQLite DLLs. ``` Original issue reported on code.google.com by `masil...@gmail.com` on 5 Dec 2011 at 5:44
non_process
use of c sqlite library what new or enhanced feature are you proposing using the csharp sqlite library instead of the existing library c sqlite is written in c and is compatible with sqlite according to the unit tests it s sqlite written in c i d be more than happy to make this change what goal would this enhancement help you achieve it removes the dependence on a bit or bit sqlite dll after all isn t platform independence one of the reasons we use c dealing with the bit issue is a pain when deploying across that barrier the downside is perhaps the code will go stale this isn t much of an issue since this is a risk even with the existing library in fact the existing one did go dormant until sqlite took over maintenance there was even talk about it using the c sqlite library instead of the sqlite dlls original issue reported on code google com by masil gmail com on dec at
0
444,134
12,806,862,475
IssuesEvent
2020-07-03 10:15:44
enso-org/enso
https://api.github.com/repos/enso-org/enso
closed
Implement Undo
Category: Tooling Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement
### Summary With the design for undo specified as part of luna/enso#551, we now need to implement the underlying functionality. ### Value Clients will be able to undo edits. ### Specification - [ ] Implement the underlying functionality for tracking user action stacks as specified in the [design document](https://github.com/luna/enso/blob/master/doc/design/engine/engine-services.md). - [ ] Implement the `text/undo` message as specified in the same design document. - [ ] Document any errors that are added as part of this proposal in the design document. ### Acceptance Criteria & Test Cases - The undo functionality has been implemented. - The undo functionality has been rigorously tested.
1.0
Implement Undo - ### Summary With the design for undo specified as part of luna/enso#551, we now need to implement the underlying functionality. ### Value Clients will be able to undo edits. ### Specification - [ ] Implement the underlying functionality for tracking user action stacks as specified in the [design document](https://github.com/luna/enso/blob/master/doc/design/engine/engine-services.md). - [ ] Implement the `text/undo` message as specified in the same design document. - [ ] Document any errors that are added as part of this proposal in the design document. ### Acceptance Criteria & Test Cases - The undo functionality has been implemented. - The undo functionality has been rigorously tested.
non_process
implement undo summary with the design for undo specified as part of luna enso we now need to implement the underlying functionality value clients will be able to undo edits specification implement the underlying functionality for tracking user action stacks as specified in the implement the text undo message as specified in the same design document document any errors that are added as part of this proposal in the design document acceptance criteria test cases the undo functionality has been implemented the undo functionality has been rigorously tested
0
17,059
22,492,218,193
IssuesEvent
2022-06-23 03:13:07
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Set M Value processing tool does not behave as expected
Feedback stale Processing Bug
### What is the bug or the crash? The [documentation and help](https://docs.qgis.org/3.22/en/docs/user_manual/processing_algs/qgis/vectorgeometry.html#qgissetmvalue) for the Set M Value processing tool states that "If no M values exist, the geometry will be upgraded to include M values and the specified value used as the initial M value for all geometries". Existing features have their M value updated according to the value set in the tool. New features created default to a NULL M value regardless of what is set using the processing tool. I've tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool. ### Steps to reproduce the issue Go to processing toolbox > set m values. Choose the input layer, set the m value to a number. Click Run. With the created temporary layer begin editing and add a new feature. Save your edits and stop editing. Again using the processing toolbox find the tool "Extract M Values". Set the summaries to first, last and count and extract the M values. Open the attribute table of the resultant layer. Newly added features through editing will have an M Value of NULL. Features existing prior to running the set M values tool will have the M values as set using the set M values tools. ### Versions 3.16.9 3.22.3 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context It may be that this is the desired behaviour in which case I would suggest an update to the documentation to make this clear. I'd be happy to make said update.
1.0
Set M Value processing tool does not behave as expected - ### What is the bug or the crash? The [documentation and help](https://docs.qgis.org/3.22/en/docs/user_manual/processing_algs/qgis/vectorgeometry.html#qgissetmvalue) for the Set M Value processing tool states that "If no M values exist, the geometry will be upgraded to include M values and the specified value used as the initial M value for all geometries". Existing features have their M value updated according to the value set in the tool. New features created default to a NULL M value regardless of what is set using the processing tool. I've tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool. ### Steps to reproduce the issue Go to processing toolbox > set m values. Choose the input layer, set the m value to a number. Click Run. With the created temporary layer begin editing and add a new feature. Save your edits and stop editing. Again using the processing toolbox find the tool "Extract M Values". Set the summaries to first, last and count and extract the M values. Open the attribute table of the resultant layer. Newly added features through editing will have an M Value of NULL. Features existing prior to running the set M values tool will have the M values as set using the set M values tools. ### Versions 3.16.9 3.22.3 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context It may be that this is the desired behaviour in which case I would suggest an update to the documentation to make this clear. I'd be happy to make said update.
process
set m value processing tool does not behave as expected what is the bug or the crash the for the set m value processing tool states that if no m values exist the geometry will be upgraded to include m values and the specified value used as the initial m value for all geometries existing features have their m value updated according to the value set in the tool new features created default to a null m value regardless of what is set using the processing tool i ve tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool steps to reproduce the issue go to processing toolbox set m values choose the input layer set the m value to a number click run with the created temporary layer begin editing and add a new feature save your edits and stop editing again using the processing toolbox find the tool extract m values set the summaries to first last and count and extract the m values open the attribute table of the resultant layer newly added features through editing will have an m value of null features existing prior to running the set m values tool will have the m values as set using the set m values tools versions supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context it may be that this is the desired behaviour in which case i would suggest an update to the documentation to make this clear i d be happy to make said update
1
12,966
15,342,672,656
IssuesEvent
2021-02-27 17:09:56
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
opened
Filter Challenge Type is displayed as Type
P4 ShapeupProcess challenge- recommender-tool
Filter `Challenge Type` is displayed as `Type` <img width="1440" alt="Screenshot 2021-02-27 at 10 36 47 PM" src="https://user-images.githubusercontent.com/58783823/109394368-a228f280-794c-11eb-812e-6b46fcb16617.png">
1.0
Filter Challenge Type is displayed as Type - Filter `Challenge Type` is displayed as `Type` <img width="1440" alt="Screenshot 2021-02-27 at 10 36 47 PM" src="https://user-images.githubusercontent.com/58783823/109394368-a228f280-794c-11eb-812e-6b46fcb16617.png">
process
filter challenge type is displayed as type filter challenge type is displayed as type img width alt screenshot at pm src
1
108,689
16,812,562,592
IssuesEvent
2021-06-17 01:01:58
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
opened
CVE-2021-0512 (High) detected in linuxv5.2
security vulnerability
## CVE-2021-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was found in android kernel common. The HID subsystem allows an "HID report field" to have a different number of "values" and "usages" when it is allocated. When a field struct is created, the size of the usage array is guaranteed to be at least as large as the values array, but it may be larger. This leads to a potential out-of-bounds write in. <p>Publish Date: 2020-11-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0512>CVE-2021-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-06-01">https://source.android.com/security/bulletin/2021-06-01</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: ASB-2021-02-05_mainline</p> </p> </details> <p></p>
True
CVE-2021-0512 (High) detected in linuxv5.2 - ## CVE-2021-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was found in android kernel common. The HID subsystem allows an "HID report field" to have a different number of "values" and "usages" when it is allocated. When a field struct is created, the size of the usage array is guaranteed to be at least as large as the values array, but it may be larger. This leads to a potential out-of-bounds write in. <p>Publish Date: 2020-11-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0512>CVE-2021-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-06-01">https://source.android.com/security/bulletin/2021-06-01</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: ASB-2021-02-05_mainline</p> </p> </details> <p></p>
non_process
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files vulnerability details an issue was found in android kernel common the hid subsystem allows an hid report field to have a different number of values and usages when it is allocated when a field struct is created the size of the usage array is guaranteed to be at least as large as the values array but it may be larger this leads to a potential out of bounds write in publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution asb mainline
0
16,531
21,558,572,967
IssuesEvent
2022-04-30 21:01:35
knative/serving
https://api.github.com/repos/knative/serving
closed
Auto TLS Beta
kind/feature area/networking kind/process triage/accepted
<!-- If you need to report a security issue with Knative, send an email to knative-security@googlegroups.com. --> <!-- ## In what area(s)? Remove the '> ' to select > /area API > /area autoscale > /area build > /area monitoring /area networking > /area test-and-release Other classifications: > /kind good-first-issue /kind process > /kind spec --> ## Describe the feature This is the issue to tracking the work for Auto TLS Beta. Below are the work items I think we need for Auto TLS Beta: 1. Support cert-manager 0.11 and above with their v1alpha2 APIs https://github.com/knative/serving/issues/6011 2. E2E tests for Auto TLS feature https://github.com/knative/serving/issues/4066, including testing the following cases: - Certificate provision per Knative Service - Certificate provision per namespace - Certificate provision with HTTP01 challenge 3. Conformance test for Certificate: https://github.com/knative/serving/issues/6445 4. Detailed documentation about Auto TLS feature and the modes it supports (DNS challenge/HTTP challenge, per ksvc/per namespace). Related issue: https://github.com/knative/docs/issues/1949 Feel free to comment in this thread if anyone has thoughts about the work items needed for Beta launch. /cc @mattmoor @tcnghia
1.0
Auto TLS Beta - <!-- If you need to report a security issue with Knative, send an email to knative-security@googlegroups.com. --> <!-- ## In what area(s)? Remove the '> ' to select > /area API > /area autoscale > /area build > /area monitoring /area networking > /area test-and-release Other classifications: > /kind good-first-issue /kind process > /kind spec --> ## Describe the feature This is the issue to tracking the work for Auto TLS Beta. Below are the work items I think we need for Auto TLS Beta: 1. Support cert-manager 0.11 and above with their v1alpha2 APIs https://github.com/knative/serving/issues/6011 2. E2E tests for Auto TLS feature https://github.com/knative/serving/issues/4066, including testing the following cases: - Certificate provision per Knative Service - Certificate provision per namespace - Certificate provision with HTTP01 challenge 3. Conformance test for Certificate: https://github.com/knative/serving/issues/6445 4. Detailed documentation about Auto TLS feature and the modes it supports (DNS challenge/HTTP challenge, per ksvc/per namespace). Related issue: https://github.com/knative/docs/issues/1949 Feel free to comment in this thread if anyone has thoughts about the work items needed for Beta launch. /cc @mattmoor @tcnghia
process
auto tls beta in what area s remove the to select area api area autoscale area build area monitoring area networking area test and release other classifications kind good first issue kind process kind spec describe the feature this is the issue to tracking the work for auto tls beta below are the work items i think we need for auto tls beta support cert manager and above with their apis tests for auto tls feature including testing the following cases certificate provision per knative service certificate provision per namespace certificate provision with challenge conformance test for certificate detailed documentation about auto tls feature and the modes it supports dns challenge http challenge per ksvc per namespace related issue feel free to comment in this thread if anyone has thoughts about the work items needed for beta launch cc mattmoor tcnghia
1
8,494
2,993,742,531
IssuesEvent
2015-07-22 07:04:29
sunlightlabs/the-phantom-mask
https://api.github.com/repos/sunlightlabs/the-phantom-mask
opened
Improved copy needed in failure notifications
bug user testing
Failure message (in italics) is lengthy and awkwardly phrased. Links to congresspersons' contact forms are unclear (and what to do with them). These links can be confusingβ€”when a contact page is down or has moved, the user will seem to gets a dead/broken link.
1.0
Improved copy needed in failure notifications - Failure message (in italics) is lengthy and awkwardly phrased. Links to congresspersons' contact forms are unclear (and what to do with them). These links can be confusingβ€”when a contact page is down or has moved, the user will seem to gets a dead/broken link.
non_process
improved copy needed in failure notifications failure message in italics is lengthy and awkwardly phrased links to congresspersons contact forms are unclear and what to do with them these links can be confusingβ€”when a contact page is down or has moved the user will seem to gets a dead broken link
0
34,941
4,599,883,598
IssuesEvent
2016-09-22 01:08:52
MozillaFoundation/Mozfest2016_production
https://api.github.com/repos/MozillaFoundation/Mozfest2016_production
closed
Test and Finalize One Web Theme Colours
Design Production
<img width="1027" alt="screen shot 2016-09-07 at 4 32 10 pm" src="https://cloud.githubusercontent.com/assets/2492510/18332095/ac28b5e6-7518-11e6-9788-08d12bdfc1a1.png"> We are using a bold palette with neons this year and we need to test: - for print in CMYK - for print in Pantone - How the palette works as a whole (example: what if yellow was neon and the rest were not?) - finalize the palette with referencing a pantone book or printed samples We should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do.
1.0
Test and Finalize One Web Theme Colours - <img width="1027" alt="screen shot 2016-09-07 at 4 32 10 pm" src="https://cloud.githubusercontent.com/assets/2492510/18332095/ac28b5e6-7518-11e6-9788-08d12bdfc1a1.png"> We are using a bold palette with neons this year and we need to test: - for print in CMYK - for print in Pantone - How the palette works as a whole (example: what if yellow was neon and the rest were not?) - finalize the palette with referencing a pantone book or printed samples We should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do.
non_process
test and finalize one web theme colours img width alt screen shot at pm src we are using a bold palette with neons this year and we need to test for print in cmyk for print in pantone how the palette works as a whole example what if yellow was neon and the rest were not finalize the palette with referencing a pantone book or printed samples we should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do
0
20,339
26,994,633,445
IssuesEvent
2023-02-09 23:16:59
python/cpython
https://api.github.com/repos/python/cpython
closed
multiprocessing's "spawn" doesn't actually use spawn
type-bug stdlib 3.11 3.10 3.9 expert-multiprocessing
BPO | [46367](https://bugs.python.org/issue46367) --- | :--- Nosy | @vstinner, @jakirkham <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2022-01-13.19:20:59.356> labels = ['type-bug', 'library', '3.9', '3.10', '3.11'] title = 'multiprocessing\'s "spawn" doesn\'t actually use spawn' updated_at = <Date 2022-01-15.00:10:58.327> user = 'https://github.com/jakirkham' ``` bugs.python.org fields: ```python activity = <Date 2022-01-15.00:10:58.327> actor = 'vstinner' assignee = 'none' closed = False closed_date = None closer = None components = ['Library (Lib)'] creation = <Date 2022-01-13.19:20:59.356> creator = 'jakirkham' dependencies = [] files = [] hgrepos = [] issue_num = 46367 keywords = [] message_count = 2.0 messages = ['410512', '410612'] nosy_count = 2.0 nosy_names = ['vstinner', 'jakirkham'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue46367' versions = ['Python 3.9', 'Python 3.10', 'Python 3.11'] ``` </p></details>
1.0
multiprocessing's "spawn" doesn't actually use spawn - BPO | [46367](https://bugs.python.org/issue46367) --- | :--- Nosy | @vstinner, @jakirkham <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2022-01-13.19:20:59.356> labels = ['type-bug', 'library', '3.9', '3.10', '3.11'] title = 'multiprocessing\'s "spawn" doesn\'t actually use spawn' updated_at = <Date 2022-01-15.00:10:58.327> user = 'https://github.com/jakirkham' ``` bugs.python.org fields: ```python activity = <Date 2022-01-15.00:10:58.327> actor = 'vstinner' assignee = 'none' closed = False closed_date = None closer = None components = ['Library (Lib)'] creation = <Date 2022-01-13.19:20:59.356> creator = 'jakirkham' dependencies = [] files = [] hgrepos = [] issue_num = 46367 keywords = [] message_count = 2.0 messages = ['410512', '410612'] nosy_count = 2.0 nosy_names = ['vstinner', 'jakirkham'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue46367' versions = ['Python 3.9', 'Python 3.10', 'Python 3.11'] ``` </p></details>
process
multiprocessing s spawn doesn t actually use spawn bpo nosy vstinner jakirkham note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing s spawn doesn t actually use spawn updated at user bugs python org fields python activity actor vstinner assignee none closed false closed date none closer none components creation creator jakirkham dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions
1
196,143
22,440,849,364
IssuesEvent
2022-06-21 01:05:52
kedacore/test-tools
https://api.github.com/repos/kedacore/test-tools
opened
CVE-2022-33987 (Medium) detected in got-11.8.2.tgz
security vulnerability
## CVE-2022-33987 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>got-11.8.2.tgz</b></p></summary> <p>Human-friendly and powerful HTTP request library for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-11.8.2.tgz">https://registry.npmjs.org/got/-/got-11.8.2.tgz</a></p> <p>Path to dependency file: /e2e/images/selenium-grid/package.json</p> <p>Path to vulnerable library: /e2e/images/selenium-grid/node_modules/got/package.json</p> <p> Dependency Hierarchy: - cli-7.8.0.tgz (Root Library) - webdriverio-7.8.0.tgz - webdriver-7.8.0.tgz - :x: **got-11.8.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-33987 (Medium) detected in got-11.8.2.tgz - ## CVE-2022-33987 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>got-11.8.2.tgz</b></p></summary> <p>Human-friendly and powerful HTTP request library for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-11.8.2.tgz">https://registry.npmjs.org/got/-/got-11.8.2.tgz</a></p> <p>Path to dependency file: /e2e/images/selenium-grid/package.json</p> <p>Path to vulnerable library: /e2e/images/selenium-grid/node_modules/got/package.json</p> <p> Dependency Hierarchy: - cli-7.8.0.tgz (Root Library) - webdriverio-7.8.0.tgz - webdriver-7.8.0.tgz - :x: **got-11.8.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in got tgz cve medium severity vulnerability vulnerable library got tgz human friendly and powerful http request library for node js library home page a href path to dependency file images selenium grid package json path to vulnerable library images selenium grid node modules got package json dependency hierarchy cli tgz root library webdriverio tgz webdriver tgz x got tgz vulnerable library found in head commit a href found in base branch main vulnerability details the got package before for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend
0
14,342
17,369,298,841
IssuesEvent
2021-07-30 11:49:51
encode/uvicorn
https://api.github.com/repos/encode/uvicorn
closed
Sending SIGTERM to parent process when running with --workers hangs indefinitely
bug multiprocessing
### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [x] The bug is reproducible against the latest release and/or `master`. - [x] There are no similar issues or pull requests to fix it yet. ### Describe the bug <!-- A clear and concise description of what the bug is. --> When running uvicorn with multiple workers, sending a SIGTERM ie `kill -15 ppid` there's no graceful shutdown and the process hangs indefinitely on https://github.com/encode/uvicorn/blob/ff4af12d6902bc9d535fe2a948d1df3ffa02b0d3/uvicorn/supervisors/multiprocess.py#L57 ### To reproduce <!-- Provide a *minimal* example with steps to reproduce the bug locally. 1. Run `uvicorn app:app --workers 2 --log-level=debug` 2. Get the ppid in the logs `INFO: Started parent process [38237]` 3. Send the SIGTERM `kill -15 38237` NOTE: try to keep any external dependencies *at an absolute minimum* . In other words, remove anything that doesn't make the bug go away. --> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> A graceful shutdown ### Actual behavior <!-- A clear and concise description of what actually happens. --> ### Debugging material <!-- Any tracebacks, screenshots, etc. that can help understanding the problem. NOTE: - Please list tracebacks in full (don't truncate them). - If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag). - Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d). --> ### Environment - OS / Python / Uvicorn version: just run `uvicorn --version` `Running uvicorn 0.12.2 with CPython 3.8.6 on Linux` - The exact command you're running uvicorn with, all flags you passed included. If you run it with gunicorn please do the same. If there is a reverse-proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce. ### Additional context <!-- Any additional information that can help understanding the problem. Eg. linked issues, or a description of what you were trying to achieve. --> originally discovered in https://github.com/encode/uvicorn/issues/364 and mistakenly taken for a docker issue, it's in fact a SIGTERM issue afaiu
1.0
Sending SIGTERM to parent process when running with --workers hangs indefinitely - ### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [x] The bug is reproducible against the latest release and/or `master`. - [x] There are no similar issues or pull requests to fix it yet. ### Describe the bug <!-- A clear and concise description of what the bug is. --> When running uvicorn with multiple workers, sending a SIGTERM ie `kill -15 ppid` there's no graceful shutdown and the process hangs indefinitely on https://github.com/encode/uvicorn/blob/ff4af12d6902bc9d535fe2a948d1df3ffa02b0d3/uvicorn/supervisors/multiprocess.py#L57 ### To reproduce <!-- Provide a *minimal* example with steps to reproduce the bug locally. 1. Run `uvicorn app:app --workers 2 --log-level=debug` 2. Get the ppid in the logs `INFO: Started parent process [38237]` 3. Send the SIGTERM `kill -15 38237` NOTE: try to keep any external dependencies *at an absolute minimum* . In other words, remove anything that doesn't make the bug go away. --> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> A graceful shutdown ### Actual behavior <!-- A clear and concise description of what actually happens. --> ### Debugging material <!-- Any tracebacks, screenshots, etc. that can help understanding the problem. NOTE: - Please list tracebacks in full (don't truncate them). - If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag). - Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d). --> ### Environment - OS / Python / Uvicorn version: just run `uvicorn --version` `Running uvicorn 0.12.2 with CPython 3.8.6 on Linux` - The exact command you're running uvicorn with, all flags you passed included. If you run it with gunicorn please do the same. If there is a reverse-proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce. ### Additional context <!-- Any additional information that can help understanding the problem. Eg. linked issues, or a description of what you were trying to achieve. --> originally discovered in https://github.com/encode/uvicorn/issues/364 and mistakenly taken for a docker issue, it's in fact a SIGTERM issue afaiu
process
sending sigterm to parent process when running with workers hangs indefinitely checklist the bug is reproducible against the latest release and or master there are no similar issues or pull requests to fix it yet describe the bug when running uvicorn with multiple workers sending a sigterm ie kill ppid there s no graceful shutdown and the process hangs indefinitely on to reproduce provide a minimal example with steps to reproduce the bug locally run uvicorn app app workers log level debug get the ppid in the logs info started parent process send the sigterm kill note try to keep any external dependencies at an absolute minimum in other words remove anything that doesn t make the bug go away expected behavior a graceful shutdown actual behavior debugging material any tracebacks screenshots etc that can help understanding the problem note please list tracebacks in full don t truncate them if relevant consider turning on debug or trace logs for additional details see the logging section on specifically the log level flag consider using to make tracebacks logs collapsible if they re very large see environment os python uvicorn version just run uvicorn version running uvicorn with cpython on linux the exact command you re running uvicorn with all flags you passed included if you run it with gunicorn please do the same if there is a reverse proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce additional context any additional information that can help understanding the problem eg linked issues or a description of what you were trying to achieve originally discovered in and mistakenly taken for a docker issue it s in fact a sigterm issue afaiu
1
19,574
25,894,937,892
IssuesEvent
2022-12-14 21:27:44
googleapis/python-monitoring-dashboards
https://api.github.com/repos/googleapis/python-monitoring-dashboards
closed
Your .repo-metadata.json file has a problem πŸ€’
type: process api: monitoring repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan πŸ“ˆ: * api_shortname 'monitoring-dashboards' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem πŸ€’ - You have a problem with your .repo-metadata.json file: Result of scan πŸ“ˆ: * api_shortname 'monitoring-dashboards' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem πŸ€’ you have a problem with your repo metadata json file result of scan πŸ“ˆ api shortname monitoring dashboards invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
7,136
10,278,674,188
IssuesEvent
2019-08-25 16:18:20
xethya/framework
https://api.github.com/repos/xethya/framework
opened
[publish process] Add "files" property to all packages
monorepo publish process
Packages are being published with folders such as `.rpt2_cache` or `coverage`.
1.0
[publish process] Add "files" property to all packages - Packages are being published with folders such as `.rpt2_cache` or `coverage`.
process
add files property to all packages packages are being published with folders such as cache or coverage
1
5,103
7,882,665,732
IssuesEvent
2018-06-27 00:00:29
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Grammar - Mid statement is not correct
bug parse-tree-processing
Per https://msdn.microsoft.com/en-us/library/ee156918.aspx, this parser rule is not correct: https://github.com/rubberduck-vba/Rubberduck/blob/e2468806b7da88f20ca28a4ea0f08edb92cacd78/Rubberduck.Parsing/Grammar/VBAParser.g4#L467 `ASSIGN` should be `EQ`. This should be fixed, and also a test added. Related: Not sure if it's desirable to have `ASSIGN` defined in the lexer. The only other place it appears to be used is for named arguments, which could equally well be parsed with `COLON EQ`. IMO, having a lexer rule named `ASSIGN` suggests that `:=` is a general-use token in VBA, which it is not.
1.0
Grammar - Mid statement is not correct - Per https://msdn.microsoft.com/en-us/library/ee156918.aspx, this parser rule is not correct: https://github.com/rubberduck-vba/Rubberduck/blob/e2468806b7da88f20ca28a4ea0f08edb92cacd78/Rubberduck.Parsing/Grammar/VBAParser.g4#L467 `ASSIGN` should be `EQ`. This should be fixed, and also a test added. Related: Not sure if it's desirable to have `ASSIGN` defined in the lexer. The only other place it appears to be used is for named arguments, which could equally well be parsed with `COLON EQ`. IMO, having a lexer rule named `ASSIGN` suggests that `:=` is a general-use token in VBA, which it is not.
process
grammar mid statement is not correct per this parser rule is not correct assign should be eq this should be fixed and also a test added related not sure if it s desirable to have assign defined in the lexer the only other place it appears to be used is for named arguments which could equally well be parsed with colon eq imo having a lexer rule named assign suggests that is a general use token in vba which it is not
1
109,942
13,879,680,220
IssuesEvent
2020-10-17 15:29:38
kambiz-frounchi/Foodie
https://api.github.com/repos/kambiz-frounchi/Foodie
opened
Define Express routes
design
These are the routes that come to mind: 1. **HTML routes**: - '/' - '/feed' - '/recipe' - '/postrecipe' - `/findrecipe' - 'userprofile' 2. **API routes**: -
1.0
Define Express routes - These are the routes that come to mind: 1. **HTML routes**: - '/' - '/feed' - '/recipe' - '/postrecipe' - `/findrecipe' - 'userprofile' 2. **API routes**: -
non_process
define express routes these are the routes that come to mind html routes feed recipe postrecipe findrecipe userprofile api routes
0
4,744
7,603,030,396
IssuesEvent
2018-04-29 09:51:12
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
opened
HTML subfigures via flexbox rather than tables?
enhancement postprocessing
I was inspecting an arXiv document (1802.06832) which has 4 subfigures a-d in its Results section and noticed they don't reflow in the HTML. They are currently marked as 4 `td` table cells with class `ltx_subfigure` in the same `tr` row. The fixed single-row display looks quite bad even on large displays, as it is significantly wider than the main article width. And on small displays you end up with a huge horizontal scroll, which is just as bad. As I recently did some flexbox work with the moderncv binding, I'm wondering if @brucemiller would see that as a better mechanism for presenting the subfigures. That would give us decent reflow, and we can contain the max-width to a reasonable responsive width.
1.0
HTML subfigures via flexbox rather than tables? - I was inspecting an arXiv document (1802.06832) which has 4 subfigures a-d in its Results section and noticed they don't reflow in the HTML. They are currently marked as 4 `td` table cells with class `ltx_subfigure` in the same `tr` row. The fixed single-row display looks quite bad even on large displays, as it is significantly wider than the main article width. And on small displays you end up with a huge horizontal scroll, which is just as bad. As I recently did some flexbox work with the moderncv binding, I'm wondering if @brucemiller would see that as a better mechanism for presenting the subfigures. That would give us decent reflow, and we can contain the max-width to a reasonable responsive width.
process
html subfigures via flexbox rather than tables i was inspecting an arxiv document which has subfigures a d in its results section and noticed they don t reflow in the html they are currently marked as td table cells with class ltx subfigure in the same tr row the fixed single row display looks quite bad even on large displays as it is significantly wider than the main article width and on small displays you end up with a huge horizontal scroll which is just as bad as i recently did some flexbox work with the moderncv binding i m wondering if brucemiller would see that as a better mechanism for presenting the subfigures that would give us decent reflow and we can contain the max width to a reasonable responsive width
1
57,900
6,560,279,965
IssuesEvent
2017-09-07 08:42:48
LiskHQ/lisk-js
https://api.github.com/repos/LiskHQ/lisk-js
closed
[178] Add tests for constants
easy test
Part of #178 Not the values themselves, but the shape of the exported object so we don't e.g. accidentally delete a constant we need.
1.0
[178] Add tests for constants - Part of #178 Not the values themselves, but the shape of the exported object so we don't e.g. accidentally delete a constant we need.
non_process
add tests for constants part of not the values themselves but the shape of the exported object so we don t e g accidentally delete a constant we need
0
20,546
27,195,718,281
IssuesEvent
2023-02-20 04:58:02
AvaloniaUI/Avalonia
https://api.github.com/repos/AvaloniaUI/Avalonia
closed
An error occurs when a font with a Chinese name is used
bug area-textprocessing
``` csharp System.InvalidOperationException: Could not create glyph typeface for: 216-δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“. at Avalonia.Skia.FontManagerImpl.CreateGlyphTypeface(Typeface typeface) in /_/src/Skia/Avalonia.Skia/FontManagerImpl.cs:line 146 at Avalonia.Media.GlyphTypeface..ctor(Typeface typeface) in /_/src/Avalonia.Visuals/Media/GlyphTypeface.cs:line 9 at Avalonia.Media.FontManager.GetOrAddGlyphTypeface(Typeface typeface) in /_/src/Avalonia.Visuals/Media/FontManager.cs:line 99 at Avalonia.Media.Typeface.get_GlyphTypeface() in /_/src/Avalonia.Visuals/Media/Typeface.cs:line 69 at Avalonia.Media.TextFormatting.TextCharacters.TryGetRunProperties(ReadOnlySlice`1 text, Typeface typeface, Typeface defaultTypeface, Int32& count) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 132 at Avalonia.Media.TextFormatting.TextCharacters.CreateShapeableRun(ReadOnlySlice`1 text, TextRunProperties defaultProperties) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 62 at Avalonia.Media.TextFormatting.TextCharacters.GetShapeableCharacters() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 40 at Avalonia.Media.TextFormatting.TextFormatterImpl.FetchTextRuns(ITextSource textSource, Int32 firstTextSourceIndex, TextLineBreak previousLineBreak, TextLineBreak& nextLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 286 at Avalonia.Media.TextFormatting.TextFormatterImpl.FormatLine(ITextSource textSource, Int32 firstTextSourceIndex, Double paragraphWidth, TextParagraphProperties paragraphProperties, TextLineBreak previousLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 15 at Avalonia.Media.TextFormatting.TextLayout.UpdateLayout() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 225 at Avalonia.Media.TextFormatting.TextLayout..ctor(String text, Typeface typeface, Double fontSize, IBrush foreground, TextAlignment textAlignment, TextWrapping textWrapping, TextTrimming textTrimming, TextDecorationCollection textDecorations, Double maxWidth, Double maxHeight, Double lineHeight, Int32 maxLines, IReadOnlyList`1 textStyleOverrides) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 71 at Avalonia.Controls.TextBlock.CreateTextLayout(Size constraint, String text) in /_/src/Avalonia.Controls/TextBlock.cs:line 475 at Avalonia.Controls.TextBlock.get_TextLayout() in /_/src/Avalonia.Controls/TextBlock.cs:line 166 at Avalonia.Controls.TextBlock.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/TextBlock.cs:line 521 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Controls.StackPanel.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/StackPanel.cs:line 207 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding, Thickness borderThickness) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 39 at Avalonia.Controls.Presenters.ContentPresenter.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Presenters/ContentPresenter.cs:line 366 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46 at Avalonia.Controls.Decorator.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Decorator.cs:line 54 at Avalonia.Controls.Primitives.VisualLayerManager.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Primitives/VisualLayerManager.cs:line 133 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625 at Avalonia.Controls.Window.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Window.cs:line 937 at Avalonia.Controls.WindowBase.MeasureCore(Size availableSize) in /_/src/Avalonia.Controls/WindowBase.cs:line 247 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutManager.Measure(ILayoutable control) in /_/src/Avalonia.Layout/LayoutManager.cs:line 297 at Avalonia.Layout.LayoutManager.ExecuteInitialLayoutPass() in /_/src/Avalonia.Layout/LayoutManager.cs:line 184 at Avalonia.Controls.Window.ShowCore(Window parent) in /_/src/Avalonia.Controls/Window.cs:line 693 at Avalonia.Controls.Window.Show() in /_/src/Avalonia.Controls/Window.cs:line 631 at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.ShowMainWindow() in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 129 at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.Start(String[] args) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 118 at Avalonia.ClassicDesktopStyleApplicationLifetimeExtensions.StartWithClassicDesktopLifetime[T](T builder, String[] args, ShutdownMode shutdownMode) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 209 at LearnAvalonia.Program.Main(String[] args) in D:\Desktop\LearnAvalonia\LearnAvalonia\Program.cs:line 13 ``` # That's my way of quoting ```xaml FontFamily="resm:LearnAvalonia.Assets.Fonts.No.216-ShangShouZhaoCaiMaoTi-2.ttf?assembly=LearnAvalonia#216-δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“" ``` <img width="911" alt="image" src="https://user-images.githubusercontent.com/108921607/219945857-7588a20d-c66f-4fca-8930-9dea9e9b22ea.png">
1.0
An error occurs when a font with a Chinese name is used - ``` csharp System.InvalidOperationException: Could not create glyph typeface for: 216-δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“. at Avalonia.Skia.FontManagerImpl.CreateGlyphTypeface(Typeface typeface) in /_/src/Skia/Avalonia.Skia/FontManagerImpl.cs:line 146 at Avalonia.Media.GlyphTypeface..ctor(Typeface typeface) in /_/src/Avalonia.Visuals/Media/GlyphTypeface.cs:line 9 at Avalonia.Media.FontManager.GetOrAddGlyphTypeface(Typeface typeface) in /_/src/Avalonia.Visuals/Media/FontManager.cs:line 99 at Avalonia.Media.Typeface.get_GlyphTypeface() in /_/src/Avalonia.Visuals/Media/Typeface.cs:line 69 at Avalonia.Media.TextFormatting.TextCharacters.TryGetRunProperties(ReadOnlySlice`1 text, Typeface typeface, Typeface defaultTypeface, Int32& count) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 132 at Avalonia.Media.TextFormatting.TextCharacters.CreateShapeableRun(ReadOnlySlice`1 text, TextRunProperties defaultProperties) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 62 at Avalonia.Media.TextFormatting.TextCharacters.GetShapeableCharacters() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 40 at Avalonia.Media.TextFormatting.TextFormatterImpl.FetchTextRuns(ITextSource textSource, Int32 firstTextSourceIndex, TextLineBreak previousLineBreak, TextLineBreak& nextLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 286 at Avalonia.Media.TextFormatting.TextFormatterImpl.FormatLine(ITextSource textSource, Int32 firstTextSourceIndex, Double paragraphWidth, TextParagraphProperties paragraphProperties, TextLineBreak previousLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 15 at Avalonia.Media.TextFormatting.TextLayout.UpdateLayout() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 225 at Avalonia.Media.TextFormatting.TextLayout..ctor(String text, Typeface typeface, Double fontSize, IBrush foreground, TextAlignment textAlignment, TextWrapping textWrapping, TextTrimming textTrimming, TextDecorationCollection textDecorations, Double maxWidth, Double maxHeight, Double lineHeight, Int32 maxLines, IReadOnlyList`1 textStyleOverrides) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 71 at Avalonia.Controls.TextBlock.CreateTextLayout(Size constraint, String text) in /_/src/Avalonia.Controls/TextBlock.cs:line 475 at Avalonia.Controls.TextBlock.get_TextLayout() in /_/src/Avalonia.Controls/TextBlock.cs:line 166 at Avalonia.Controls.TextBlock.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/TextBlock.cs:line 521 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Controls.StackPanel.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/StackPanel.cs:line 207 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding, Thickness borderThickness) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 39 at Avalonia.Controls.Presenters.ContentPresenter.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Presenters/ContentPresenter.cs:line 366 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46 at Avalonia.Controls.Decorator.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Decorator.cs:line 54 at Avalonia.Controls.Primitives.VisualLayerManager.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Primitives/VisualLayerManager.cs:line 133 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625 at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625 at Avalonia.Controls.Window.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Window.cs:line 937 at Avalonia.Controls.WindowBase.MeasureCore(Size availableSize) in /_/src/Avalonia.Controls/WindowBase.cs:line 247 at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364 at Avalonia.Layout.LayoutManager.Measure(ILayoutable control) in /_/src/Avalonia.Layout/LayoutManager.cs:line 297 at Avalonia.Layout.LayoutManager.ExecuteInitialLayoutPass() in /_/src/Avalonia.Layout/LayoutManager.cs:line 184 at Avalonia.Controls.Window.ShowCore(Window parent) in /_/src/Avalonia.Controls/Window.cs:line 693 at Avalonia.Controls.Window.Show() in /_/src/Avalonia.Controls/Window.cs:line 631 at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.ShowMainWindow() in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 129 at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.Start(String[] args) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 118 at Avalonia.ClassicDesktopStyleApplicationLifetimeExtensions.StartWithClassicDesktopLifetime[T](T builder, String[] args, ShutdownMode shutdownMode) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 209 at LearnAvalonia.Program.Main(String[] args) in D:\Desktop\LearnAvalonia\LearnAvalonia\Program.cs:line 13 ``` # That's my way of quoting ```xaml FontFamily="resm:LearnAvalonia.Assets.Fonts.No.216-ShangShouZhaoCaiMaoTi-2.ttf?assembly=LearnAvalonia#216-δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“" ``` <img width="911" alt="image" src="https://user-images.githubusercontent.com/108921607/219945857-7588a20d-c66f-4fca-8930-9dea9e9b22ea.png">
process
an error occurs when a font with a chinese name is used csharp system invalidoperationexception could not create glyph typeface for δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“ at avalonia skia fontmanagerimpl createglyphtypeface typeface typeface in src skia avalonia skia fontmanagerimpl cs line at avalonia media glyphtypeface ctor typeface typeface in src avalonia visuals media glyphtypeface cs line at avalonia media fontmanager getoraddglyphtypeface typeface typeface in src avalonia visuals media fontmanager cs line at avalonia media typeface get glyphtypeface in src avalonia visuals media typeface cs line at avalonia media textformatting textcharacters trygetrunproperties readonlyslice text typeface typeface typeface defaulttypeface count in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textcharacters createshapeablerun readonlyslice text textrunproperties defaultproperties in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textcharacters getshapeablecharacters in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textformatterimpl fetchtextruns itextsource textsource firsttextsourceindex textlinebreak previouslinebreak textlinebreak nextlinebreak in src avalonia visuals media textformatting textformatterimpl cs line at avalonia media textformatting textformatterimpl formatline itextsource textsource firsttextsourceindex double paragraphwidth textparagraphproperties paragraphproperties textlinebreak previouslinebreak in src avalonia visuals media textformatting textformatterimpl cs line at avalonia media textformatting textlayout updatelayout in src avalonia visuals media textformatting textlayout cs line at avalonia media textformatting textlayout ctor string text typeface typeface double fontsize ibrush foreground textalignment textalignment textwrapping textwrapping texttrimming texttrimming textdecorationcollection textdecorations double maxwidth double maxheight double lineheight maxlines ireadonlylist textstyleoverrides in src avalonia visuals media textformatting textlayout cs line at avalonia controls textblock createtextlayout size constraint string text in src avalonia controls textblock cs line at avalonia controls textblock get textlayout in src avalonia controls textblock cs line at avalonia controls textblock measureoverride size availablesize in src avalonia controls textblock cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia controls stackpanel measureoverride size availablesize in src avalonia controls stackpanel cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding in src avalonia layout layouthelper cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding thickness borderthickness in src avalonia layout layouthelper cs line at avalonia controls presenters contentpresenter measureoverride size availablesize in src avalonia controls presenters contentpresenter cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding in src avalonia layout layouthelper cs line at avalonia controls decorator measureoverride size availablesize in src avalonia controls decorator cs line at avalonia controls primitives visuallayermanager measureoverride size availablesize in src avalonia controls primitives visuallayermanager cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measureoverride size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measureoverride size availablesize in src avalonia layout layoutable cs line at avalonia controls window measureoverride size availablesize in src avalonia controls window cs line at avalonia controls windowbase measurecore size availablesize in src avalonia controls windowbase cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutmanager measure ilayoutable control in src avalonia layout layoutmanager cs line at avalonia layout layoutmanager executeinitiallayoutpass in src avalonia layout layoutmanager cs line at avalonia controls window showcore window parent in src avalonia controls window cs line at avalonia controls window show in src avalonia controls window cs line at avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime showmainwindow in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime start string args in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at avalonia classicdesktopstyleapplicationlifetimeextensions startwithclassicdesktoplifetime t builder string args shutdownmode shutdownmode in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at learnavalonia program main string args in d desktop learnavalonia learnavalonia program cs line that s my way of quoting xaml fontfamily resm learnavalonia assets fonts no shangshouzhaocaimaoti ttf assembly learnavalonia δΈŠι¦–ζ‹›θ΄’ηŒ«δ½“ img width alt image src
1
8,624
11,779,265,573
IssuesEvent
2020-03-16 17:42:01
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
PH - metric Heat Balance Operations Tab
Process Heating Quick Fix
In the process heating tool if you set the units to metric in setup the operations assessment tab still shows $/MMBtu. I think the calculations are ok, its just that the units don't change correctly. I *think* they should say GJ. Look at the unit conversion to double check. ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/6a7d8bf1-d32d-451e-adc5-6d95cff4dec7)
1.0
PH - metric Heat Balance Operations Tab - In the process heating tool if you set the units to metric in setup the operations assessment tab still shows $/MMBtu. I think the calculations are ok, its just that the units don't change correctly. I *think* they should say GJ. Look at the unit conversion to double check. ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/6a7d8bf1-d32d-451e-adc5-6d95cff4dec7)
process
ph metric heat balance operations tab in the process heating tool if you set the units to metric in setup the operations assessment tab still shows mmbtu i think the calculations are ok its just that the units don t change correctly i think they should say gj look at the unit conversion to double check
1
6,663
9,782,053,081
IssuesEvent
2019-06-07 21:48:04
googleapis/google-cloud-java
https://api.github.com/repos/googleapis/google-cloud-java
closed
Perform snapshot releases after successful builds to master
type: process
We lost releases of snapshot versions when we turned off the oraclejdk7 build on Travis (since the after_success.sh script only performed releases for that build variant). Now that we are using Circle, we should add a step in the CircleCI workflow to perform a snapshot release after all jobs finish (operating systems + integration tests).
1.0
Perform snapshot releases after successful builds to master - We lost releases of snapshot versions when we turned off the oraclejdk7 build on Travis (since the after_success.sh script only performed releases for that build variant). Now that we are using Circle, we should add a step in the CircleCI workflow to perform a snapshot release after all jobs finish (operating systems + integration tests).
process
perform snapshot releases after successful builds to master we lost releases of snapshot versions when we turned off the build on travis since the after success sh script only performed releases for that build variant now that we are using circle we should add a step in the circleci workflow to perform a snapshot release after all jobs finish operating systems integration tests
1
17,413
23,228,628,556
IssuesEvent
2022-08-03 04:44:38
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Processing history rework (Request in QGIS)
Processing 3.24
### Request for documentation From pull request QGIS/qgis#46525 Author: @nyalldawson QGIS version: 3.24 **Processing history rework** ### PR Description: This PR partially implements the framework described in https://github.com/qgis/QGIS-Enhancement-Proposals/issues/130, and ports the processing history log over to this framework. While there's no user visible changes which arise from this, it's a much more flexible framework as it allows us to store a full variant map in the history. So we can use this in future PRs to store more than just the python command in the history (e.g. storing the qgis_process command, the environment of the algorithm execution (ellipsoid settings, distance/area units, etc), AND any outputs given by the algorithm execution and the full output log. History log storage in now stored in a sqlite database instead of the text file previously used, which will allow for more efficient storage/retrieval/filtering. (This framework extends to use for more than just Processing -- it could also be used to store the history of other user operations, such as project edits, layout exports, ...) Also now includes new right-click actions in the Processing history dialog to copy entries as JSON or qgis_process commands Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature][processing] As 'copy as json' and 'copy as qgis_process\ncommand' actions to right click menu in history dialog"
1.0
Processing history rework (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#46525 Author: @nyalldawson QGIS version: 3.24 **Processing history rework** ### PR Description: This PR partially implements the framework described in https://github.com/qgis/QGIS-Enhancement-Proposals/issues/130, and ports the processing history log over to this framework. While there's no user visible changes which arise from this, it's a much more flexible framework as it allows us to store a full variant map in the history. So we can use this in future PRs to store more than just the python command in the history (e.g. storing the qgis_process command, the environment of the algorithm execution (ellipsoid settings, distance/area units, etc), AND any outputs given by the algorithm execution and the full output log. History log storage in now stored in a sqlite database instead of the text file previously used, which will allow for more efficient storage/retrieval/filtering. (This framework extends to use for more than just Processing -- it could also be used to store the history of other user operations, such as project edits, layout exports, ...) Also now includes new right-click actions in the Processing history dialog to copy entries as JSON or qgis_process commands Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature][processing] As 'copy as json' and 'copy as qgis_process\ncommand' actions to right click menu in history dialog"
process
processing history rework request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version processing history rework pr description this pr partially implements the framework described in and ports the processing history log over to this framework while there s no user visible changes which arise from this it s a much more flexible framework as it allows us to store a full variant map in the history so we can use this in future prs to store more than just the python command in the history e g storing the qgis process command the environment of the algorithm execution ellipsoid settings distance area units etc and any outputs given by the algorithm execution and the full output log history log storage in now stored in a sqlite database instead of the text file previously used which will allow for more efficient storage retrieval filtering this framework extends to use for more than just processing it could also be used to store the history of other user operations such as project edits layout exports also now includes new right click actions in the processing history dialog to copy entries as json or qgis process commands sponsored by the research institute for nature and forest flemish govt commits tagged with or as copy as json and copy as qgis process ncommand actions to right click menu in history dialog
1
19,611
25,962,622,691
IssuesEvent
2022-12-19 01:57:49
AbdElAziz333/SmoothBoot-Reloaded
https://api.github.com/repos/AbdElAziz333/SmoothBoot-Reloaded
closed
version check data
in processing
It spams my logs with [Forge Version Check/DEBUG] [net.minecraftforge.fml.VersionChecker/]: [smoothboot] Received version check data: ``` <!DOCTYPE html> <html lang="en" data-color-mode="auto" data-light-theme="light" data-dark-theme="dark" data-a11y-animated-images="system"> <head> <meta charset="utf-8"> <link rel="dns-prefetch" href="https://github.githubassets.com"> <link rel="dns-prefetch" href="https://avatars.githubusercontent.com"> <link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com"> <link rel="dns-prefetch" href="https://user-images.githubusercontent.com/"> <link rel="preconnect" href="https://github.githubassets.com" crossorigin> <link rel="preconnect" href="https://avatars.githubusercontent.com"> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/light-719f1193e0c0.css" /><link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/dark-0c343b529849.css" /><link data-color-theme="dark_dimmed" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_dimmed-f22da508b62a.css" /><link data-color-theme="dark_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_high_contrast-188ef1de59e6.css" /><link data-color-theme="dark_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_colorblind-bc6bf4eea850.css" /><link data-color-theme="light_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_colorblind-527658dec362.css" /><link data-color-theme="light_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_high_contrast-c7a7fe0cd8ec.css" /><link data-color-theme="light_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_tritanopia-6aa855bdae0f.css" /><link data-color-theme="dark_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_tritanopia-6aa5e25aacc0.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/primer-ef7ca334db1f.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/global-86610d5db319.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/github-6d0bfca6b879.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/code-fa476fb5c04d.css" /> <meta name="optimizely-datafile" content="{&quot;groups&quot;: [], &quot;environmentKey&quot;: &quot;production&quot;, &quot;rollouts&quot;: [], &quot;typedAudiences&quot;: [], &quot;projectId&quot;: &quot;16737760170&quot;, &quot;variables&quot;: [], &quot;featureFlags&quot;: [], &quot;experiments&quot;: [], &quot;version&quot;: &quot;4&quot;, &quot;audiences&quot;: [{&quot;conditions&quot;: &quot;[\&quot;or\&quot;, {\&quot;match\&quot;: \&quot;exact\&quot;, \&quot;name\&quot;: \&quot;$opt_dummy_attribute\&quot;, \&quot;type\&quot;: \&quot;custom_attribute\&quot;, \&quot;value\&quot;: \&quot;$opt_dummy_value\&quot;}]&quot;, &quot;id&quot;: &quot;$opt_dummy_audience&quot;, &quot;name&quot;: &quot;Optimizely-Generated Audience for Backwards Compatibility&quot;}], &quot;anonymizeIP&quot;: true, &quot;sdkKey&quot;: &quot;WTc6awnGuYDdG98CYRban&quot;, &quot;attributes&quot;: [{&quot;id&quot;: &quot;16822470375&quot;, &quot;key&quot;: &quot;user_id&quot;}, {&quot;id&quot;: &quot;17143601254&quot;, &quot;key&quot;: &quot;spammy&quot;}, {&quot;id&quot;: &quot;18175660309&quot;, &quot;key&quot;: [...]``` had to truncate
1.0
version check data - It spams my logs with [Forge Version Check/DEBUG] [net.minecraftforge.fml.VersionChecker/]: [smoothboot] Received version check data: ``` <!DOCTYPE html> <html lang="en" data-color-mode="auto" data-light-theme="light" data-dark-theme="dark" data-a11y-animated-images="system"> <head> <meta charset="utf-8"> <link rel="dns-prefetch" href="https://github.githubassets.com"> <link rel="dns-prefetch" href="https://avatars.githubusercontent.com"> <link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com"> <link rel="dns-prefetch" href="https://user-images.githubusercontent.com/"> <link rel="preconnect" href="https://github.githubassets.com" crossorigin> <link rel="preconnect" href="https://avatars.githubusercontent.com"> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/light-719f1193e0c0.css" /><link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/dark-0c343b529849.css" /><link data-color-theme="dark_dimmed" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_dimmed-f22da508b62a.css" /><link data-color-theme="dark_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_high_contrast-188ef1de59e6.css" /><link data-color-theme="dark_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_colorblind-bc6bf4eea850.css" /><link data-color-theme="light_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_colorblind-527658dec362.css" /><link data-color-theme="light_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_high_contrast-c7a7fe0cd8ec.css" /><link data-color-theme="light_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_tritanopia-6aa855bdae0f.css" /><link data-color-theme="dark_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_tritanopia-6aa5e25aacc0.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/primer-ef7ca334db1f.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/global-86610d5db319.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/github-6d0bfca6b879.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/code-fa476fb5c04d.css" /> <meta name="optimizely-datafile" content="{&quot;groups&quot;: [], &quot;environmentKey&quot;: &quot;production&quot;, &quot;rollouts&quot;: [], &quot;typedAudiences&quot;: [], &quot;projectId&quot;: &quot;16737760170&quot;, &quot;variables&quot;: [], &quot;featureFlags&quot;: [], &quot;experiments&quot;: [], &quot;version&quot;: &quot;4&quot;, &quot;audiences&quot;: [{&quot;conditions&quot;: &quot;[\&quot;or\&quot;, {\&quot;match\&quot;: \&quot;exact\&quot;, \&quot;name\&quot;: \&quot;$opt_dummy_attribute\&quot;, \&quot;type\&quot;: \&quot;custom_attribute\&quot;, \&quot;value\&quot;: \&quot;$opt_dummy_value\&quot;}]&quot;, &quot;id&quot;: &quot;$opt_dummy_audience&quot;, &quot;name&quot;: &quot;Optimizely-Generated Audience for Backwards Compatibility&quot;}], &quot;anonymizeIP&quot;: true, &quot;sdkKey&quot;: &quot;WTc6awnGuYDdG98CYRban&quot;, &quot;attributes&quot;: [{&quot;id&quot;: &quot;16822470375&quot;, &quot;key&quot;: &quot;user_id&quot;}, {&quot;id&quot;: &quot;17143601254&quot;, &quot;key&quot;: &quot;spammy&quot;}, {&quot;id&quot;: &quot;18175660309&quot;, &quot;key&quot;: [...]``` had to truncate
process
version check data it spams my logs with received version check data link rel dns prefetch href link rel dns prefetch href link rel dns prefetch href link rel dns prefetch href link rel preconnect href meta name optimizely datafile content quot groups quot quot environmentkey quot quot production quot quot rollouts quot quot typedaudiences quot quot projectid quot quot quot quot variables quot quot featureflags quot quot experiments quot quot version quot quot quot quot audiences quot quot quot id quot quot opt dummy audience quot quot name quot quot optimizely generated audience for backwards compatibility quot quot anonymizeip quot true quot sdkkey quot quot quot quot attributes quot quot id quot quot quot quot key quot quot user id quot quot id quot quot quot quot key quot quot spammy quot quot id quot quot quot quot key quot had to truncate
1
4,659
7,496,694,231
IssuesEvent
2018-04-08 12:17:10
ODiogoSilva/assemblerflow
https://api.github.com/repos/ODiogoSilva/assemblerflow
closed
Module to collect reports
enhancement process
Create a terminal module that collects the `.report.json` files generated throughout the pipeline and adds the fields required for parsing by [report-nf](https://github.com/ODiogoSilva/report-nf)
1.0
Module to collect reports - Create a terminal module that collects the `.report.json` files generated throughout the pipeline and adds the fields required for parsing by [report-nf](https://github.com/ODiogoSilva/report-nf)
process
module to collect reports create a terminal module that collects the report json files generated throughout the pipeline and adds the fields required for parsing by
1
12,860
15,252,209,726
IssuesEvent
2021-02-20 01:58:15
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Support drag and drop in a processing model
Feature Request Feedback Processing stale
Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [21009](https://issues.qgis.org/issues/21009) Redmine category:processing/modeller --- I would like to see support for drag and drop in Processing models. When I drop a file (CSV, GeoPackage etc.) from my Windows deskop on my model, QGIS should be able to accept it as a data source.
1.0
Support drag and drop in a processing model - Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [21009](https://issues.qgis.org/issues/21009) Redmine category:processing/modeller --- I would like to see support for drag and drop in Processing models. When I drop a file (CSV, GeoPackage etc.) from my Windows deskop on my model, QGIS should be able to accept it as a data source.
process
support drag and drop in a processing model author name magnus nilsson magnus nilsson original redmine issue redmine category processing modeller i would like to see support for drag and drop in processing models when i drop a file csv geopackage etc from my windows deskop on my model qgis should be able to accept it as a data source
1
465,300
13,382,298,460
IssuesEvent
2020-09-02 08:36:30
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
opened
Trying to input answers in landscape is impossible due to keyboard automatically dismissing.
Priority: Essential Status: In implementation Status: Not started Type: Bug Where: Exploration player Where: Skills/Questions
**Describe the bug** When you try to input anything in landscape, the keyboard automatically dismisses. **To Reproduce** Steps to reproduce the behavior: 1. Open the app 2. Navigate to the prototype exploration 3. Rotate to landscape 4. Continue to the second state (fractions) 5. Select the fraction input box 6. Observe the soft keyboard open 7. Observe the soft keyboard immediately close **Expected behavior** The soft keyboard shouldn't be automatically closing. **Screenshots** - **Device** - Device: Pixel XL - SDK version: SDK 29 **Additional context** This is definitely a regression since StateFragmentTest was passing 2 weeks ago (the last time I tried these scenarios).
1.0
Trying to input answers in landscape is impossible due to keyboard automatically dismissing. - **Describe the bug** When you try to input anything in landscape, the keyboard automatically dismisses. **To Reproduce** Steps to reproduce the behavior: 1. Open the app 2. Navigate to the prototype exploration 3. Rotate to landscape 4. Continue to the second state (fractions) 5. Select the fraction input box 6. Observe the soft keyboard open 7. Observe the soft keyboard immediately close **Expected behavior** The soft keyboard shouldn't be automatically closing. **Screenshots** - **Device** - Device: Pixel XL - SDK version: SDK 29 **Additional context** This is definitely a regression since StateFragmentTest was passing 2 weeks ago (the last time I tried these scenarios).
non_process
trying to input answers in landscape is impossible due to keyboard automatically dismissing describe the bug when you try to input anything in landscape the keyboard automatically dismisses to reproduce steps to reproduce the behavior open the app navigate to the prototype exploration rotate to landscape continue to the second state fractions select the fraction input box observe the soft keyboard open observe the soft keyboard immediately close expected behavior the soft keyboard shouldn t be automatically closing screenshots device device pixel xl sdk version sdk additional context this is definitely a regression since statefragmenttest was passing weeks ago the last time i tried these scenarios
0
72,855
31,769,565,512
IssuesEvent
2023-09-12 10:53:11
gauravrs18/issue_onboarding
https://api.github.com/repos/gauravrs18/issue_onboarding
closed
dev-angular-style-account-services-new-connection-component-activate-component -consumer-details-component -application-component -address-component
CX-account-services
dev-angular-style-account-services-new-connection-component-activate-component -consumer-details-component -application-component -address-component
1.0
dev-angular-style-account-services-new-connection-component-activate-component -consumer-details-component -application-component -address-component - dev-angular-style-account-services-new-connection-component-activate-component -consumer-details-component -application-component -address-component
non_process
dev angular style account services new connection component activate component consumer details component application component address component dev angular style account services new connection component activate component consumer details component application component address component
0
55,086
13,521,887,011
IssuesEvent
2020-09-15 07:44:38
spring-projects/spring-session
https://api.github.com/repos/spring-projects/spring-session
closed
spring-session-bom `2020.0.0-SNAPSHOT` refers to milestones
in: build type: dependency-upgrade
The latest [spring-session-bom](https://repo.spring.io/snapshot/org/springframework/session/spring-session-bom/2020.0.0-SNAPSHOT/spring-session-bom-2020.0.0-20200915.060532-123.pom) on repo.spring.io has references to milestone 1 artifacts. I believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot.
1.0
spring-session-bom `2020.0.0-SNAPSHOT` refers to milestones - The latest [spring-session-bom](https://repo.spring.io/snapshot/org/springframework/session/spring-session-bom/2020.0.0-SNAPSHOT/spring-session-bom-2020.0.0-20200915.060532-123.pom) on repo.spring.io has references to milestone 1 artifacts. I believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot.
non_process
spring session bom snapshot refers to milestones the latest on repo spring io has references to milestone artifacts i believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot
0
298,632
22,540,886,050
IssuesEvent
2022-06-26 00:22:46
apache/camel-k
https://api.github.com/repos/apache/camel-k
closed
In troubleshooting documentation section, provide information on which logs to look at
area/documentation status/stale
In troubleshooting documentation section provide information on which log to look at operator log/kit log, others? using kubectl/VS Code?
1.0
In troubleshooting documentation section, provide information on which logs to look at - In troubleshooting documentation section provide information on which log to look at operator log/kit log, others? using kubectl/VS Code?
non_process
in troubleshooting documentation section provide information on which logs to look at in troubleshooting documentation section provide information on which log to look at operator log kit log others using kubectl vs code
0
4,614
7,459,552,181
IssuesEvent
2018-03-30 15:47:58
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
bigquery parquet loading code not available on pypi, please push dev release
api: bigquery type: process
I'm evaluating the BigQuery new support for loading from parquet files and following the code samples from: https://cloud.google.com/bigquery/docs/loading-data-cloud-storage-parquet The above uses code that is not yet on pypi. I can build it locally but for rolling out our real tests it would be convenient if a release could be made. It doesn't have to be an offical release, just pushing the dev release would be enough: https://github.com/GoogleCloudPlatform/google-cloud-python/commit/fc91c877eb6aee2a8edfd5a8950bc5ca51e8255c /cc @tswast
1.0
bigquery parquet loading code not available on pypi, please push dev release - I'm evaluating the BigQuery new support for loading from parquet files and following the code samples from: https://cloud.google.com/bigquery/docs/loading-data-cloud-storage-parquet The above uses code that is not yet on pypi. I can build it locally but for rolling out our real tests it would be convenient if a release could be made. It doesn't have to be an offical release, just pushing the dev release would be enough: https://github.com/GoogleCloudPlatform/google-cloud-python/commit/fc91c877eb6aee2a8edfd5a8950bc5ca51e8255c /cc @tswast
process
bigquery parquet loading code not available on pypi please push dev release i m evaluating the bigquery new support for loading from parquet files and following the code samples from the above uses code that is not yet on pypi i can build it locally but for rolling out our real tests it would be convenient if a release could be made it doesn t have to be an offical release just pushing the dev release would be enough cc tswast
1
13,715
16,480,396,106
IssuesEvent
2021-05-24 10:50:33
New-Time-Development/OmeCord
https://api.github.com/repos/New-Time-Development/OmeCord
closed
Chat system does work not correct
Big issue Known bug bug work in process
I fix it today. Everybody can write in the chat. Thats bad
1.0
Chat system does work not correct - I fix it today. Everybody can write in the chat. Thats bad
process
chat system does work not correct i fix it today everybody can write in the chat thats bad
1
3,282
6,369,676,148
IssuesEvent
2017-08-01 12:32:05
gatewayapps/ims-feedback
https://api.github.com/repos/gatewayapps/ims-feedback
closed
It shouldn't be scoring the activities, it should be scoring the aspects.??
bug ims.packages.process
Version: 1.1.1-master.53(9e63209) User: CHRISSIE GRAY(chrissie.gray@nucor-yamato.com)
1.0
It shouldn't be scoring the activities, it should be scoring the aspects.?? - Version: 1.1.1-master.53(9e63209) User: CHRISSIE GRAY(chrissie.gray@nucor-yamato.com)
process
it shouldn t be scoring the activities it should be scoring the aspects version master user chrissie gray chrissie gray nucor yamato com
1
29,625
11,764,607,048
IssuesEvent
2020-03-14 13:31:45
Molizo/FTC-Scouting-App-Skystone
https://api.github.com/repos/Molizo/FTC-Scouting-App-Skystone
opened
CVE-2019-0820 (High) detected in msbump.2.3.2.nupkg
security vulnerability
## CVE-2019-0820 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>msbump.2.3.2.nupkg</b></p></summary> <p>MSBuild task that bumps the version of a Visual Studio 2017 project before build and pack.</p> <p>Library home page: <a href="https://api.nuget.org/packages/msbump.2.3.2.nupkg">https://api.nuget.org/packages/msbump.2.3.2.nupkg</a></p> <p>Path to dependency file: /tmp/ws-scm/FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p> <p>Path to vulnerable library: /FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p> <p> Dependency Hierarchy: - :x: **msbump.2.3.2.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Molizo/FTC-Scouting-App-Skystone/commit/7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f">7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981. <p>Publish Date: 2019-05-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820</a></p> <p>Release Date: 2019-05-16</p> <p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-0820 (High) detected in msbump.2.3.2.nupkg - ## CVE-2019-0820 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>msbump.2.3.2.nupkg</b></p></summary> <p>MSBuild task that bumps the version of a Visual Studio 2017 project before build and pack.</p> <p>Library home page: <a href="https://api.nuget.org/packages/msbump.2.3.2.nupkg">https://api.nuget.org/packages/msbump.2.3.2.nupkg</a></p> <p>Path to dependency file: /tmp/ws-scm/FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p> <p>Path to vulnerable library: /FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p> <p> Dependency Hierarchy: - :x: **msbump.2.3.2.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Molizo/FTC-Scouting-App-Skystone/commit/7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f">7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981. <p>Publish Date: 2019-05-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820</a></p> <p>Release Date: 2019-05-16</p> <p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in msbump nupkg cve high severity vulnerability vulnerable library msbump nupkg msbuild task that bumps the version of a visual studio project before build and pack library home page a href path to dependency file tmp ws scm ftc scouting app skystone skystonescouting skystonescouting csproj path to vulnerable library ftc scouting app skystone skystonescouting skystonescouting csproj dependency hierarchy x msbump nupkg vulnerable library found in head commit a href vulnerability details a denial of service vulnerability exists when net framework and net core improperly process regex strings aka net framework and net core denial of service vulnerability this cve id is unique from cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text regularexpressions step up your open source security game with whitesource
0
18,461
24,549,627,108
IssuesEvent
2022-10-12 11:34:00
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Admins > Add new admin > UI issue in add new admin screen
Bug P2 Participant manager Process: Fixed Process: Tested dev
**Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Verify the email field **AR:** Dropdown icon is getting displayed in the email field **ER:** Dropdown icon should not get displayed in the email field ![PmUI](https://user-images.githubusercontent.com/86007179/173838121-47c86858-ea04-482c-b67c-9367e0a67389.png)
2.0
[PM] Admins > Add new admin > UI issue in add new admin screen - **Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Verify the email field **AR:** Dropdown icon is getting displayed in the email field **ER:** Dropdown icon should not get displayed in the email field ![PmUI](https://user-images.githubusercontent.com/86007179/173838121-47c86858-ea04-482c-b67c-9367e0a67389.png)
process
admins add new admin ui issue in add new admin screen steps login to pm click on admins tab click on add new admin button verify the email field ar dropdown icon is getting displayed in the email field er dropdown icon should not get displayed in the email field
1
300,114
25,945,916,576
IssuesEvent
2022-12-17 00:57:54
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Retain LGTM through squashes
sig/contributor-experience kind/feature help wanted sig/testing lifecycle/frozen triage/accepted
#### What would you like to be added: Configure this repo ([k/kubernetes](https://github.com/kubernetes/kubernetes)) to retain LGTM labels when people squash commits. PR https://github.com/kubernetes/test-infra/pull/9138 made this possible. #### Why is this needed: This feature will make it easier for people to squash commits, and should give reviewers confidence to ask other contributors to squash when appropriate. To implement this change, make a PR against ([k/test-infra](https://github.com/kubernetes/test-infra)) that is similar to https://github.com/kubernetes/test-infra/pull/19304.
1.0
Retain LGTM through squashes - #### What would you like to be added: Configure this repo ([k/kubernetes](https://github.com/kubernetes/kubernetes)) to retain LGTM labels when people squash commits. PR https://github.com/kubernetes/test-infra/pull/9138 made this possible. #### Why is this needed: This feature will make it easier for people to squash commits, and should give reviewers confidence to ask other contributors to squash when appropriate. To implement this change, make a PR against ([k/test-infra](https://github.com/kubernetes/test-infra)) that is similar to https://github.com/kubernetes/test-infra/pull/19304.
non_process
retain lgtm through squashes what would you like to be added configure this repo to retain lgtm labels when people squash commits pr made this possible why is this needed this feature will make it easier for people to squash commits and should give reviewers confidence to ask other contributors to squash when appropriate to implement this change make a pr against that is similar to
0
150,677
11,980,556,182
IssuesEvent
2020-04-07 09:32:24
status-im/status-react
https://api.github.com/repos/status-im/status-react
closed
Can't find web3 in browser (on simpledapp.eth) when refreshing the page
bug e2e test blocker fix them all webview
# Bug Report ## Problem When refreshing browser page with simpledapp.eth then it's often "..fails to find web3 library" While I could not reproduce it with usual opening Dapp in browser, it's reproduced 2 of 3 times when refreshing the page #### Expected behavior Refreshing the page and finds web3 fine #### Actual behavior <img width="303" alt="Screenshot 2020-03-26 at 12 34 46" src="https://user-images.githubusercontent.com/8749671/77637368-3bda6980-6f5e-11ea-8105-272544503246.png"> ### Reproduction - Open Status and create new account - Open `simpledapp.eth` in browser - Tap `Refresh` button (top-right) ### Additional Information - Status version: develop nightly 26th of March - Operating System: Android #### Logs ``` 03-26 12:28:37.445 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 0 03-26 12:28:37.542 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 1 03-26 12:28:37.719 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in 03-26 12:28:37.937 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/loading-started 03-26 12:28:37.939 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in 03-26 12:28:38.456 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/navigation-state-changed 03-26 12:28:42.128 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :mailserver/check-connection-timeout ```
1.0
Can't find web3 in browser (on simpledapp.eth) when refreshing the page - # Bug Report ## Problem When refreshing browser page with simpledapp.eth then it's often "..fails to find web3 library" While I could not reproduce it with usual opening Dapp in browser, it's reproduced 2 of 3 times when refreshing the page #### Expected behavior Refreshing the page and finds web3 fine #### Actual behavior <img width="303" alt="Screenshot 2020-03-26 at 12 34 46" src="https://user-images.githubusercontent.com/8749671/77637368-3bda6980-6f5e-11ea-8105-272544503246.png"> ### Reproduction - Open Status and create new account - Open `simpledapp.eth` in browser - Tap `Refresh` button (top-right) ### Additional Information - Status version: develop nightly 26th of March - Operating System: Android #### Logs ``` 03-26 12:28:37.445 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 0 03-26 12:28:37.542 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 1 03-26 12:28:37.719 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in 03-26 12:28:37.937 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/loading-started 03-26 12:28:37.939 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in 03-26 12:28:38.456 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/navigation-state-changed 03-26 12:28:42.128 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :mailserver/check-connection-timeout ```
non_process
can t find in browser on simpledapp eth when refreshing the page bug report problem when refreshing browser page with simpledapp eth then it s often fails to find library while i could not reproduce it with usual opening dapp in browser it s reproduced of times when refreshing the page expected behavior refreshing the page and finds fine actual behavior img width alt screenshot at src reproduction open status and create new account open simpledapp eth in browser tap refresh button top right additional information status version develop nightly of march operating system android logs d viewrootimpl viewpostime pointer d viewrootimpl viewpostime pointer d reactnativejs debug handling re frame event set in d reactnativejs debug handling re frame event browser loading started d reactnativejs debug handling re frame event set in d reactnativejs debug handling re frame event browser navigation state changed d reactnativejs debug handling re frame event mailserver check connection timeout
0
1,534
4,120,220,867
IssuesEvent
2016-06-08 17:12:25
nodejs/node
https://api.github.com/repos/nodejs/node
closed
execFile arguments appear to be in the wrong order
child_process
* **Version**: 6.0.0 * **Platform**: MacOS Reading the docs for execFile, I expected the invocation with environment variables to look something like: `var ls = childProcess.execFile(path, args, {env: {SOME_ENV_VARIABLE: "somethign"}});` However, this signature did not work. Instead, we found both: `var ls = childProcess.execFile(path, args, function() {}, {env: {SOME_ENV_VARIABLE: "somethign"}});` and `var ls = childProcess.execFile(path, args, [], {env: {SOME_ENV_VARIABLE: "somethign"}});` worked. Am I misreading the docs, or is the behavior misdocumented?
1.0
execFile arguments appear to be in the wrong order - * **Version**: 6.0.0 * **Platform**: MacOS Reading the docs for execFile, I expected the invocation with environment variables to look something like: `var ls = childProcess.execFile(path, args, {env: {SOME_ENV_VARIABLE: "somethign"}});` However, this signature did not work. Instead, we found both: `var ls = childProcess.execFile(path, args, function() {}, {env: {SOME_ENV_VARIABLE: "somethign"}});` and `var ls = childProcess.execFile(path, args, [], {env: {SOME_ENV_VARIABLE: "somethign"}});` worked. Am I misreading the docs, or is the behavior misdocumented?
process
execfile arguments appear to be in the wrong order version platform macos reading the docs for execfile i expected the invocation with environment variables to look something like var ls childprocess execfile path args env some env variable somethign however this signature did not work instead we found both var ls childprocess execfile path args function env some env variable somethign and var ls childprocess execfile path args env some env variable somethign worked am i misreading the docs or is the behavior misdocumented
1
19,754
26,123,717,478
IssuesEvent
2022-12-28 15:39:29
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_fs (__main__.TestMultiprocessing)
module: multiprocessing triaged module: flaky-tests skipped
Platforms: asan, linux This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_fs%2C%20TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6297446390). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @VitalyFedyunin
1.0
DISABLED test_fs (__main__.TestMultiprocessing) - Platforms: asan, linux This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_fs%2C%20TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6297446390). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @VitalyFedyunin
process
disabled test fs main testmultiprocessing platforms asan linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with red and green cc vitalyfedyunin
1
604,349
18,682,315,818
IssuesEvent
2021-11-01 07:54:28
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
enewspaper.latimes.com - site is not usable
browser-firefox priority-normal engine-gecko
<!-- @browser: Firefox 93.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/91731 --> **URL**: https://enewspaper.latimes.com/desktop/latimes/default.aspx?token=42e23962a5d74614be16bae3d62d13e7&utm_id=39150&sfmc_id=3035949&edid=95390162-66b6-4b40-9414-680b46c0eb7a **Browser / Version**: Firefox 93.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: The site used to load with Firefox with no problem, now it does not but does if I switch to Chrome. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/5eb1c0fb-f144-462f-906c-aca31fbba112.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❀️_
1.0
enewspaper.latimes.com - site is not usable - <!-- @browser: Firefox 93.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/91731 --> **URL**: https://enewspaper.latimes.com/desktop/latimes/default.aspx?token=42e23962a5d74614be16bae3d62d13e7&utm_id=39150&sfmc_id=3035949&edid=95390162-66b6-4b40-9414-680b46c0eb7a **Browser / Version**: Firefox 93.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: The site used to load with Firefox with no problem, now it does not but does if I switch to Chrome. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/5eb1c0fb-f144-462f-906c-aca31fbba112.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❀️_
non_process
enewspaper latimes com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce the site used to load with firefox with no problem now it does not but does if i switch to chrome view the screenshot img alt screenshot src browser configuration none from with ❀️
0
9,940
12,974,088,184
IssuesEvent
2020-07-21 14:56:43
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
opened
Update OWNERs of central dashboard
area/community area/front-end kind/process priority/p1
/kind process The OWNERs file of the centraldashboard was last updated over a year ago. https://github.com/kubeflow/kubeflow/blob/master/components/centraldashboard/OWNERS It looks like at least @prodonjs has indicated [here](https://github.com/kubeflow/kubeflow/issues/3576#issuecomment-660055405) that he is no longer actively working on this. I think @avdaredevil is also primarily focused on other projects but ocassionally pitches in. @swiftdiaries are you still actively working on centraldashboard or willing to? It looks like we may need to find an appropriate set of owners in order to continue to drive this forward.
1.0
Update OWNERs of central dashboard - /kind process The OWNERs file of the centraldashboard was last updated over a year ago. https://github.com/kubeflow/kubeflow/blob/master/components/centraldashboard/OWNERS It looks like at least @prodonjs has indicated [here](https://github.com/kubeflow/kubeflow/issues/3576#issuecomment-660055405) that he is no longer actively working on this. I think @avdaredevil is also primarily focused on other projects but ocassionally pitches in. @swiftdiaries are you still actively working on centraldashboard or willing to? It looks like we may need to find an appropriate set of owners in order to continue to drive this forward.
process
update owners of central dashboard kind process the owners file of the centraldashboard was last updated over a year ago it looks like at least prodonjs has indicated that he is no longer actively working on this i think avdaredevil is also primarily focused on other projects but ocassionally pitches in swiftdiaries are you still actively working on centraldashboard or willing to it looks like we may need to find an appropriate set of owners in order to continue to drive this forward
1
518,376
15,027,646,564
IssuesEvent
2021-02-02 01:15:42
Conteo-Inc/conteo
https://api.github.com/repos/Conteo-Inc/conteo
closed
FR-M2: Create Match List
0_Frontend 0_Priority 1
Communicate with @sb3958 about the design requirements. While the list itself does not need to be a separate component, the list items need at least one separate component. Do not include the request, fill with dummy data until <Linked Issue> is completed.
1.0
FR-M2: Create Match List - Communicate with @sb3958 about the design requirements. While the list itself does not need to be a separate component, the list items need at least one separate component. Do not include the request, fill with dummy data until <Linked Issue> is completed.
non_process
fr create match list communicate with about the design requirements while the list itself does not need to be a separate component the list items need at least one separate component do not include the request fill with dummy data until is completed
0
161,896
25,415,093,448
IssuesEvent
2022-11-22 22:56:38
WordPress/pattern-directory
https://api.github.com/repos/WordPress/pattern-directory
opened
Create a new "featured" view
[Component] Theme [Status] Needs Design
As mentioned in https://github.com/WordPress/gutenberg/issues/44501, there should be a "featured" tag for the pattern directory. Currently we have a "featured" category, which could work for this, or we could take the opportunity to switch this over to an internal keyword (since it's not _really_ meant to be used in the same way as the other categories), as you should be able to filter featured by category. - [ ] Create a new keyword term "featured", and apply the new term to patterns by wordpressdotorg - [ ] Add a β€œfeatured” toggle to the filter navigation (needs design) - [ ] Support showing only featured posts in categories, homepage, etc Tech noteβ€” currently, we've done a lot of permalink rewriting to allow complex filtering of patterns with "pretty" urls. Adding in this filter will make it even more complex, so I would stick with a query parameter instead of trying to make the URLs pretty (ex, leave it as `/patterns/categories/header/?featured=true` rather than trying to inject `/featured/` in there somewhere). @jasmussen I've added this issue separately as a "needs design", but you can combine it with the design for #395 if that's easier. There are two other design/experience questions here though: Should the front page show all patterns or only featured by default? Is there a use case for showing just β€œcommunity”, or should it be β€œAll | Featured”?
1.0
Create a new "featured" view - As mentioned in https://github.com/WordPress/gutenberg/issues/44501, there should be a "featured" tag for the pattern directory. Currently we have a "featured" category, which could work for this, or we could take the opportunity to switch this over to an internal keyword (since it's not _really_ meant to be used in the same way as the other categories), as you should be able to filter featured by category. - [ ] Create a new keyword term "featured", and apply the new term to patterns by wordpressdotorg - [ ] Add a β€œfeatured” toggle to the filter navigation (needs design) - [ ] Support showing only featured posts in categories, homepage, etc Tech noteβ€” currently, we've done a lot of permalink rewriting to allow complex filtering of patterns with "pretty" urls. Adding in this filter will make it even more complex, so I would stick with a query parameter instead of trying to make the URLs pretty (ex, leave it as `/patterns/categories/header/?featured=true` rather than trying to inject `/featured/` in there somewhere). @jasmussen I've added this issue separately as a "needs design", but you can combine it with the design for #395 if that's easier. There are two other design/experience questions here though: Should the front page show all patterns or only featured by default? Is there a use case for showing just β€œcommunity”, or should it be β€œAll | Featured”?
non_process
create a new featured view as mentioned in there should be a featured tag for the pattern directory currently we have a featured category which could work for this or we could take the opportunity to switch this over to an internal keyword since it s not really meant to be used in the same way as the other categories as you should be able to filter featured by category create a new keyword term featured and apply the new term to patterns by wordpressdotorg add a β€œfeatured” toggle to the filter navigation needs design support showing only featured posts in categories homepage etc tech noteβ€” currently we ve done a lot of permalink rewriting to allow complex filtering of patterns with pretty urls adding in this filter will make it even more complex so i would stick with a query parameter instead of trying to make the urls pretty ex leave it as patterns categories header featured true rather than trying to inject featured in there somewhere jasmussen i ve added this issue separately as a needs design but you can combine it with the design for if that s easier there are two other design experience questions here though should the front page show all patterns or only featured by default is there a use case for showing just β€œcommunity” or should it be β€œall featured”
0
4,166
7,107,918,947
IssuesEvent
2018-01-16 21:45:53
18F/product-guide
https://api.github.com/repos/18F/product-guide
closed
UPDATE SECTION (Project Comms) - Dashboard
process change
Additional guidance to come from @gboone re: how PMs should use the dashboard per comment in googledoc - @awfrancisco: Any additional guidance you want to add for how PMs should use the Dashboard? @gboone: I am sure we will eventually but since we're up in the air on redesigning it ATM I'll hold off and update this when we have a real answer.
1.0
UPDATE SECTION (Project Comms) - Dashboard - Additional guidance to come from @gboone re: how PMs should use the dashboard per comment in googledoc - @awfrancisco: Any additional guidance you want to add for how PMs should use the Dashboard? @gboone: I am sure we will eventually but since we're up in the air on redesigning it ATM I'll hold off and update this when we have a real answer.
process
update section project comms dashboard additional guidance to come from gboone re how pms should use the dashboard per comment in googledoc awfrancisco any additional guidance you want to add for how pms should use the dashboard gboone i am sure we will eventually but since we re up in the air on redesigning it atm i ll hold off and update this when we have a real answer
1
14,895
18,291,056,115
IssuesEvent
2021-10-05 15:17:28
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Change Public Access Prevention "unspecified" to "inherited"
api: storage priority: p2 type: process
GCS is changing the `unspecified` public access prevention value to `inherited`. This will require the following changes: 1. Modify the values in the library. Constants/enum values for unspecified: these should be marked as deprecated. They can stay in the library if deletion would constitute a breaking change 2. Update the system tests. 3. Create a new sample with region tag public_access_prevention_inherited. So we don't break any documentation links, please do not delete the unspecified sample or sample test from your repo at this point in time. Please see nodejs example here: https://github.com/googleapis/nodejs-storage/pull/1614/files Please wait for approval from myself or @tritone before merging or releasing this change so we can coordinate around release dates. New inherited value should be rolled out to allowlisted testing projects this coming week; should be at 100% by 9/30
1.0
Change Public Access Prevention "unspecified" to "inherited" - GCS is changing the `unspecified` public access prevention value to `inherited`. This will require the following changes: 1. Modify the values in the library. Constants/enum values for unspecified: these should be marked as deprecated. They can stay in the library if deletion would constitute a breaking change 2. Update the system tests. 3. Create a new sample with region tag public_access_prevention_inherited. So we don't break any documentation links, please do not delete the unspecified sample or sample test from your repo at this point in time. Please see nodejs example here: https://github.com/googleapis/nodejs-storage/pull/1614/files Please wait for approval from myself or @tritone before merging or releasing this change so we can coordinate around release dates. New inherited value should be rolled out to allowlisted testing projects this coming week; should be at 100% by 9/30
process
change public access prevention unspecified to inherited gcs is changing the unspecified public access prevention value to inherited this will require the following changes modify the values in the library constants enum values for unspecified these should be marked as deprecated they can stay in the library if deletion would constitute a breaking change update the system tests create a new sample with region tag public access prevention inherited so we don t break any documentation links please do not delete the unspecified sample or sample test from your repo at this point in time please see nodejs example here please wait for approval from myself or tritone before merging or releasing this change so we can coordinate around release dates new inherited value should be rolled out to allowlisted testing projects this coming week should be at by
1
21,024
27,969,912,736
IssuesEvent
2023-03-25 00:17:32
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
rotated, mirrored cloning
feature: enhancement scope: image processing no-issue-activity
**Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is, e.g. "I'm always frustrated when [...]" --> Cloning is not easy if we have only a limited amount of sources to pick from, and a bigger area to cover / repair **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> my idea is, to manipulate the output of the cloned source to be more flexible: - mirroring the source input along the x/y axis, - rotating of the source input In this way, we avoid to see "repetitive patterns" when we have to pick multiple times from the same source The only problem which i see is, light and shadows may influence the cloned object, and the cloned object might look fake. But, if our cloned output needs only a slight rotation, the end result should look acceptable, imo. **Alternatives** <!-- A clear and concise description of any alternative solutions or features you've considered. -->
1.0
rotated, mirrored cloning - **Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is, e.g. "I'm always frustrated when [...]" --> Cloning is not easy if we have only a limited amount of sources to pick from, and a bigger area to cover / repair **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> my idea is, to manipulate the output of the cloned source to be more flexible: - mirroring the source input along the x/y axis, - rotating of the source input In this way, we avoid to see "repetitive patterns" when we have to pick multiple times from the same source The only problem which i see is, light and shadows may influence the cloned object, and the cloned object might look fake. But, if our cloned output needs only a slight rotation, the end result should look acceptable, imo. **Alternatives** <!-- A clear and concise description of any alternative solutions or features you've considered. -->
process
rotated mirrored cloning is your feature request related to a problem please describe cloning is not easy if we have only a limited amount of sources to pick from and a bigger area to cover repair describe the solution you d like my idea is to manipulate the output of the cloned source to be more flexible mirroring the source input along the x y axis rotating of the source input in this way we avoid to see repetitive patterns when we have to pick multiple times from the same source the only problem which i see is light and shadows may influence the cloned object and the cloned object might look fake but if our cloned output needs only a slight rotation the end result should look acceptable imo alternatives
1
493,418
14,231,703,352
IssuesEvent
2020-11-18 09:55:53
kubermatic/kubermatic
https://api.github.com/repos/kubermatic/kubermatic
closed
Make pod and service networks configurable
customer-request lifecycle/rotten priority/normal team/lifecycle
**User Story** The pod and service network cidr is part of the cluster CRD. Since the networks cannot be change for existing clusters, there is not effective way to configure the networks. Customers requested to use a different cidr as problems are expected when the pod / service networks overlap with other infrastructure. **Acceptance criteria** In best case, the pod and service cidr should be configurable per cluster. The option to change the defaults globally would also help.
1.0
Make pod and service networks configurable - **User Story** The pod and service network cidr is part of the cluster CRD. Since the networks cannot be change for existing clusters, there is not effective way to configure the networks. Customers requested to use a different cidr as problems are expected when the pod / service networks overlap with other infrastructure. **Acceptance criteria** In best case, the pod and service cidr should be configurable per cluster. The option to change the defaults globally would also help.
non_process
make pod and service networks configurable user story the pod and service network cidr is part of the cluster crd since the networks cannot be change for existing clusters there is not effective way to configure the networks customers requested to use a different cidr as problems are expected when the pod service networks overlap with other infrastructure acceptance criteria in best case the pod and service cidr should be configurable per cluster the option to change the defaults globally would also help
0
96,878
28,038,744,115
IssuesEvent
2023-03-28 16:50:41
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
reopened
evdi broken with linux 6.0
6.topic: kernel 6.topic: hardware 0.kind: build failure
### Steps To Reproduce ```nix boot.kernelPackages = pkgs.linuxPackages_6_0; services.xserver.videoDrivers = [ "modesetting" "displaylink" ]; ``` ### Build log ``` error: builder for '/nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv' failed with exit code 2; last 10 log lines: > 1060 | drm_framebuffer_get(&efb->base); > | ^~~~~~~~~~~~~~~~~~~ > make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_fb.o] Error 1 > cc1: some warnings being treated as errors > make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_painter.o] Error 1 > make[2]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/Makefile:1852: /build/source/module] Error 2 > make[2]: Leaving directory '/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/build' > make[1]: *** [Makefile:81: module] Error 2 > make[1]: Leaving directory '/build/source/module' > make: *** [Makefile:8: all] Error 2 For full logs, run 'nix log /nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv'. error: 1 dependencies of derivation '/nix/store/49lczfqfpp17y4vmpr04pkss0cwrb4pf-kernel-modules.drv' failed to build error: 1 dependencies of derivation '/nix/store/q6p42l3r50xv4jphgpd9bx81y32r7m6y-linux-6.0.7-modules.drv' failed to build error: 1 dependencies of derivation '/nix/store/pw5f02p73pvpj03a60pm0iysl88csdmn-nixos-system-Ryan-Laptop-22.11.20221108.0932685.drv' failed to build error: 1 dependencies of derivation '/nix/store/zs7ma4qvymdrb999cqk1hyqpziadb0yc-nixops-machines.drv' failed to build ``` ### Additional context ### Notify maintainers @eyjhb ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console [user@system:~]$ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.19.16, NixOS, 22.11 (Raccoon), 22.11.20221031.d40fea9` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.0` - channels(root): `"nixos"` - nixpkgs: `/nix/store/mrpx61vyrmhqhr2ms9svscyc1h523gf1-source` ```
1.0
evdi broken with linux 6.0 - ### Steps To Reproduce ```nix boot.kernelPackages = pkgs.linuxPackages_6_0; services.xserver.videoDrivers = [ "modesetting" "displaylink" ]; ``` ### Build log ``` error: builder for '/nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv' failed with exit code 2; last 10 log lines: > 1060 | drm_framebuffer_get(&efb->base); > | ^~~~~~~~~~~~~~~~~~~ > make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_fb.o] Error 1 > cc1: some warnings being treated as errors > make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_painter.o] Error 1 > make[2]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/Makefile:1852: /build/source/module] Error 2 > make[2]: Leaving directory '/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/build' > make[1]: *** [Makefile:81: module] Error 2 > make[1]: Leaving directory '/build/source/module' > make: *** [Makefile:8: all] Error 2 For full logs, run 'nix log /nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv'. error: 1 dependencies of derivation '/nix/store/49lczfqfpp17y4vmpr04pkss0cwrb4pf-kernel-modules.drv' failed to build error: 1 dependencies of derivation '/nix/store/q6p42l3r50xv4jphgpd9bx81y32r7m6y-linux-6.0.7-modules.drv' failed to build error: 1 dependencies of derivation '/nix/store/pw5f02p73pvpj03a60pm0iysl88csdmn-nixos-system-Ryan-Laptop-22.11.20221108.0932685.drv' failed to build error: 1 dependencies of derivation '/nix/store/zs7ma4qvymdrb999cqk1hyqpziadb0yc-nixops-machines.drv' failed to build ``` ### Additional context ### Notify maintainers @eyjhb ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console [user@system:~]$ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.19.16, NixOS, 22.11 (Raccoon), 22.11.20221031.d40fea9` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.0` - channels(root): `"nixos"` - nixpkgs: `/nix/store/mrpx61vyrmhqhr2ms9svscyc1h523gf1-source` ```
non_process
evdi broken with linux steps to reproduce nix boot kernelpackages pkgs linuxpackages services xserver videodrivers build log error builder for nix store evdi drv failed with exit code last log lines drm framebuffer get efb base make error some warnings being treated as errors make error make error make leaving directory nix store linux dev lib modules build make error make leaving directory build source module make error for full logs run nix log nix store evdi drv error dependencies of derivation nix store kernel modules drv failed to build error dependencies of derivation nix store linux modules drv failed to build error dependencies of derivation nix store nixos system ryan laptop drv failed to build error dependencies of derivation nix store nixops machines drv failed to build additional context notify maintainers eyjhb metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels root nixos nixpkgs nix store source
0
98,957
16,389,571,236
IssuesEvent
2021-05-17 14:34:39
Thanraj/linux-1
https://api.github.com/repos/Thanraj/linux-1
opened
CVE-2019-19241 (High) detected in linuxv5.0
security vulnerability
## CVE-2019-19241 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.4.2, the io_uring feature leads to requests that inadvertently have UID 0 and full capabilities, aka CID-181e448d8709. This is related to fs/io-wq.c, fs/io_uring.c, and net/socket.c. For example, an attacker can bypass intended restrictions on adding an IPv4 address to the loopback interface. This occurs because IORING_OP_SENDMSG operations, although requested in the context of an unprivileged user, are sometimes performed by a kernel worker thread without considering that context. <p>Publish Date: 2019-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19241>CVE-2019-19241</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241</a></p> <p>Release Date: 2019-12-17</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19241 (High) detected in linuxv5.0 - ## CVE-2019-19241 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.4.2, the io_uring feature leads to requests that inadvertently have UID 0 and full capabilities, aka CID-181e448d8709. This is related to fs/io-wq.c, fs/io_uring.c, and net/socket.c. For example, an attacker can bypass intended restrictions on adding an IPv4 address to the loopback interface. This occurs because IORING_OP_SENDMSG operations, although requested in the context of an unprivileged user, are sometimes performed by a kernel worker thread without considering that context. <p>Publish Date: 2019-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19241>CVE-2019-19241</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241</a></p> <p>Release Date: 2019-12-17</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details in the linux kernel before the io uring feature leads to requests that inadvertently have uid and full capabilities aka cid this is related to fs io wq c fs io uring c and net socket c for example an attacker can bypass intended restrictions on adding an address to the loopback interface this occurs because ioring op sendmsg operations although requested in the context of an unprivileged user are sometimes performed by a kernel worker thread without considering that context publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
667,319
22,465,441,359
IssuesEvent
2022-06-22 00:56:43
pusharray/new-collections
https://api.github.com/repos/pusharray/new-collections
closed
Reduce popup width and use standard layout for list items
feature priority: high effort: low
- In popup or narrow layout use same style for list item as in tab selector - When there is only 1 item display either hosname or title from first tab to prevent empty looking panel list - ...
1.0
Reduce popup width and use standard layout for list items - - In popup or narrow layout use same style for list item as in tab selector - When there is only 1 item display either hosname or title from first tab to prevent empty looking panel list - ...
non_process
reduce popup width and use standard layout for list items in popup or narrow layout use same style for list item as in tab selector when there is only item display either hosname or title from first tab to prevent empty looking panel list
0
16,595
21,647,948,965
IssuesEvent
2022-05-06 05:51:01
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Hybrid Worker Certificate Installation
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
Reviewing the .\Export-RunAsCertificateToHybridWorker, it is not clear as to how it would identify the hybrid worker where the certificate is to be installed. Given that is the case, it makes me hesitate to run the script since I do not know where it is going to try to install the certificate. Plus, what if you have multiple hybrid workers? Wouldn't you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a21ca143-2f33-5cea-94a8-ace7e9de5f9c * Version Independent ID: d7f2ef01-8c25-770e-dfd9-37b98dc7ba29 * Content: [Run Azure Automation runbooks on a Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hrw-run-runbooks#runbook-auth-managed-identities) * Content Source: [articles/automation/automation-hrw-run-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-hrw-run-runbooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **sudhirsneha**
1.0
Hybrid Worker Certificate Installation - Reviewing the .\Export-RunAsCertificateToHybridWorker, it is not clear as to how it would identify the hybrid worker where the certificate is to be installed. Given that is the case, it makes me hesitate to run the script since I do not know where it is going to try to install the certificate. Plus, what if you have multiple hybrid workers? Wouldn't you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a21ca143-2f33-5cea-94a8-ace7e9de5f9c * Version Independent ID: d7f2ef01-8c25-770e-dfd9-37b98dc7ba29 * Content: [Run Azure Automation runbooks on a Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hrw-run-runbooks#runbook-auth-managed-identities) * Content Source: [articles/automation/automation-hrw-run-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-hrw-run-runbooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **sudhirsneha**
process
hybrid worker certificate installation reviewing the export runascertificatetohybridworker it is not clear as to how it would identify the hybrid worker where the certificate is to be installed given that is the case it makes me hesitate to run the script since i do not know where it is going to try to install the certificate plus what if you have multiple hybrid workers wouldn t you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias sudhirsneha
1
1,949
4,771,091,234
IssuesEvent
2016-10-26 16:59:05
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Complete System.Diagnostics namespace
area-System.Diagnostics.Process netstandard2.0
These types are either missing or incomplete. - [x] T:System.Diagnostics.CorrelationManager - [x] T:System.Diagnostics.DebuggerStepperBoundaryAttribute - [x] T:System.Diagnostics.DebuggerVisualizerAttribute - [x] T:System.Diagnostics.MonitoringDescriptionAttribute - [x] T:System.Diagnostics.Process - [x] T:System.Diagnostics.ProcessModule - [x] T:System.Diagnostics.ProcessModuleCollection - [x] T:System.Diagnostics.ProcessThread - [x] T:System.Diagnostics.ProcessThreadCollection - [x] T:System.Diagnostics.ProcessWindowStyle - [x] T:System.Diagnostics.StackFrame - [x] T:System.Diagnostics.StackTrace - [x] T:System.Diagnostics.SwitchAttribute - [x] T:System.Diagnostics.SwitchLevelAttribute - [x] T:System.Diagnostics.TraceListener Some of these will need to throw PNSE, eg., many of the missing members of Process.
1.0
Complete System.Diagnostics namespace - These types are either missing or incomplete. - [x] T:System.Diagnostics.CorrelationManager - [x] T:System.Diagnostics.DebuggerStepperBoundaryAttribute - [x] T:System.Diagnostics.DebuggerVisualizerAttribute - [x] T:System.Diagnostics.MonitoringDescriptionAttribute - [x] T:System.Diagnostics.Process - [x] T:System.Diagnostics.ProcessModule - [x] T:System.Diagnostics.ProcessModuleCollection - [x] T:System.Diagnostics.ProcessThread - [x] T:System.Diagnostics.ProcessThreadCollection - [x] T:System.Diagnostics.ProcessWindowStyle - [x] T:System.Diagnostics.StackFrame - [x] T:System.Diagnostics.StackTrace - [x] T:System.Diagnostics.SwitchAttribute - [x] T:System.Diagnostics.SwitchLevelAttribute - [x] T:System.Diagnostics.TraceListener Some of these will need to throw PNSE, eg., many of the missing members of Process.
process
complete system diagnostics namespace these types are either missing or incomplete t system diagnostics correlationmanager t system diagnostics debuggerstepperboundaryattribute t system diagnostics debuggervisualizerattribute t system diagnostics monitoringdescriptionattribute t system diagnostics process t system diagnostics processmodule t system diagnostics processmodulecollection t system diagnostics processthread t system diagnostics processthreadcollection t system diagnostics processwindowstyle t system diagnostics stackframe t system diagnostics stacktrace t system diagnostics switchattribute t system diagnostics switchlevelattribute t system diagnostics tracelistener some of these will need to throw pnse eg many of the missing members of process
1
10,214
13,079,062,103
IssuesEvent
2020-08-01 01:46:38
kubeflow/kfserving
https://api.github.com/repos/kubeflow/kfserving
closed
Define KFServing related area labels
kind/process
It would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels. Labels should be added here: https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml In order to get added to all the Kubeflow repos automatically. Owners (people responsible for issue triage) can be listed in https://github.com/kubeflow/community/blob/master/labels-owners.yaml It would probably be good to define a top level label e.g. "area/kfserving". As well as possible subareas e.g. "area/kfserving/modelservers" as appropriate.
1.0
Define KFServing related area labels - It would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels. Labels should be added here: https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml In order to get added to all the Kubeflow repos automatically. Owners (people responsible for issue triage) can be listed in https://github.com/kubeflow/community/blob/master/labels-owners.yaml It would probably be good to define a top level label e.g. "area/kfserving". As well as possible subareas e.g. "area/kfserving/modelservers" as appropriate.
process
define kfserving related area labels it would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels labels should be added here in order to get added to all the kubeflow repos automatically owners people responsible for issue triage can be listed in it would probably be good to define a top level label e g area kfserving as well as possible subareas e g area kfserving modelservers as appropriate
1
5,637
2,957,575,958
IssuesEvent
2015-07-08 17:01:11
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
closed
Embed doc URLs in swagger documentation
area/cluster-api area/usability kind/documentation priority/P2 team/UX
We should embed links to github docs in the API documentation. I think there's a swagger field for that. /cc @nikhiljindal
1.0
Embed doc URLs in swagger documentation - We should embed links to github docs in the API documentation. I think there's a swagger field for that. /cc @nikhiljindal
non_process
embed doc urls in swagger documentation we should embed links to github docs in the api documentation i think there s a swagger field for that cc nikhiljindal
0
17,040
22,420,243,723
IssuesEvent
2022-06-20 01:42:26
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Lauren is Explorin'
suggested title in process
**Is your feature request related to a problem? Please describe.** [Watch classic episodes like Lauren is Explorin' is learning about Colors.] **Describe the solution you'd like** [TV-Y - 2 seasons - tv series, comedy] **Describe alternatives you've considered** Director: Craig McCracken As seen in: Foster's Home for Imaginary Friends (2004-2004) **Additional context** ![lauren is exploring](https://user-images.githubusercontent.com/88994668/129502354-5a1cdf4b-b070-4ff5-be42-723d1eef5950.png) ![lauren is exploring LOGO](https://user-images.githubusercontent.com/88994668/129502740-b8537dd5-1253-409c-82eb-2d42e09adb99.png)
1.0
Lauren is Explorin' - **Is your feature request related to a problem? Please describe.** [Watch classic episodes like Lauren is Explorin' is learning about Colors.] **Describe the solution you'd like** [TV-Y - 2 seasons - tv series, comedy] **Describe alternatives you've considered** Director: Craig McCracken As seen in: Foster's Home for Imaginary Friends (2004-2004) **Additional context** ![lauren is exploring](https://user-images.githubusercontent.com/88994668/129502354-5a1cdf4b-b070-4ff5-be42-723d1eef5950.png) ![lauren is exploring LOGO](https://user-images.githubusercontent.com/88994668/129502740-b8537dd5-1253-409c-82eb-2d42e09adb99.png)
process
lauren is explorin is your feature request related to a problem please describe describe the solution you d like describe alternatives you ve considered director craig mccracken as seen in foster s home for imaginary friends additional context
1
89,949
15,856,029,081
IssuesEvent
2021-04-08 01:20:50
akshat702/cart-ionic
https://api.github.com/repos/akshat702/cart-ionic
opened
CVE-2020-36048 (High) detected in engine.io-3.2.1.tgz
security vulnerability
## CVE-2020-36048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-3.2.1.tgz</b></p></summary> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz</a></p> <p>Path to dependency file: cart-ionic/cart/package.json</p> <p>Path to vulnerable library: cart-ionic/cart/e2e/node_modules/engine.io/package.json,cart-ionic/cart/e2e/node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - socket.io-2.1.1.tgz - :x: **engine.io-3.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport. <p>Publish Date: 2021-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p> <p>Release Date: 2021-01-08</p> <p>Fix Resolution: engine.io - 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36048 (High) detected in engine.io-3.2.1.tgz - ## CVE-2020-36048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-3.2.1.tgz</b></p></summary> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz</a></p> <p>Path to dependency file: cart-ionic/cart/package.json</p> <p>Path to vulnerable library: cart-ionic/cart/e2e/node_modules/engine.io/package.json,cart-ionic/cart/e2e/node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - socket.io-2.1.1.tgz - :x: **engine.io-3.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport. <p>Publish Date: 2021-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p> <p>Release Date: 2021-01-08</p> <p>Fix Resolution: engine.io - 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in engine io tgz cve high severity vulnerability vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file cart ionic cart package json path to vulnerable library cart ionic cart node modules engine io package json cart ionic cart node modules engine io package json dependency hierarchy karma tgz root library socket io tgz x engine io tgz vulnerable library vulnerability details engine io before allows attackers to cause a denial of service resource consumption via a post request to the long polling transport publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io step up your open source security game with whitesource
0
15,069
18,765,576,505
IssuesEvent
2021-11-05 23:15:01
esmero/strawberryfield
https://api.github.com/repos/esmero/strawberryfield
closed
Add SOLR SBF Flavor Document removal when File gets removed
JSON Preprocessors Events and Subscriber Typed Data and Search
# What? When we remove a file, old SBR documents could end being left in Solr. We may want to remove them automatically (using the Solr tracker) when a node and/or file gets removed. - Different ways of removing depending if the Node is removed v/s a File from a Node Important: Revision deletion triggers the same event as a full Node removal. Check that we do not remove everything if a revision is deleted.
1.0
Add SOLR SBF Flavor Document removal when File gets removed - # What? When we remove a file, old SBR documents could end being left in Solr. We may want to remove them automatically (using the Solr tracker) when a node and/or file gets removed. - Different ways of removing depending if the Node is removed v/s a File from a Node Important: Revision deletion triggers the same event as a full Node removal. Check that we do not remove everything if a revision is deleted.
process
add solr sbf flavor document removal when file gets removed what when we remove a file old sbr documents could end being left in solr we may want to remove them automatically using the solr tracker when a node and or file gets removed different ways of removing depending if the node is removed v s a file from a node important revision deletion triggers the same event as a full node removal check that we do not remove everything if a revision is deleted
1
20,554
27,211,219,303
IssuesEvent
2023-02-20 16:41:54
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
closed
[Entry] [5506.03.618]: Edit coordinate
entry webform delete-after-processing
Hello, I would like to add this coordinate to the roomfinder: ```yaml "5506.03.618": { lat: 48.266430538130976, lon: 11.668311542516562 } ```
1.0
[Entry] [5506.03.618]: Edit coordinate - Hello, I would like to add this coordinate to the roomfinder: ```yaml "5506.03.618": { lat: 48.266430538130976, lon: 11.668311542516562 } ```
process
edit coordinate hello i would like to add this coordinate to the roomfinder yaml lat lon
1
98,141
16,360,163,061
IssuesEvent
2021-05-14 08:13:26
Galaxy-Software-Service/Express-Demo
https://api.github.com/repos/Galaxy-Software-Service/Express-Demo
opened
CVE-2021-29469 (High) detected in redis-2.8.0.tgz
security vulnerability
## CVE-2021-29469 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis-2.8.0.tgz</b></p></summary> <p>Redis client library</p> <p>Library home page: <a href="https://registry.npmjs.org/redis/-/redis-2.8.0.tgz">https://registry.npmjs.org/redis/-/redis-2.8.0.tgz</a></p> <p>Path to dependency file: Express-Demo/package.json</p> <p>Path to vulnerable library: Express-Demo/node_modules/redis/package.json</p> <p> Dependency Hierarchy: - connect-redis-3.4.2.tgz (Root Library) - :x: **redis-2.8.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Express-Demo/commit/41acf582b4ea1792c35f65c02df57132590f1702">41acf582b4ea1792c35f65c02df57132590f1702</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1. <p>Publish Date: 2021-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469>CVE-2021-29469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3">https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3</a></p> <p>Release Date: 2021-04-23</p> <p>Fix Resolution: redis - 3.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"redis","packageVersion":"2.8.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect-redis:3.4.2;redis:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"redis - 3.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29469","vulnerabilityDetails":"Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-29469 (High) detected in redis-2.8.0.tgz - ## CVE-2021-29469 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis-2.8.0.tgz</b></p></summary> <p>Redis client library</p> <p>Library home page: <a href="https://registry.npmjs.org/redis/-/redis-2.8.0.tgz">https://registry.npmjs.org/redis/-/redis-2.8.0.tgz</a></p> <p>Path to dependency file: Express-Demo/package.json</p> <p>Path to vulnerable library: Express-Demo/node_modules/redis/package.json</p> <p> Dependency Hierarchy: - connect-redis-3.4.2.tgz (Root Library) - :x: **redis-2.8.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Express-Demo/commit/41acf582b4ea1792c35f65c02df57132590f1702">41acf582b4ea1792c35f65c02df57132590f1702</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1. <p>Publish Date: 2021-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469>CVE-2021-29469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3">https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3</a></p> <p>Release Date: 2021-04-23</p> <p>Fix Resolution: redis - 3.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"redis","packageVersion":"2.8.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect-redis:3.4.2;redis:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"redis - 3.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29469","vulnerabilityDetails":"Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in redis tgz cve high severity vulnerability vulnerable library redis tgz redis client library library home page a href path to dependency file express demo package json path to vulnerable library express demo node modules redis package json dependency hierarchy connect redis tgz root library x redis tgz vulnerable library found in head commit a href found in base branch master vulnerability details node redis is a node js redis client before version when a client is in monitoring mode the regex begin used to detected monitor messages could cause exponential backtracking on some strings this issue could lead to a denial of service the issue is patched in version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution redis isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree connect redis redis isminimumfixversionavailable true minimumfixversion redis basebranches vulnerabilityidentifier cve vulnerabilitydetails node redis is a node js redis client before version when a client is in monitoring mode the regex begin used to detected monitor messages could cause exponential backtracking on some strings this issue could lead to a denial of service the issue is patched in version vulnerabilityurl
0
15,171
18,942,926,941
IssuesEvent
2021-11-18 06:33:49
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371>
high priority triage review module: multiprocessing
## πŸ› Bug <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: run a bert model based simcse traing. <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> 67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 11156/16617 [1:04:38<31:12, 2.92it/s]Traceback (most recent call last): File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/queues.py", line 234, in _feed obj = _ForkingPickler.dumps(obj) File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/reduction.py", line 51, in dumps cls(buf, protocol).dump(obj) File "/root/miniconda3/envs/se/lib/python3.6/site-packages/torch/multiprocessing/reductions.py", line 328, in reduce_storage fd, size = storage._share_fd_() RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> In fact, the error only appeared today. Before, I could train and save the model normally. ## Environment Collecting environment information... PyTorch version: 1.10.0+cu102 Is debug build: False CUDA used to build PyTorch: 10.2 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.3 LTS (x86_64) GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 Clang version: Could not collect CMake version: version 3.10.2 Libc version: glibc-2.17 Python version: 3.6.13 |Anaconda, Inc.| (default, Jun 4 2021, 14:25:59) [GCC 7.5.0] (64-bit runtime) Python platform: Linux-3.10.0-957.el7.x86_64-x86_64-with-debian-buster-sid Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce RTX 2080 Ti GPU 1: GeForce RTX 2080 Ti GPU 2: GeForce RTX 2080 Ti GPU 3: GeForce RTX 2080 Ti GPU 4: GeForce RTX 2080 Ti GPU 5: GeForce RTX 2080 Ti GPU 6: GeForce RTX 2080 Ti GPU 7: GeForce RTX 2080 Ti Nvidia driver version: 440.33.01 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.5 [pip3] torch==1.10.0 [pip3] torchaudio==0.10.0 [pip3] torchvision==0.11.1 [conda] cudatoolkit 10.1.243 h6bb024c_0 [conda] faiss-gpu 1.7.1 py3.6_hf4b0e32_1_cuda10.1 pytorch [conda] libfaiss 1.7.1 hbf11de5_1_cuda10.1 pytorch [conda] mkl 2021.3.0 h06a4308_520 [conda] numpy 1.19.5 pypi_0 pypi [conda] numpy-base 1.17.0 py36h2f8d375_0 [conda] torch 1.10.0 pypi_0 pypi [conda] torchaudio 0.10.0 pypi_0 pypi [conda] torchvision 0.11.1 pypi_0 pypi ## Additional context <!-- Add any other context about the problem here. --> cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @VitalyFedyunin
1.0
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371> - ## πŸ› Bug <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: run a bert model based simcse traing. <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> 67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 11156/16617 [1:04:38<31:12, 2.92it/s]Traceback (most recent call last): File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/queues.py", line 234, in _feed obj = _ForkingPickler.dumps(obj) File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/reduction.py", line 51, in dumps cls(buf, protocol).dump(obj) File "/root/miniconda3/envs/se/lib/python3.6/site-packages/torch/multiprocessing/reductions.py", line 328, in reduce_storage fd, size = storage._share_fd_() RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> In fact, the error only appeared today. Before, I could train and save the model normally. ## Environment Collecting environment information... PyTorch version: 1.10.0+cu102 Is debug build: False CUDA used to build PyTorch: 10.2 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.3 LTS (x86_64) GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 Clang version: Could not collect CMake version: version 3.10.2 Libc version: glibc-2.17 Python version: 3.6.13 |Anaconda, Inc.| (default, Jun 4 2021, 14:25:59) [GCC 7.5.0] (64-bit runtime) Python platform: Linux-3.10.0-957.el7.x86_64-x86_64-with-debian-buster-sid Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce RTX 2080 Ti GPU 1: GeForce RTX 2080 Ti GPU 2: GeForce RTX 2080 Ti GPU 3: GeForce RTX 2080 Ti GPU 4: GeForce RTX 2080 Ti GPU 5: GeForce RTX 2080 Ti GPU 6: GeForce RTX 2080 Ti GPU 7: GeForce RTX 2080 Ti Nvidia driver version: 440.33.01 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.5 [pip3] torch==1.10.0 [pip3] torchaudio==0.10.0 [pip3] torchvision==0.11.1 [conda] cudatoolkit 10.1.243 h6bb024c_0 [conda] faiss-gpu 1.7.1 py3.6_hf4b0e32_1_cuda10.1 pytorch [conda] libfaiss 1.7.1 hbf11de5_1_cuda10.1 pytorch [conda] mkl 2021.3.0 h06a4308_520 [conda] numpy 1.19.5 pypi_0 pypi [conda] numpy-base 1.17.0 py36h2f8d375_0 [conda] torch 1.10.0 pypi_0 pypi [conda] torchaudio 0.10.0 pypi_0 pypi [conda] torchvision 0.11.1 pypi_0 pypi ## Additional context <!-- Add any other context about the problem here. --> cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @VitalyFedyunin
process
runtimeerror falseinternal assert failed at aten src aten mapallocator cpp please report a bug to pytorch unable to write to file πŸ› bug to reproduce steps to reproduce the behavior run a bert model based simcse traing β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– traceback most recent call last file root envs se lib multiprocessing queues py line in feed obj forkingpickler dumps obj file root envs se lib multiprocessing reduction py line in dumps cls buf protocol dump obj file root envs se lib site packages torch multiprocessing reductions py line in reduce storage fd size storage share fd runtimeerror falseinternal assert failed at aten src aten mapallocator cpp please report a bug to pytorch unable to write to file expected behavior in fact the error only appeared today before i could train and save the model normally environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version anaconda inc default jun bit runtime python platform linux with debian buster sid is cuda available true cuda runtime version could not collect gpu models and configuration gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti nvidia driver version cudnn version usr lib linux gnu libcudnn so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy torch torchaudio torchvision cudatoolkit faiss gpu pytorch libfaiss pytorch mkl numpy pypi pypi numpy base torch pypi pypi torchaudio pypi pypi torchvision pypi pypi additional context cc ezyang gchanan bdhirsh jbschlosser vitalyfedyunin
1