Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
18,897
| 2,616,011,931
|
IssuesEvent
|
2015-03-02 00:54:56
|
jasonhall/bwapi
|
https://api.github.com/repos/jasonhall/bwapi
|
closed
|
Create functionality that can disable starcraft rendering / GUI
|
auto-migrated Priority-Low Type-Enhancement
|
```
Create a function that can disable/enable the starcraft GUI and rendering code.
This will be useful for speeding up starcraft games.
```
Original issue reported on code.google.com by `lowerlo...@gmail.com` on 16 Oct 2010 at 3:40
|
1.0
|
Create functionality that can disable starcraft rendering / GUI - ```
Create a function that can disable/enable the starcraft GUI and rendering code.
This will be useful for speeding up starcraft games.
```
Original issue reported on code.google.com by `lowerlo...@gmail.com` on 16 Oct 2010 at 3:40
|
non_process
|
create functionality that can disable starcraft rendering gui create a function that can disable enable the starcraft gui and rendering code this will be useful for speeding up starcraft games original issue reported on code google com by lowerlo gmail com on oct at
| 0
|
74,460
| 20,167,167,746
|
IssuesEvent
|
2022-02-10 06:27:48
|
buildingSMART/IFC4.3.x-development
|
https://api.github.com/repos/buildingSMART/IFC4.3.x-development
|
opened
|
Is ParameterTakesPrecedence for IfcWindowType and IfcDoorType still a thing?
|
allocated-core proposal allocated-building
|
Originally brought up [here](https://forums.buildingsmart.org/t/using-parametertakesprecedence-in-ifcwindowtype-and-ifcdoortype/2574).
I personally have never seen ParameterTakesPrecedence used.
In the implementer agreement https://standards.buildingsmart.org/documents/Implementation/IFC_Implementation_Agreements/CV-2x3-181.html it basically states that it should always be false.
I propose to mark it as deprecated.
|
1.0
|
Is ParameterTakesPrecedence for IfcWindowType and IfcDoorType still a thing? - Originally brought up [here](https://forums.buildingsmart.org/t/using-parametertakesprecedence-in-ifcwindowtype-and-ifcdoortype/2574).
I personally have never seen ParameterTakesPrecedence used.
In the implementer agreement https://standards.buildingsmart.org/documents/Implementation/IFC_Implementation_Agreements/CV-2x3-181.html it basically states that it should always be false.
I propose to mark it as deprecated.
|
non_process
|
is parametertakesprecedence for ifcwindowtype and ifcdoortype still a thing originally brought up i personally have never seen parametertakesprecedence used in the implementer agreement it basically states that it should always be false i propose to mark it as deprecated
| 0
|
542,222
| 15,857,304,261
|
IssuesEvent
|
2021-04-08 04:28:10
|
ankidroid/Anki-Android
|
https://api.github.com/repos/ankidroid/Anki-Android
|
closed
|
[Bug] Notification service crash the app if it runs during sync
|
2.15 Bug Priority-High
|
###### Reproduction Steps
0. Turn on notification
1. Reinstall AnkiDroid
2. Start ankidroid
3. Sync
###### Expected Result
Syncing.
A notification.
Not crashing
###### Actual Result
Crash with the stacktrace:
> 2021-03-18 04:53:05.640 22370-22370/com.ichi2.anki E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.ichi2.anki, PID: 22370
java.lang.RuntimeException: Unable to start receiver com.ichi2.anki.services.ReminderService: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference
at android.app.ActivityThread.handleReceiver(ActivityThread.java:4207)
at android.app.ActivityThread.access$1600(ActivityThread.java:273)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)
at android.os.Handler.dispatchMessage(Handler.java:107)
at android.os.Looper.loop(Looper.java:237)
at android.app.ActivityThread.main(ActivityThread.java:8107)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)
Caused by: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference
at com.ichi2.libanki.sched.SchedV2._lrnForDeck(SchedV2.java:1478)
at com.ichi2.libanki.sched.SchedV2.deckDueList(SchedV2.java:552)
at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:592)
at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:587)
at com.ichi2.anki.services.ReminderService.getDeckOptionDue(ReminderService.java:172)
at com.ichi2.anki.services.ReminderService.onReceive(ReminderService.java:110)
at android.app.ActivityThread.handleReceiver(ActivityThread.java:4198)
at android.app.ActivityThread.access$1600(ActivityThread.java:273)Β
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)Β
at android.os.Handler.dispatchMessage(Handler.java:107)Β
at android.os.Looper.loop(Looper.java:237)Β
at android.app.ActivityThread.main(ActivityThread.java:8107)Β
at java.lang.reflect.Method.invoke(Native Method)Β
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)Β
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)Β
My understanding is that the notification service started running. However, sync closed the database, which means that notification suddenly can't access the database anymore
One way to deal with it would be to try/catch the NullPointerException here and state that notification is not so important that it should crash the program. Or that sync interact with notification, to pause notification until database is reoladed
|
1.0
|
[Bug] Notification service crash the app if it runs during sync - ###### Reproduction Steps
0. Turn on notification
1. Reinstall AnkiDroid
2. Start ankidroid
3. Sync
###### Expected Result
Syncing.
A notification.
Not crashing
###### Actual Result
Crash with the stacktrace:
> 2021-03-18 04:53:05.640 22370-22370/com.ichi2.anki E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.ichi2.anki, PID: 22370
java.lang.RuntimeException: Unable to start receiver com.ichi2.anki.services.ReminderService: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference
at android.app.ActivityThread.handleReceiver(ActivityThread.java:4207)
at android.app.ActivityThread.access$1600(ActivityThread.java:273)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)
at android.os.Handler.dispatchMessage(Handler.java:107)
at android.os.Looper.loop(Looper.java:237)
at android.app.ActivityThread.main(ActivityThread.java:8107)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)
Caused by: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.ichi2.libanki.DB.queryScalar(java.lang.String, java.lang.Object[])' on a null object reference
at com.ichi2.libanki.sched.SchedV2._lrnForDeck(SchedV2.java:1478)
at com.ichi2.libanki.sched.SchedV2.deckDueList(SchedV2.java:552)
at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:592)
at com.ichi2.libanki.sched.SchedV2.deckDueTree(SchedV2.java:587)
at com.ichi2.anki.services.ReminderService.getDeckOptionDue(ReminderService.java:172)
at com.ichi2.anki.services.ReminderService.onReceive(ReminderService.java:110)
at android.app.ActivityThread.handleReceiver(ActivityThread.java:4198)
at android.app.ActivityThread.access$1600(ActivityThread.java:273)Β
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:2107)Β
at android.os.Handler.dispatchMessage(Handler.java:107)Β
at android.os.Looper.loop(Looper.java:237)Β
at android.app.ActivityThread.main(ActivityThread.java:8107)Β
at java.lang.reflect.Method.invoke(Native Method)Β
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:496)Β
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1100)Β
My understanding is that the notification service started running. However, sync closed the database, which means that notification suddenly can't access the database anymore
One way to deal with it would be to try/catch the NullPointerException here and state that notification is not so important that it should crash the program. Or that sync interact with notification, to pause notification until database is reoladed
|
non_process
|
notification service crash the app if it runs during sync reproduction steps turn on notification reinstall ankidroid start ankidroid sync expected result syncing a notification not crashing actual result crash with the stacktrace com anki e androidruntime fatal exception main process com anki pid java lang runtimeexception unable to start receiver com anki services reminderservice java lang nullpointerexception attempt to invoke virtual method int com libanki db queryscalar java lang string java lang object on a null object reference at android app activitythread handlereceiver activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java caused by java lang nullpointerexception attempt to invoke virtual method int com libanki db queryscalar java lang string java lang object on a null object reference at com libanki sched lrnfordeck java at com libanki sched deckduelist java at com libanki sched deckduetree java at com libanki sched deckduetree java at com anki services reminderservice getdeckoptiondue reminderservice java at com anki services reminderservice onreceive reminderservice java at android app activitythread handlereceiver activitythread java at android app activitythread access activitythread java Β at android app activitythread h handlemessage activitythread java Β at android os handler dispatchmessage handler java Β at android os looper loop looper java Β at android app activitythread main activitythread java Β at java lang reflect method invoke native method Β at com android internal os runtimeinit methodandargscaller run runtimeinit java Β at com android internal os zygoteinit main zygoteinit java Β my understanding is that the notification service started running however sync closed the database which means that notification suddenly can t access the database anymore one way to deal with it would be to try catch the nullpointerexception here and state that notification is not so important that it should crash the program or that sync interact with notification to pause notification until database is reoladed
| 0
|
145,039
| 19,319,018,374
|
IssuesEvent
|
2021-12-14 01:50:26
|
jozseftiborcz/sast-eval-springboot1
|
https://api.github.com/repos/jozseftiborcz/sast-eval-springboot1
|
opened
|
CVE-2021-22096 (Medium) detected in spring-webmvc-5.2.2.RELEASE.jar, spring-web-5.2.2.RELEASE.jar
|
security vulnerability
|
## CVE-2021-22096 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-webmvc-5.2.2.RELEASE.jar</b>, <b>spring-web-5.2.2.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-webmvc-5.2.2.RELEASE.jar</b></p></summary>
<p>Spring Web MVC</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: sast-eval-springboot1/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library)
- :x: **spring-webmvc-5.2.2.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-web-5.2.2.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: sast-eval-springboot1/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-22096 (Medium) detected in spring-webmvc-5.2.2.RELEASE.jar, spring-web-5.2.2.RELEASE.jar - ## CVE-2021-22096 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-webmvc-5.2.2.RELEASE.jar</b>, <b>spring-web-5.2.2.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-webmvc-5.2.2.RELEASE.jar</b></p></summary>
<p>Spring Web MVC</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: sast-eval-springboot1/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/5.2.2.RELEASE/spring-webmvc-5.2.2.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library)
- :x: **spring-webmvc-5.2.2.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-web-5.2.2.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: sast-eval-springboot1/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in spring webmvc release jar spring web release jar cve medium severity vulnerability vulnerable libraries spring webmvc release jar spring web release jar spring webmvc release jar spring web mvc library home page a href path to dependency file sast eval build gradle path to vulnerable library home wss scanner repository org springframework spring webmvc release spring webmvc release jar home wss scanner repository org springframework spring webmvc release spring webmvc release jar dependency hierarchy spring boot starter web release jar root library x spring webmvc release jar vulnerable library spring web release jar spring web library home page a href path to dependency file sast eval pom xml path to vulnerable library home wss scanner repository org springframework spring web release spring web release jar home wss scanner repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in base branch master vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release org springframework spring web release org springframework spring webmvc release org springframework spring webflux release step up your open source security game with whitesource
| 0
|
143,971
| 5,533,498,813
|
IssuesEvent
|
2017-03-21 13:30:52
|
djoproject/pyshell
|
https://api.github.com/repos/djoproject/pyshell
|
closed
|
EVERYWHERE/make code compliant with python3
|
enhancement Middle priority To reCheck
|
### Description
Just execute pytest with a python3 interpreter and fix the issues until the test pass, then try to execute the whole software.
### Stuff to check/improve
- encoding #99
- decorator (pyshell.arg.decorator)
- ...
|
1.0
|
EVERYWHERE/make code compliant with python3 - ### Description
Just execute pytest with a python3 interpreter and fix the issues until the test pass, then try to execute the whole software.
### Stuff to check/improve
- encoding #99
- decorator (pyshell.arg.decorator)
- ...
|
non_process
|
everywhere make code compliant with description just execute pytest with a interpreter and fix the issues until the test pass then try to execute the whole software stuff to check improve encoding decorator pyshell arg decorator
| 0
|
53,615
| 13,185,437,429
|
IssuesEvent
|
2020-08-12 21:24:26
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
closed
|
FNX2-15613 β Rename mozilla_ui_publicsuffixlist to mozilla_lib_publicsuffixlist
|
eng:build good first issue help wanted
|
`Deps.mozilla_ui_publicsuffixlist` should be renamed to `Deps.mozilla_lib`_publicsuffixlist and updated in the build.gradle file. The corresponding dependency is named "lib-publicsuffixlist", so the current name is incorrect.
|
1.0
|
FNX2-15613 β Rename mozilla_ui_publicsuffixlist to mozilla_lib_publicsuffixlist - `Deps.mozilla_ui_publicsuffixlist` should be renamed to `Deps.mozilla_lib`_publicsuffixlist and updated in the build.gradle file. The corresponding dependency is named "lib-publicsuffixlist", so the current name is incorrect.
|
non_process
|
β rename mozilla ui publicsuffixlist to mozilla lib publicsuffixlist deps mozilla ui publicsuffixlist should be renamed to deps mozilla lib publicsuffixlist and updated in the build gradle file the corresponding dependency is named lib publicsuffixlist so the current name is incorrect
| 0
|
73,018
| 7,320,547,729
|
IssuesEvent
|
2018-03-02 07:53:03
|
TEAMMATES/teammates
|
https://api.github.com/repos/TEAMMATES/teammates
|
closed
|
FeedbackSessionQuestionTypeTest.json: fix malformed question metadata
|
a-Testing d.Contributors e.1 p.Low
|
Discovered while looking into https://github.com/TEAMMATES/teammates/issues/8415#issuecomment-364003271. The test data of one question is malformed (it's an MSQ question with `msqChoices` but `numOfMcqChoices` instead of `numOfMsqChoices` in the metadata):
https://github.com/TEAMMATES/teammates/blob/85faa7829f4edf19383b033e2d59825652b4b835/src/test/resources/data/FeedbackSessionQuestionTypeTest.json#L711
To investigate whether there are other similar cases, and why the deserialisation does not fail.
|
1.0
|
FeedbackSessionQuestionTypeTest.json: fix malformed question metadata - Discovered while looking into https://github.com/TEAMMATES/teammates/issues/8415#issuecomment-364003271. The test data of one question is malformed (it's an MSQ question with `msqChoices` but `numOfMcqChoices` instead of `numOfMsqChoices` in the metadata):
https://github.com/TEAMMATES/teammates/blob/85faa7829f4edf19383b033e2d59825652b4b835/src/test/resources/data/FeedbackSessionQuestionTypeTest.json#L711
To investigate whether there are other similar cases, and why the deserialisation does not fail.
|
non_process
|
feedbacksessionquestiontypetest json fix malformed question metadata discovered while looking into the test data of one question is malformed it s an msq question with msqchoices but numofmcqchoices instead of numofmsqchoices in the metadata to investigate whether there are other similar cases and why the deserialisation does not fail
| 0
|
17,679
| 23,512,656,480
|
IssuesEvent
|
2022-08-18 18:08:23
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
Port code samples of GAE Python bundled APIs from GCS
|
type: process api: appengine samples
|
## In which file did you encounter the issue?
N/A. I'm about to add a set of code samples.
### Did you change the file? If so, how?
Yes. `gsutil cp gs://python3-wormhole-api-samples/python3-wormhole-samples.zip .`
## Describe the issue
GAE Python bundled APIs went GA in April. The code samples for these APIs are still in a GCS bucket.
https://storage.googleapis.com/python3-wormhole-api-samples/python3-wormhole-samples.zip
This issue is to publish these samples to
[python-docs-samples](https://github.com/GoogleCloudPlatform/python-docs-samples)
|
1.0
|
Port code samples of GAE Python bundled APIs from GCS - ## In which file did you encounter the issue?
N/A. I'm about to add a set of code samples.
### Did you change the file? If so, how?
Yes. `gsutil cp gs://python3-wormhole-api-samples/python3-wormhole-samples.zip .`
## Describe the issue
GAE Python bundled APIs went GA in April. The code samples for these APIs are still in a GCS bucket.
https://storage.googleapis.com/python3-wormhole-api-samples/python3-wormhole-samples.zip
This issue is to publish these samples to
[python-docs-samples](https://github.com/GoogleCloudPlatform/python-docs-samples)
|
process
|
port code samples of gae python bundled apis from gcs in which file did you encounter the issue n a i m about to add a set of code samples did you change the file if so how yes gsutil cp gs wormhole api samples wormhole samples zip describe the issue gae python bundled apis went ga in april the code samples for these apis are still in a gcs bucket this issue is to publish these samples to
| 1
|
319,271
| 23,763,092,604
|
IssuesEvent
|
2022-09-01 10:30:30
|
celestiaorg/docs
|
https://api.github.com/repos/celestiaorg/docs
|
closed
|
Document Celestia's Reserved Namespaces
|
documentation
|
It's important for developers to know what Celestia's Reserved Namespaces are
|
1.0
|
Document Celestia's Reserved Namespaces - It's important for developers to know what Celestia's Reserved Namespaces are
|
non_process
|
document celestia s reserved namespaces it s important for developers to know what celestia s reserved namespaces are
| 0
|
1,316
| 3,867,055,676
|
IssuesEvent
|
2016-04-09 01:37:57
|
MaretEngineering/MROV
|
https://api.github.com/repos/MaretEngineering/MROV
|
closed
|
Properly initialize servo values
|
Necessary Addition Processing
|
The servos all start at 0 but they should start at the "center" value for the camera and open for the claw (make these constants because we will have to change them when we get the mechanical specs from @ddworken or @vishnuramasawmy)
Also, create variables (probably arrays) for the range of values that the servos operate on and replace those in the code.
|
1.0
|
Properly initialize servo values - The servos all start at 0 but they should start at the "center" value for the camera and open for the claw (make these constants because we will have to change them when we get the mechanical specs from @ddworken or @vishnuramasawmy)
Also, create variables (probably arrays) for the range of values that the servos operate on and replace those in the code.
|
process
|
properly initialize servo values the servos all start at but they should start at the center value for the camera and open for the claw make these constants because we will have to change them when we get the mechanical specs from ddworken or vishnuramasawmy also create variables probably arrays for the range of values that the servos operate on and replace those in the code
| 1
|
278,192
| 30,702,228,143
|
IssuesEvent
|
2023-07-27 01:13:08
|
nidhi7598/linux-4.1.15_CVE-2019-10220
|
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220
|
closed
|
CVE-2016-4485 (High) detected in linuxlinux-4.4.302 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2016-4485 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The llc_cmsg_rcv function in net/llc/af_llc.c in the Linux kernel before 4.5.5 does not initialize a certain data structure, which allows attackers to obtain sensitive information from kernel stack memory by reading a message.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4485>CVE-2016-4485</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4485">https://nvd.nist.gov/vuln/detail/CVE-2016-4485</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-4485 (High) detected in linuxlinux-4.4.302 - autoclosed - ## CVE-2016-4485 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/llc/af_llc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The llc_cmsg_rcv function in net/llc/af_llc.c in the Linux kernel before 4.5.5 does not initialize a certain data structure, which allows attackers to obtain sensitive information from kernel stack memory by reading a message.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4485>CVE-2016-4485</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4485">https://nvd.nist.gov/vuln/detail/CVE-2016-4485</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net llc af llc c net llc af llc c vulnerability details the llc cmsg rcv function in net llc af llc c in the linux kernel before does not initialize a certain data structure which allows attackers to obtain sensitive information from kernel stack memory by reading a message publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
15,952
| 20,171,691,257
|
IssuesEvent
|
2022-02-10 10:59:05
|
ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed
|
https://api.github.com/repos/ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed
|
opened
|
π Processing failed: KeyError
|
process
|
## Overview
`KeyError` found in `processing_task` task during run ended on 2022-02-10T10:59:05.295909.
## Details
Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed`
Task name: `processing_task`
Error type: `KeyError`
Error message: 'vadcp_eastward_seawater_velocity_dim_0'
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing
process_dataset(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset
append_to_zarr(mod_ds, store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 346, in append_to_zarr
dim_indexer, modify_zarr_dims = _validate_dims(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 76, in _validate_dims
existing_var = existing_zarr[dim]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 349, in __getitem__
raise KeyError(item)
KeyError: 'vadcp_eastward_seawater_velocity_dim_0'
```
</details>
|
1.0
|
π Processing failed: KeyError - ## Overview
`KeyError` found in `processing_task` task during run ended on 2022-02-10T10:59:05.295909.
## Details
Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_pd0_beam_parsed`
Task name: `processing_task`
Error type: `KeyError`
Error message: 'vadcp_eastward_seawater_velocity_dim_0'
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing
process_dataset(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset
append_to_zarr(mod_ds, store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 346, in append_to_zarr
dim_indexer, modify_zarr_dims = _validate_dims(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 76, in _validate_dims
existing_var = existing_zarr[dim]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 349, in __getitem__
raise KeyError(item)
KeyError: 'vadcp_eastward_seawater_velocity_dim_0'
```
</details>
|
process
|
π processing failed keyerror overview keyerror found in processing task task during run ended on details flow name streamed vadcp beam parsed task name processing task error type keyerror error message vadcp eastward seawater velocity dim traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing process dataset file srv conda envs notebook lib site packages ooi harvester processor init py line in process dataset append to zarr mod ds store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr dim indexer modify zarr dims validate dims file srv conda envs notebook lib site packages ooi harvester processor utils py line in validate dims existing var existing zarr file srv conda envs notebook lib site packages zarr hierarchy py line in getitem raise keyerror item keyerror vadcp eastward seawater velocity dim
| 1
|
695,035
| 23,841,273,886
|
IssuesEvent
|
2022-09-06 10:25:43
|
RinMinase/anidb
|
https://api.github.com/repos/RinMinase/anidb
|
closed
|
Add checkbox for filtering download list
|
Type: Feature Request Priority: Normal
|
- To not include or exclude watched
- By default exclude watched titles
|
1.0
|
Add checkbox for filtering download list - - To not include or exclude watched
- By default exclude watched titles
|
non_process
|
add checkbox for filtering download list to not include or exclude watched by default exclude watched titles
| 0
|
14,402
| 17,456,177,591
|
IssuesEvent
|
2021-08-06 01:47:13
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Rasterize mesh dataset tool: python error when running it in batch mode
|
Feedback stale Processing Bug Mesh
|
When running the said tool in batch mode (Right click>Execute as Batch Process) with several hundred processes (100s of lines), the tool crashes intermittently after successfully generating several, but not all of the output rasters. The batch processes are populated by a JSON file via the Open (Batch) button in the tool dialog. I am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully.
Here's a sample of the JSON batch file:
`[{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5038000, 5039000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5038.tif"}},
{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5039000, 5040000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5039.tif"}},
{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5040000, 5041000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5040.tif"}}]`
I am also able to run the same entire batch process successfully with the old Crayfish plugin version of the tool on QGIS 3.10 without errors/crashes.
Error message:
`An error has occurred while executing Python code:
Exception: unknown
Traceback (most recent call last):
File "C:/QGIS_L~1/apps/qgis/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 157, in runAlgorithm
results, ok = self.algorithm().run(parameters, context, multi_feedback)
Exception: unknown
Python version: 3.7.0 (v3.7.0:1bf9cc5093, Jun 27 2018, 04:59:51) [MSC v.1914 64 bit (AMD64)]
QGIS version: 3.18.0-ZΓΌrich ZΓΌrich, bdef9fb328
Python Path:
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins\LAStools
C:/QGIS_L~1/apps/qgis/./python
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins
C:/QGIS_L~1/apps/qgis/./python/plugins
C:\QGIS_L~1\apps\Python37
C:\QGIS_L~1\apps\Python37\Scripts
C:\QGIS_latest\bin\python37.zip
C:\QGIS_L~1\apps\Python37\DLLs
C:\QGIS_L~1\apps\Python37\lib
C:\QGIS_latest\bin
C:\QGIS_L~1\apps\Python37\lib\site-packages
C:\QGIS_L~1\apps\Python37\lib\site-packages\win32
C:\QGIS_L~1\apps\Python37\lib\site-packages\win32\lib
C:\QGIS_L~1\apps\Python37\lib\site-packages\Pythonwin
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python
C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\crayfish\gui\..\pyqtgraph-0.10.0-py2.py3-none-any.whl
C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\QuickMultiAttributeEdit3/forms`
QGIS version info:
`
QGIS version
3.18.0-ZΓΌrich
QGIS code revision
bdef9fb328
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
3.1.4
Running against GDAL/OGR
3.1.4
Compiled against GEOS
3.8.1-CAPI-1.13.3
Running against GEOS
3.8.1-CAPI-1.13.3
Compiled against SQLite
3.29.0
Running against SQLite
3.29.0
PostgreSQL Client Version
11.5
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
6.3.2
Running against PROJ
Rel. 6.3.2, May 1st, 2020
OS Version
Windows 10 (10.0)
Active python plugins
changeDataSource;
crayfish;
ImportPhotos;
LAStools;
latlontools;
LecoS;
Mergin;
profiletool;
qgis-maptiler-plugin;
QPackage;
QuickMultiAttributeEdit3;
quick_map_services;
splitmultipart;
valuetool;
db_manager;
processing
`
|
1.0
|
Rasterize mesh dataset tool: python error when running it in batch mode - When running the said tool in batch mode (Right click>Execute as Batch Process) with several hundred processes (100s of lines), the tool crashes intermittently after successfully generating several, but not all of the output rasters. The batch processes are populated by a JSON file via the Open (Batch) button in the tool dialog. I am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully.
Here's a sample of the JSON batch file:
`[{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5038000, 5039000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5038.tif"}},
{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5039000, 5040000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5039.tif"}},
{"PARAMETERS": {"INPUT": "'//path/to/my/mesh.2dm'", "DATASET_GROUPS": "[0]", "DATASET_TIME": "{'type': 'static'}", "EXTENT": "'300000, 301000, 5040000, 5041000'", "PIXEL_SIZE": "0.25", "CRS_OUTPUT": "QgsCoordinateReferenceSystem('EPSG:2950')"}, "OUTPUTS": {"OUTPUT": "//path/to/my/output/rasters/tuile_300-5040.tif"}}]`
I am also able to run the same entire batch process successfully with the old Crayfish plugin version of the tool on QGIS 3.10 without errors/crashes.
Error message:
`An error has occurred while executing Python code:
Exception: unknown
Traceback (most recent call last):
File "C:/QGIS_L~1/apps/qgis/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 157, in runAlgorithm
results, ok = self.algorithm().run(parameters, context, multi_feedback)
Exception: unknown
Python version: 3.7.0 (v3.7.0:1bf9cc5093, Jun 27 2018, 04:59:51) [MSC v.1914 64 bit (AMD64)]
QGIS version: 3.18.0-ZΓΌrich ZΓΌrich, bdef9fb328
Python Path:
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins\LAStools
C:/QGIS_L~1/apps/qgis/./python
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python/plugins
C:/QGIS_L~1/apps/qgis/./python/plugins
C:\QGIS_L~1\apps\Python37
C:\QGIS_L~1\apps\Python37\Scripts
C:\QGIS_latest\bin\python37.zip
C:\QGIS_L~1\apps\Python37\DLLs
C:\QGIS_L~1\apps\Python37\lib
C:\QGIS_latest\bin
C:\QGIS_L~1\apps\Python37\lib\site-packages
C:\QGIS_L~1\apps\Python37\lib\site-packages\win32
C:\QGIS_L~1\apps\Python37\lib\site-packages\win32\lib
C:\QGIS_L~1\apps\Python37\lib\site-packages\Pythonwin
C:/Users/mywinusername/AppData/Roaming/QGIS/QGIS3\profiles\default/python
C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\crayfish\gui\..\pyqtgraph-0.10.0-py2.py3-none-any.whl
C:\Users\mywinusername\AppData\Roaming\QGIS\QGIS3\profiles\default\python\plugins\QuickMultiAttributeEdit3/forms`
QGIS version info:
`
QGIS version
3.18.0-ZΓΌrich
QGIS code revision
bdef9fb328
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
3.1.4
Running against GDAL/OGR
3.1.4
Compiled against GEOS
3.8.1-CAPI-1.13.3
Running against GEOS
3.8.1-CAPI-1.13.3
Compiled against SQLite
3.29.0
Running against SQLite
3.29.0
PostgreSQL Client Version
11.5
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
6.3.2
Running against PROJ
Rel. 6.3.2, May 1st, 2020
OS Version
Windows 10 (10.0)
Active python plugins
changeDataSource;
crayfish;
ImportPhotos;
LAStools;
latlontools;
LecoS;
Mergin;
profiletool;
qgis-maptiler-plugin;
QPackage;
QuickMultiAttributeEdit3;
quick_map_services;
splitmultipart;
valuetool;
db_manager;
processing
`
|
process
|
rasterize mesh dataset tool python error when running it in batch mode when running the said tool in batch mode right click execute as batch process with several hundred processes of lines the tool crashes intermittently after successfully generating several but not all of the output rasters the batch processes are populated by a json file via the open batch button in the tool dialog i am able to rerun the batch iteratively starting from the last successfully generated output raster and all of the processes execute successfully here s a sample of the json batch file dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif parameters input path to my mesh dataset groups dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif parameters input path to my mesh dataset groups dataset time type static extent pixel size crs output qgscoordinatereferencesystem epsg outputs output path to my output rasters tuile tif i am also able to run the same entire batch process successfully with the old crayfish plugin version of the tool on qgis without errors crashes error message an error has occurred while executing python code exception unknown traceback most recent call last file c qgis l apps qgis python plugins processing gui batchalgorithmdialog py line in runalgorithm results ok self algorithm run parameters context multi feedback exception unknown python version jun qgis version zΓΌrich zΓΌrich python path c users mywinusername appdata roaming qgis profiles default python plugins lastools c qgis l apps qgis python c users mywinusername appdata roaming qgis profiles default python c users mywinusername appdata roaming qgis profiles default python plugins c qgis l apps qgis python plugins c qgis l apps c qgis l apps scripts c qgis latest bin zip c qgis l apps dlls c qgis l apps lib c qgis latest bin c qgis l apps lib site packages c qgis l apps lib site packages c qgis l apps lib site packages lib c qgis l apps lib site packages pythonwin c users mywinusername appdata roaming qgis profiles default python c users mywinusername appdata roaming qgis profiles default python plugins crayfish gui pyqtgraph none any whl c users mywinusername appdata roaming qgis profiles default python plugins forms qgis version info qgis version zΓΌrich qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins changedatasource crayfish importphotos lastools latlontools lecos mergin profiletool qgis maptiler plugin qpackage quick map services splitmultipart valuetool db manager processing
| 1
|
20,656
| 27,329,693,305
|
IssuesEvent
|
2023-02-25 13:13:25
|
cse442-at-ub/project_s23-cinco
|
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
|
closed
|
Connect to UB Servers and create a static "Hello World" website.
|
Processing Task Sprint 1
|
Test
1. User will go to this website:
https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/
2. Click on the hello.php file
3. Confirm the user can see the text "Hello World".
|
1.0
|
Connect to UB Servers and create a static "Hello World" website. - Test
1. User will go to this website:
https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/
2. Click on the hello.php file
3. Confirm the user can see the text "Hello World".
|
process
|
connect to ub servers and create a static hello world website test user will go to this website click on the hello php file confirm the user can see the text hello world
| 1
|
45,821
| 24,228,833,655
|
IssuesEvent
|
2022-09-26 16:24:56
|
Automattic/woocommerce-payments
|
https://api.github.com/repos/Automattic/woocommerce-payments
|
closed
|
Consider prefetching API data on page load
|
type: enhancement category: performance category: engineering
|
At the cost of skipping the loading placeholder view, we could **speed up loading time on initial page load** by pre<s>fetch</s>loading the data that will be needed by the view, in order to avoid round-trips on what I'd expect to be the most latency-prone leg of the stack, browser to site.
Currently, descriptors of what data is needed to populate the view are trapped deep inside the JS bundle, which must first be received and evaluated by the browser, making it inaccessible to the PHP server rendering piece.
However, we can write an adapter to derive required resources from the route and query params β see [prototype](https://github.com/Automattic/woocommerce-payments/commit/1c527064a43c4db1a866c8998579d6a72d1c1d2c) implementing one possible prefetching scheme for a few routes.
In many cases, the query params can be mapped directed to what the REST API handler expects (as is already done in client-side hooks [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/hooks.js#L33)] and resolvers [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/resolvers.js#L47)]). It looks like this will be good enough for all current routes except possibly when advanced filters are present, which I haven't attempted to prefetch due to the complexity of the mapping.
(It would be nice to have framework-level support for this but the routing logic [[code](https://github.com/woocommerce/woocommerce-admin/blob/d2fab15cb2b6e8cce8fd258e0d07580241334e9f/src/PageController.php#L116-L122)] doesn't currently even support patterns so we cannot do `/dispute/:dispute_id` even though the bundled client-side routing library fully supports it.)
Other considerations:
- One advantage of fetching the data from JS rather than PHP is that multiple requests are made in parallel β is there a way to do this from PHP without adding too much complexity?
- Since some of the same mapping logic would be implemented in the plugin PHP as in the browser code, we'd want to be sure they stayed in sync rather than drifting apart by accident β bidirectional references in code comments, perhaps
- Are there any good alternative ways of determining which requests will be needed by the view, or having the same specification be ingested in both contexts?
- Should the page query params be made to match the API query params more closely, to simplify the mapping? AFAIK the only hard limitation here is that other query params (`page`, in particular) are reserved, but we could make explicit exceptions for mapping those (just that one?)
- Would be worth doing some timing comparisons, primary consideration being the _time to load all data_, with _time to first UI render_ as another concern. (Some casual testing with a stopwatch _seems_ to indicate a quicker loading experience even with the deposit details screen preloading 3 resources in series)
|
True
|
Consider prefetching API data on page load - At the cost of skipping the loading placeholder view, we could **speed up loading time on initial page load** by pre<s>fetch</s>loading the data that will be needed by the view, in order to avoid round-trips on what I'd expect to be the most latency-prone leg of the stack, browser to site.
Currently, descriptors of what data is needed to populate the view are trapped deep inside the JS bundle, which must first be received and evaluated by the browser, making it inaccessible to the PHP server rendering piece.
However, we can write an adapter to derive required resources from the route and query params β see [prototype](https://github.com/Automattic/woocommerce-payments/commit/1c527064a43c4db1a866c8998579d6a72d1c1d2c) implementing one possible prefetching scheme for a few routes.
In many cases, the query params can be mapped directed to what the REST API handler expects (as is already done in client-side hooks [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/hooks.js#L33)] and resolvers [[code](https://github.com/Automattic/woocommerce-payments/blob/de82b68dee1e2233b8951065906c589ee8e0d3d4/client/data/transactions/resolvers.js#L47)]). It looks like this will be good enough for all current routes except possibly when advanced filters are present, which I haven't attempted to prefetch due to the complexity of the mapping.
(It would be nice to have framework-level support for this but the routing logic [[code](https://github.com/woocommerce/woocommerce-admin/blob/d2fab15cb2b6e8cce8fd258e0d07580241334e9f/src/PageController.php#L116-L122)] doesn't currently even support patterns so we cannot do `/dispute/:dispute_id` even though the bundled client-side routing library fully supports it.)
Other considerations:
- One advantage of fetching the data from JS rather than PHP is that multiple requests are made in parallel β is there a way to do this from PHP without adding too much complexity?
- Since some of the same mapping logic would be implemented in the plugin PHP as in the browser code, we'd want to be sure they stayed in sync rather than drifting apart by accident β bidirectional references in code comments, perhaps
- Are there any good alternative ways of determining which requests will be needed by the view, or having the same specification be ingested in both contexts?
- Should the page query params be made to match the API query params more closely, to simplify the mapping? AFAIK the only hard limitation here is that other query params (`page`, in particular) are reserved, but we could make explicit exceptions for mapping those (just that one?)
- Would be worth doing some timing comparisons, primary consideration being the _time to load all data_, with _time to first UI render_ as another concern. (Some casual testing with a stopwatch _seems_ to indicate a quicker loading experience even with the deposit details screen preloading 3 resources in series)
|
non_process
|
consider prefetching api data on page load at the cost of skipping the loading placeholder view we could speed up loading time on initial page load by pre fetch loading the data that will be needed by the view in order to avoid round trips on what i d expect to be the most latency prone leg of the stack browser to site currently descriptors of what data is needed to populate the view are trapped deep inside the js bundle which must first be received and evaluated by the browser making it inaccessible to the php server rendering piece however we can write an adapter to derive required resources from the route and query params β see implementing one possible prefetching scheme for a few routes in many cases the query params can be mapped directed to what the rest api handler expects as is already done in client side hooks and resolvers it looks like this will be good enough for all current routes except possibly when advanced filters are present which i haven t attempted to prefetch due to the complexity of the mapping it would be nice to have framework level support for this but the routing logic doesn t currently even support patterns so we cannot do dispute dispute id even though the bundled client side routing library fully supports it other considerations one advantage of fetching the data from js rather than php is that multiple requests are made in parallel β is there a way to do this from php without adding too much complexity since some of the same mapping logic would be implemented in the plugin php as in the browser code we d want to be sure they stayed in sync rather than drifting apart by accident β bidirectional references in code comments perhaps are there any good alternative ways of determining which requests will be needed by the view or having the same specification be ingested in both contexts should the page query params be made to match the api query params more closely to simplify the mapping afaik the only hard limitation here is that other query params page in particular are reserved but we could make explicit exceptions for mapping those just that one would be worth doing some timing comparisons primary consideration being the time to load all data with time to first ui render as another concern some casual testing with a stopwatch seems to indicate a quicker loading experience even with the deposit details screen preloading resources in series
| 0
|
485,330
| 13,964,181,530
|
IssuesEvent
|
2020-10-25 17:06:17
|
AY2021S1-CS2103T-T13-2/tp
|
https://api.github.com/repos/AY2021S1-CS2103T-T13-2/tp
|
closed
|
As an expert user, I want to be able to see the statistics of my overall performance of a certain tag
|
priority.High type.Story
|
Statistics aggregated by `tag`.
|
1.0
|
As an expert user, I want to be able to see the statistics of my overall performance of a certain tag - Statistics aggregated by `tag`.
|
non_process
|
as an expert user i want to be able to see the statistics of my overall performance of a certain tag statistics aggregated by tag
| 0
|
342,111
| 10,312,180,807
|
IssuesEvent
|
2019-08-29 19:12:28
|
inverse-inc/packetfence
|
https://api.github.com/repos/inverse-inc/packetfence
|
closed
|
Default portal is not created by default
|
Priority: Low Type: Bug
|
I see that in the logs:
Aug 29 08:40:28 unicorn packetfence: WARN pfperl-api(11281): Can't stat /usr/local/pf/html/captive-portal/profile-templates/default: No such file or directory
at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/ConnectionProfiles.pm line 393.
I wonder if we should create by default ?
|
1.0
|
Default portal is not created by default - I see that in the logs:
Aug 29 08:40:28 unicorn packetfence: WARN pfperl-api(11281): Can't stat /usr/local/pf/html/captive-portal/profile-templates/default: No such file or directory
at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/ConnectionProfiles.pm line 393.
I wonder if we should create by default ?
|
non_process
|
default portal is not created by default i see that in the logs aug unicorn packetfence warn pfperl api can t stat usr local pf html captive portal profile templates default no such file or directory at usr local pf lib pf unifiedapi controller config connectionprofiles pm line i wonder if we should create by default
| 0
|
40,255
| 10,478,844,346
|
IssuesEvent
|
2019-09-24 01:46:10
|
filecoin-project/go-filecoin
|
https://api.github.com/repos/filecoin-project/go-filecoin
|
closed
|
[bug] go run ./build/ lint warning messages after libp2p upgrade
|
A-build C-bug
|
**Describe the bug**
We updated the version of `github.com/golangci/golangci-lint` (v1.17.0) we use when upgrading go-libp2p. This update results in the following warning logs being returned when the linter is ran:
```
$ go run ./build/ lint
Linting ./... ...
go run github.com/golangci/golangci-lint/cmd/golangci-lint run
level=warning msg="[runner/golint] Golint: can't lint 3 files: no file name for file &{Doc:<nil> Package:6844383 Name:bls Decls:[0xc008dc4780 0xc008dc47c0 0xc008dc4800 0xc008dc48c0 0xc008dc4900 0xc008dc4940 0xc008dc4980 0xc008dc49c0 0xc008dc4a40] Scope:scope 0xc003ec3000 {\n\tconst SignatureBytes\n\tconst PublicKeyBytes\n\ttype PublicKey\n\ttype PrivateKey\n\ttype Message\n\ttype Digest\n\tconst PrivateKeyBytes\n\tconst DigestBytes\n\ttype Signature\n}\n Imports:[] Unresolved:[byte byte byte byte byte] Comments:[0xc007b9e2e0 0xc007b9e360 0xc007b9e3e0 0xc007b9e4a0 0xc007b9e520 0xc007b9e5c0 0xc007b9e660 0xc007b9e720 0xc007b9e7a0]}"
```
This was introduced in a1771a863590be1432979fea9dc98d5bc99e4d0e
**To Reproduce**
run:
`go run ./build/ lint`
**Expected behavior**
No warning logs are returned when the command is ran.
**Version information**
<!-- If you are having issue building go-filecoin please include the following: -->
- Go: go1.12.1
- Commit: a1771a863590be1432979fea9dc98d5bc99e4d0e
|
1.0
|
[bug] go run ./build/ lint warning messages after libp2p upgrade - **Describe the bug**
We updated the version of `github.com/golangci/golangci-lint` (v1.17.0) we use when upgrading go-libp2p. This update results in the following warning logs being returned when the linter is ran:
```
$ go run ./build/ lint
Linting ./... ...
go run github.com/golangci/golangci-lint/cmd/golangci-lint run
level=warning msg="[runner/golint] Golint: can't lint 3 files: no file name for file &{Doc:<nil> Package:6844383 Name:bls Decls:[0xc008dc4780 0xc008dc47c0 0xc008dc4800 0xc008dc48c0 0xc008dc4900 0xc008dc4940 0xc008dc4980 0xc008dc49c0 0xc008dc4a40] Scope:scope 0xc003ec3000 {\n\tconst SignatureBytes\n\tconst PublicKeyBytes\n\ttype PublicKey\n\ttype PrivateKey\n\ttype Message\n\ttype Digest\n\tconst PrivateKeyBytes\n\tconst DigestBytes\n\ttype Signature\n}\n Imports:[] Unresolved:[byte byte byte byte byte] Comments:[0xc007b9e2e0 0xc007b9e360 0xc007b9e3e0 0xc007b9e4a0 0xc007b9e520 0xc007b9e5c0 0xc007b9e660 0xc007b9e720 0xc007b9e7a0]}"
```
This was introduced in a1771a863590be1432979fea9dc98d5bc99e4d0e
**To Reproduce**
run:
`go run ./build/ lint`
**Expected behavior**
No warning logs are returned when the command is ran.
**Version information**
<!-- If you are having issue building go-filecoin please include the following: -->
- Go: go1.12.1
- Commit: a1771a863590be1432979fea9dc98d5bc99e4d0e
|
non_process
|
go run build lint warning messages after upgrade describe the bug we updated the version of github com golangci golangci lint we use when upgrading go this update results in the following warning logs being returned when the linter is ran go run build lint linting go run github com golangci golangci lint cmd golangci lint run level warning msg golint can t lint files no file name for file doc package name bls decls scope scope n tconst signaturebytes n tconst publickeybytes n ttype publickey n ttype privatekey n ttype message n ttype digest n tconst privatekeybytes n tconst digestbytes n ttype signature n n imports unresolved comments this was introduced in to reproduce run go run build lint expected behavior no warning logs are returned when the command is ran version information go commit
| 0
|
21,614
| 30,017,402,656
|
IssuesEvent
|
2023-06-26 19:54:19
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Measure and optimize pty host reconnection
|
feature-request verified perf terminal-process
|
Reconnecting/restoring terminals becomes quite slow when you have many terminals. We should measure what's happening with the pty host trace logs and try optimize.
|
1.0
|
Measure and optimize pty host reconnection - Reconnecting/restoring terminals becomes quite slow when you have many terminals. We should measure what's happening with the pty host trace logs and try optimize.
|
process
|
measure and optimize pty host reconnection reconnecting restoring terminals becomes quite slow when you have many terminals we should measure what s happening with the pty host trace logs and try optimize
| 1
|
17,879
| 23,832,320,577
|
IssuesEvent
|
2022-09-05 23:22:04
|
OctopusDeploy/Issues
|
https://api.github.com/repos/OctopusDeploy/Issues
|
closed
|
Out of date Tentacle and Calamari warnings are no longer useful
|
kind/enhancement feature/usability size/small feature/ops-processes p3 area/core team/fire-and-motion
|
Currently, users of Octopus Server using Machine Policy to automatically upgrade tentacles during deployments can see warnings regarding out of date Tentacle and Calamari instances. [Relevant documentation here](https://octopus.com/docs/infrastructure/machine-policies#MachinePolicies-ConfigurehowCalamariandTentacleareupdated)
As evidenced by customer queries, we are finding that these warnings do not hold much value anymore, now that the default behaviour is 'on demand' upgrades. For e.g.
```
As a user I want to be able to upgrade my Octopus server without negatively impacting deployment schedules nor do I want to cause undue consternation among the operations support folks.
```
The 'Healthy with warnings' icons shown in the deployment targets view (and other locations) are also causing some confusion.
## Proposed solution
We think that removing these warnings entirely will encourage use of the default behaviour, which carries better performance characteristics overall, and not make unnecessary alarm bells sound off for customers.
## Before

'Healthy with warnings' icons are displayed for such deployment targets.
## After

Such targets would not be labelled as 'Healthy with warnings'
## Workarounds
At this stage, ignore the warnings, or move to the 'Always up to date' model. The trade-off for the latter is that in large ecosystems, tentacle upgrades can take a significant amount of time due to a CPU bound limit on how many upgrades can be done at once.
## Links
[public support ticket](https://help.octopus.com/t/increase-number-of-concurrent-tentacle-updates/23368/6)
[internal discussion thread](https://octopusdeploy.slack.com/archives/C033W4273/p1561418888036100)
[related issue](https://github.com/OctopusDeploy/Issues/issues/4374)
# Prerequisites
- [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this feature request
- [x] I have tagged the issue appropriately (area/*, kind/enhancement)
|
1.0
|
Out of date Tentacle and Calamari warnings are no longer useful - Currently, users of Octopus Server using Machine Policy to automatically upgrade tentacles during deployments can see warnings regarding out of date Tentacle and Calamari instances. [Relevant documentation here](https://octopus.com/docs/infrastructure/machine-policies#MachinePolicies-ConfigurehowCalamariandTentacleareupdated)
As evidenced by customer queries, we are finding that these warnings do not hold much value anymore, now that the default behaviour is 'on demand' upgrades. For e.g.
```
As a user I want to be able to upgrade my Octopus server without negatively impacting deployment schedules nor do I want to cause undue consternation among the operations support folks.
```
The 'Healthy with warnings' icons shown in the deployment targets view (and other locations) are also causing some confusion.
## Proposed solution
We think that removing these warnings entirely will encourage use of the default behaviour, which carries better performance characteristics overall, and not make unnecessary alarm bells sound off for customers.
## Before

'Healthy with warnings' icons are displayed for such deployment targets.
## After

Such targets would not be labelled as 'Healthy with warnings'
## Workarounds
At this stage, ignore the warnings, or move to the 'Always up to date' model. The trade-off for the latter is that in large ecosystems, tentacle upgrades can take a significant amount of time due to a CPU bound limit on how many upgrades can be done at once.
## Links
[public support ticket](https://help.octopus.com/t/increase-number-of-concurrent-tentacle-updates/23368/6)
[internal discussion thread](https://octopusdeploy.slack.com/archives/C033W4273/p1561418888036100)
[related issue](https://github.com/OctopusDeploy/Issues/issues/4374)
# Prerequisites
- [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this feature request
- [x] I have tagged the issue appropriately (area/*, kind/enhancement)
|
process
|
out of date tentacle and calamari warnings are no longer useful currently users of octopus server using machine policy to automatically upgrade tentacles during deployments can see warnings regarding out of date tentacle and calamari instances as evidenced by customer queries we are finding that these warnings do not hold much value anymore now that the default behaviour is on demand upgrades for e g as a user i want to be able to upgrade my octopus server without negatively impacting deployment schedules nor do i want to cause undue consternation among the operations support folks the healthy with warnings icons shown in the deployment targets view and other locations are also causing some confusion proposed solution we think that removing these warnings entirely will encourage use of the default behaviour which carries better performance characteristics overall and not make unnecessary alarm bells sound off for customers before healthy with warnings icons are displayed for such deployment targets after such targets would not be labelled as healthy with warnings workarounds at this stage ignore the warnings or move to the always up to date model the trade off for the latter is that in large ecosystems tentacle upgrades can take a significant amount of time due to a cpu bound limit on how many upgrades can be done at once links prerequisites i have searched and issues to make sure it isn t already requested i have written a descriptive issue title i have linked the original source of this feature request i have tagged the issue appropriately area kind enhancement
| 1
|
350,819
| 31,932,351,254
|
IssuesEvent
|
2023-09-19 08:17:42
|
masters2023-project-06-second-hand/be-a
|
https://api.github.com/repos/masters2023-project-06-second-hand/be-a
|
opened
|
[BE] μ±ν
λ°© μμ±, μ±ν
μ°Έμ¬μ 보 μν°ν° μμ±
|
π¦test β¨ feature
|
## β¨ ν΄λΉ κΈ°λ₯μ ꡬννκΈ° μν΄ ν μΌμ΄ 무μμΈκ°μ?
- [ ] μ±ν
λ°© μ‘°ν/μμ±
- [ ] μ±ν
μ°Έμ¬μ 보 μν°ν°
- [ ]
## β
Test Case
- [ ]
- [ ]
|
1.0
|
[BE] μ±ν
λ°© μμ±, μ±ν
μ°Έμ¬μ 보 μν°ν° μμ± - ## β¨ ν΄λΉ κΈ°λ₯μ ꡬννκΈ° μν΄ ν μΌμ΄ 무μμΈκ°μ?
- [ ] μ±ν
λ°© μ‘°ν/μμ±
- [ ] μ±ν
μ°Έμ¬μ 보 μν°ν°
- [ ]
## β
Test Case
- [ ]
- [ ]
|
non_process
|
μ±ν
λ°© μμ± μ±ν
μ°Έμ¬μ 보 μν°ν° μμ± β¨ ν΄λΉ κΈ°λ₯μ ꡬννκΈ° μν΄ ν μΌμ΄ 무μμΈκ°μ μ±ν
λ°© μ‘°ν μμ± μ±ν
μ°Έμ¬μ 보 μν°ν° β
test case
| 0
|
10,090
| 13,044,162,044
|
IssuesEvent
|
2020-07-29 03:47:28
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `SubDatetimeAndDuration` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `SubDatetimeAndDuration` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `SubDatetimeAndDuration` from TiDB -
## Description
Port the scalar function `SubDatetimeAndDuration` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function subdatetimeandduration from tidb description port the scalar function subdatetimeandduration from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
1,873
| 4,699,340,790
|
IssuesEvent
|
2016-10-12 15:25:00
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
49 Kay Street
|
processed regression
|
Is it more important for the house number to match in the result or the street name?
We're seeing `49 De Kay Street` show up higher than `3`,`4`, and `5 KAY STREET`.
Separate issue here is that there are OA records with all caps coming through now, due to the fact that OA stopped normalizing their data for capitalization.
|
1.0
|
49 Kay Street - Is it more important for the house number to match in the result or the street name?
We're seeing `49 De Kay Street` show up higher than `3`,`4`, and `5 KAY STREET`.
Separate issue here is that there are OA records with all caps coming through now, due to the fact that OA stopped normalizing their data for capitalization.
|
process
|
kay street is it more important for the house number to match in the result or the street name we re seeing de kay street show up higher than and kay street separate issue here is that there are oa records with all caps coming through now due to the fact that oa stopped normalizing their data for capitalization
| 1
|
330,307
| 10,038,306,541
|
IssuesEvent
|
2019-07-18 14:54:13
|
Citykleta/web-app
|
https://api.github.com/repos/Citykleta/web-app
|
closed
|
Route exploration
|
component:Itinerary enhancement priority:medium
|
At the moment, a user can quickly find a route going from A to B going through various intermediate points.
It would be however important to improve this part of the application.
- [ ] change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown
- [ ] propose alternatives
- [ ] give a way to the user to preview the instructions
|
1.0
|
Route exploration - At the moment, a user can quickly find a route going from A to B going through various intermediate points.
It would be however important to improve this part of the application.
- [ ] change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown
- [ ] propose alternatives
- [ ] give a way to the user to preview the instructions
|
non_process
|
route exploration at the moment a user can quickly find a route going from a to b going through various intermediate points it would be however important to improve this part of the application change the style of the displayed route to convey the sens of the route as the moment it just displays the path but no direction is shown propose alternatives give a way to the user to preview the instructions
| 0
|
135,988
| 5,267,789,695
|
IssuesEvent
|
2017-02-05 02:40:20
|
TrinityCore/TrinityCore
|
https://api.github.com/repos/TrinityCore/TrinityCore
|
opened
|
[3.3.5] Extra lashtails in Barrens
|
Branch-3.3.5a Comp-Database Priority-VeryLow Sub-Spawns
|
[//]: # (***************************************************************)
[//]: # (** DON'T DELETE THIS TEMPLATE ELSE YOUR ISSUE WILL BE CLOSED **)
[//]: # (***************************************************************)
**Description:**
While looking at some videos of the game, on the subject of pooling. This one in particular at around this point:
https://youtu.be/xyhkTcJuOoE?t=168
There are some extra lashtails in the Barrens that don't belong there, for sure you would really not see them spawned like this in reality. As seen here:

The one that belongs can be seen to be roaming, the other 3 are static. The one that belongs is spawnId: 19816.
Those that don't are: 40493, 49495, 49496.
There is also another static one added 40494:

I can't say whether that one belongs or not. For sure it shouldn't be static.
Soliciting opinion before I remove them/enable movement on the last one.
**Current behaviour:** (Tell us what happens.)
Too many lashtails.
**Expected behaviour:** (Tell us what should happen instead.)
Not so many lashtails.
**Steps to reproduce the problem:**
1. .go creature 40493
2. Observe extra lashtails
3. ?????
4. Profit
**Branch(es):** 3.3.5 / master (Tell us which branch(es) this issue affects.)
3.3.5
**TC rev. hash/commit:**
e991a15dd7cbf5ba374f8942b7ea88f3a4d7b365
**TDB version:**
TDB_full_world_335.62_2016_10_17 + as per hash above.
**Operating system:**
Linux
[//]: # (This template is for problem reports. For other types of report, edit it accordingly.)
[//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
[//]: # (For fixes containing C++, create a Pull Request.)
|
1.0
|
[3.3.5] Extra lashtails in Barrens - [//]: # (***************************************************************)
[//]: # (** DON'T DELETE THIS TEMPLATE ELSE YOUR ISSUE WILL BE CLOSED **)
[//]: # (***************************************************************)
**Description:**
While looking at some videos of the game, on the subject of pooling. This one in particular at around this point:
https://youtu.be/xyhkTcJuOoE?t=168
There are some extra lashtails in the Barrens that don't belong there, for sure you would really not see them spawned like this in reality. As seen here:

The one that belongs can be seen to be roaming, the other 3 are static. The one that belongs is spawnId: 19816.
Those that don't are: 40493, 49495, 49496.
There is also another static one added 40494:

I can't say whether that one belongs or not. For sure it shouldn't be static.
Soliciting opinion before I remove them/enable movement on the last one.
**Current behaviour:** (Tell us what happens.)
Too many lashtails.
**Expected behaviour:** (Tell us what should happen instead.)
Not so many lashtails.
**Steps to reproduce the problem:**
1. .go creature 40493
2. Observe extra lashtails
3. ?????
4. Profit
**Branch(es):** 3.3.5 / master (Tell us which branch(es) this issue affects.)
3.3.5
**TC rev. hash/commit:**
e991a15dd7cbf5ba374f8942b7ea88f3a4d7b365
**TDB version:**
TDB_full_world_335.62_2016_10_17 + as per hash above.
**Operating system:**
Linux
[//]: # (This template is for problem reports. For other types of report, edit it accordingly.)
[//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
[//]: # (For fixes containing C++, create a Pull Request.)
|
non_process
|
extra lashtails in barrens don t delete this template else your issue will be closed description while looking at some videos of the game on the subject of pooling this one in particular at around this point there are some extra lashtails in the barrens that don t belong there for sure you would really not see them spawned like this in reality as seen here the one that belongs can be seen to be roaming the other are static the one that belongs is spawnid those that don t are there is also another static one added i can t say whether that one belongs or not for sure it shouldn t be static soliciting opinion before i remove them enable movement on the last one current behaviour tell us what happens too many lashtails expected behaviour tell us what should happen instead not so many lashtails steps to reproduce the problem go creature observe extra lashtails profit branch es master tell us which branch es this issue affects tc rev hash commit tdb version tdb full world as per hash above operating system linux this template is for problem reports for other types of report edit it accordingly if this is a crash report include the crashlog with for fixes containing c create a pull request
| 0
|
16,515
| 21,527,132,701
|
IssuesEvent
|
2022-04-28 19:40:10
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Something went wrong error for card, no 'dataset_query' returned by API/card
|
Type:Bug Priority:P2 Querying/Processor .Backend .Regression
|
**Describe the bug**
When loading a user-created question, the generic "Something went wrong" (+flaming server) error page is shown.
It is not clear why this happened or what the exact repro steps are. Some notes:
- User interface default language is set to German
- Instance was upgraded from v0.35.4 to v0.39.2, question created under v0.35.4
- The GET /api/card endpoint does not include a "dataset_query" object, unlike a normal response
- The question has a join to a saved question as per #12928 / #16353
- The question has filters on:
- an expression in the saved question
- a date field, where the filter includes a time
- The expression is an integer field divided by the 'count' metric calculated in the question
If it's not able to be reproduced, I would like a pointer on how to fix the question in the backend.
Though I'm not sure if the "dataset_query" object looks correct or what should be changed if it's wrong.
**Logs**
The browser (Firefox 88) console says:
```
Error: TypeError: c.dataset_query is null
```
The Metabase database entry for the affected card (note leading space).
```
> select dataset_query from report_card where id = 123;
{"type":"query","query":{"source-table":"card__122","joins":[{"fields":"all","source-table":555,"condition":["=",["field-literal","person_id","type/Text"],["joined-field","persons",["field-id",16666]]],"alias":"persons"}],"filter":["and",["<",["datetime-field",["field-literal","date_seen","type/Date"],"minute"],"2021-05-01T12:30:00"],["=",["field-literal","keep_entry","type/Float"],1]]},"database":12}
```
When viewing the question, the Metabase logs say:
```
qtp432197459-6257 | 2021-06-08 07:51:13,041 | ERROR | metabase.models.interface | Unable to normalize:
{:type "query",
:query
{:source-table "card__122",
:joins
[{:fields "all",
:source-table 555,
:condition ["=" ["field-literal" "person_id" "type/Text"] ["joined-field" "persons" ["field-id" 16666]]],
:alias "persons"}],
:filter
["and"
["<" ["datetime-field" ["field-literal" "date_seen" "type/Date"] "minute"] "2021-05-01T12:30:00"]
["=" ["field-literal" "keep_entry" "type/Float"] 1]]},
:database 12}
clojure.lang.ExceptionInfo: Error normalizing query
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:795) ~[metabase.jar:?]
at metabase.models.interface$maybe_normalize.invokeStatic(interface.clj:68) ~[metabase.jar:?]
at metabase.models.interface$maybe_normalize.invoke(interface.clj:67) ~[metabase.jar:?]
at metabase.models.interface$catch_normalization_exceptions$fn__24676.invoke(interface.clj:78) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at toucan.models$apply_type_fns$iter__24558__24562$fn__24563.invoke(models.clj:304) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?]
at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?]
at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?]
at toucan.models$apply_type_fns.invokeStatic(models.clj:302) ~[metabase.jar:?]
at toucan.models$apply_type_fns.invoke(models.clj:299) ~[metabase.jar:?]
at toucan.models$do_post_select.invokeStatic(models.clj:349) ~[metabase.jar:?]
at toucan.models$do_post_select.invoke(models.clj:344) ~[metabase.jar:?]
at toucan.db$do_post_select$iter__28532__28536$fn__28537.invoke(db.clj:373) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.lang.LazilyPersistentVector.create(LazilyPersistentVector.java:44) ~[metabase.jar:?]
at clojure.core$vec.invokeStatic(core.clj:377) ~[metabase.jar:?]
at clojure.core$vec.invoke(core.clj:367) ~[metabase.jar:?]
at toucan.db$do_post_select.invokeStatic(db.clj:372) ~[metabase.jar:?]
at toucan.db$do_post_select.invoke(db.clj:363) ~[metabase.jar:?]
at toucan.db$simple_select.invokeStatic(db.clj:394) ~[metabase.jar:?]
at toucan.db$simple_select.invoke(db.clj:383) ~[metabase.jar:?]
at toucan.db$simple_select_one.invokeStatic(db.clj:420) ~[metabase.jar:?]
at toucan.db$simple_select_one.invoke(db.clj:409) ~[metabase.jar:?]
at toucan.db$select_one.invokeStatic(db.clj:627) ~[metabase.jar:?]
at toucan.db$select_one.doInvoke(db.clj:620) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:139) ~[metabase.jar:?]
at clojure.lang.Var.applyTo(Var.java:705) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:671) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at toucan.models$invoke_model.invokeStatic(models.clj:380) ~[metabase.jar:?]
at toucan.models$invoke_model.doInvoke(models.clj:368) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:445) ~[metabase.jar:?]
at toucan.models$invoke_model.invokeStatic(models.clj:378) ~[metabase.jar:?]
at toucan.models$invoke_model.invoke(models.clj:368) ~[metabase.jar:?]
at clojure.lang.AFn.applyToHelper(AFn.java:156) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:132) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:667) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at toucan.models$invoke_model_or_instance.invokeStatic(models.clj:395) ~[metabase.jar:?]
at toucan.models$invoke_model_or_instance.doInvoke(models.clj:391) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:423) ~[metabase.jar:?]
at metabase.models.card.CardInstance.invoke(card.clj:25) ~[metabase.jar:?]
at metabase.api.card$fn__65802.invokeStatic(card.clj:162) ~[metabase.jar:?]
at metabase.api.card$fn__65802.invoke(card.clj:159) ~[metabase.jar:?]
at compojure.core$wrap_response$fn__12050.invoke(core.clj:160) ~[metabase.jar:?]
at compojure.core$wrap_route_middleware$fn__12034.invoke(core.clj:132) ~[metabase.jar:?]
at compojure.core$wrap_route_info$fn__12039.invoke(core.clj:139) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:151) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at metabase.server.middleware.auth$enforce_authentication$fn__56336.invoke(auth.clj:14) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at clojure.lang.AFn.applyToHelper(AFn.java:160) ~[metabase.jar:?]
at clojure.lang.AFn.applyTo(AFn.java:144) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:665) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at metabase.server.routes$fn__83020$fn__83021.doInvoke(routes.clj:57) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:436) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__81060.invoke(exceptions.clj:98) ~[metabase.jar:?]
at metabase.server.middleware.exceptions$catch_api_exceptions$fn__81057.invoke(exceptions.clj:86) ~[metabase.jar:?]
at metabase.server.middleware.log$log_api_call$fn__82936$fn__82937.invoke(log.clj:195) ~[metabase.jar:?]
at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216) ~[metabase.jar:?]
at toucan.db$_do_with_call_counting.invoke(db.clj:209) ~[metabase.jar:?]
at metabase.server.middleware.log$log_api_call$fn__82936.invoke(log.clj:189) ~[metabase.jar:?]
at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__82572.invoke(browser_cookie.clj:30) ~[metabase.jar:?]
at metabase.server.middleware.security$add_security_headers$fn__81022.invoke(security.clj:142) ~[metabase.jar:?]
at metabase.server.middleware.json$wrap_json_body$fn__82717.invoke(json.clj:62) ~[metabase.jar:?]
at metabase.server.middleware.json$wrap_streamed_json_response$fn__82735.invoke(json.clj:98) ~[metabase.jar:?]
at ring.middleware.keyword_params$wrap_keyword_params$fn__83277.invoke(keyword_params.clj:55) ~[metabase.jar:?]
at ring.middleware.params$wrap_params$fn__83293.invoke(params.clj:69) ~[metabase.jar:?]
at metabase.server.middleware.misc$maybe_set_site_url$fn__35809.invoke(misc.clj:58) ~[metabase.jar:?]
at metabase.server.middleware.session$bind_current_user$fn__43391$fn__43392.invoke(session.clj:248) ~[metabase.jar:?]
at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:229) ~[metabase.jar:?]
at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:221) ~[metabase.jar:?]
at metabase.server.middleware.session$bind_current_user$fn__43391.invoke(session.clj:247) ~[metabase.jar:?]
at metabase.server.middleware.session$wrap_current_user_info$fn__43378.invoke(session.clj:207) ~[metabase.jar:?]
at metabase.server.middleware.session$wrap_session_id$fn__43364.invoke(session.clj:153) ~[metabase.jar:?]
at metabase.server.middleware.auth$wrap_api_key$fn__56344.invoke(auth.clj:27) ~[metabase.jar:?]
at ring.middleware.cookies$wrap_cookies$fn__83197.invoke(cookies.clj:216) ~[metabase.jar:?]
at metabase.server.middleware.misc$add_content_type$fn__35792.invoke(misc.clj:27) ~[metabase.jar:?]
at metabase.server.middleware.misc$disable_streaming_buffering$fn__35817.invoke(misc.clj:75) ~[metabase.jar:?]
at ring.middleware.gzip$wrap_gzip$fn__83239.invoke(gzip.clj:86) ~[metabase.jar:?]
at metabase.server.middleware.misc$bind_request$fn__35820.invoke(misc.clj:92) ~[metabase.jar:?]
at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__82953.invoke(ssl.clj:48) ~[metabase.jar:?]
at metabase.server$async_proxy_handler$fn__82516.invoke(server.clj:71) ~[metabase.jar:?]
at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) ~[metabase.jar:?]
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) ~[metabase.jar:?]
at org.eclipse.jetty.server.Server.handle(Server.java:516) ~[metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:383) ~[metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:556) [metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:375) [metabase.jar:?]
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:273) [metabase.jar:?]
at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311) [metabase.jar:?]
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) [metabase.jar:?]
at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:336) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:313) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:171) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:129) [metabase.jar:?]
at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:375) [metabase.jar:?]
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:773) [metabase.jar:?]
at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:905) [metabase.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?]
Caused by: clojure.lang.ExceptionInfo: Error canonicalizing query
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:677) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
Caused by: clojure.lang.ExceptionInfo: Invalid MBQL clause
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:555) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?]
at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?]
at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?]
at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?]
at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?]
at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?]
at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?]
at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?]
at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?]
at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?]
at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?]
at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
Caused by: clojure.lang.ExceptionInfo: Output of update-field-options does not match schema:
[nil nil (named (not ("Invalid :temporal-unit for the specified :base-type." a-clojure.lang.PersistentArrayMap)) "options")]
at metabase.mbql.util$fn__20573$update_field_options__20578.doInvoke(util.cljc:581) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:142) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:669) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at metabase.mbql.util$assoc_field_options.invokeStatic(util.cljc:590) ~[metabase.jar:?]
at metabase.mbql.util$assoc_field_options.doInvoke(util.cljc:587) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:439) ~[metabase.jar:?]
at metabase.mbql.util$with_temporal_unit.invokeStatic(util.cljc:595) ~[metabase.jar:?]
at metabase.mbql.util$with_temporal_unit.invoke(util.cljc:592) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20861.invokeStatic(normalize.cljc:404) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20861.invoke(normalize.cljc:398) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_implicit_field_id.invokeStatic(normalize.cljc:363) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_implicit_field_id.invoke(normalize.cljc:360) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_simple_filter_clause.invokeStatic(normalize.cljc:455) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_simple_filter_clause.invoke(normalize.cljc:452) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20918$fn__20919.invoke(normalize.cljc:464) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at clojure.core$map$fn__5866.invoke(core.clj:2753) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?]
at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?]
at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_compound_filter_clause.invokeStatic(normalize.cljc:425) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_compound_filter_clause.invoke(normalize.cljc:423) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20888$fn__20889.invoke(normalize.cljc:433) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:552) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?]
at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?]
at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?]
at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?]
at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?]
at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?]
at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?]
at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?]
at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?]
at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?]
at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?]
at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
```
**To Reproduce**
Not sure - have not found exactly how to reproduce it with the Sample Database.
A similar question complains about having a time filter on a date field, but no "something went wrong"
```
Unsupported temporal bucketing: You can't bucket a :type/Date Field by :minute.
```
Maybe it is about the combination of metric + expression + saved question + filters?
**Expected behavior**
The question should load without error, or at least an error that the user can rectify.
The "something went wrong" page does not show any of the normal Question UI controls.
**Information about your Metabase Installation:**
You can get this information by going to Admin -> Troubleshooting.
- Your browser and the version: (e.x. Chrome 52.1, Firefox 48.0, Safari 11.1, β¦) Firefox 88
- Your operating system: (e.x. OS X 10.10, Windows 10.1809, Ubuntu 16.04, β¦) CentOS 7
- Your databases: (e.x. MySQL, Postgres, MongoDB, β¦) PostgreSQL 10
- Metabase version: (e.x. 0.19.3) Metabase v0.39.2
- Metabase hosting environment: (e.x. Mac app, Elastic Beanstalk, Docker, Heroku, Jar-file on Ubuntu, β¦) CentOS 7
- Metabase internal database: (e.x. H2 (default), Postgres or MySQL) PostgreSQL 10
**Severity**
Blocking with respect to not being able to view or fix this particular question
|
1.0
|
Something went wrong error for card, no 'dataset_query' returned by API/card - **Describe the bug**
When loading a user-created question, the generic "Something went wrong" (+flaming server) error page is shown.
It is not clear why this happened or what the exact repro steps are. Some notes:
- User interface default language is set to German
- Instance was upgraded from v0.35.4 to v0.39.2, question created under v0.35.4
- The GET /api/card endpoint does not include a "dataset_query" object, unlike a normal response
- The question has a join to a saved question as per #12928 / #16353
- The question has filters on:
- an expression in the saved question
- a date field, where the filter includes a time
- The expression is an integer field divided by the 'count' metric calculated in the question
If it's not able to be reproduced, I would like a pointer on how to fix the question in the backend.
Though I'm not sure if the "dataset_query" object looks correct or what should be changed if it's wrong.
**Logs**
The browser (Firefox 88) console says:
```
Error: TypeError: c.dataset_query is null
```
The Metabase database entry for the affected card (note leading space).
```
> select dataset_query from report_card where id = 123;
{"type":"query","query":{"source-table":"card__122","joins":[{"fields":"all","source-table":555,"condition":["=",["field-literal","person_id","type/Text"],["joined-field","persons",["field-id",16666]]],"alias":"persons"}],"filter":["and",["<",["datetime-field",["field-literal","date_seen","type/Date"],"minute"],"2021-05-01T12:30:00"],["=",["field-literal","keep_entry","type/Float"],1]]},"database":12}
```
When viewing the question, the Metabase logs say:
```
qtp432197459-6257 | 2021-06-08 07:51:13,041 | ERROR | metabase.models.interface | Unable to normalize:
{:type "query",
:query
{:source-table "card__122",
:joins
[{:fields "all",
:source-table 555,
:condition ["=" ["field-literal" "person_id" "type/Text"] ["joined-field" "persons" ["field-id" 16666]]],
:alias "persons"}],
:filter
["and"
["<" ["datetime-field" ["field-literal" "date_seen" "type/Date"] "minute"] "2021-05-01T12:30:00"]
["=" ["field-literal" "keep_entry" "type/Float"] 1]]},
:database 12}
clojure.lang.ExceptionInfo: Error normalizing query
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:795) ~[metabase.jar:?]
at metabase.models.interface$maybe_normalize.invokeStatic(interface.clj:68) ~[metabase.jar:?]
at metabase.models.interface$maybe_normalize.invoke(interface.clj:67) ~[metabase.jar:?]
at metabase.models.interface$catch_normalization_exceptions$fn__24676.invoke(interface.clj:78) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at toucan.models$apply_type_fns$iter__24558__24562$fn__24563.invoke(models.clj:304) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?]
at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?]
at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?]
at toucan.models$apply_type_fns.invokeStatic(models.clj:302) ~[metabase.jar:?]
at toucan.models$apply_type_fns.invoke(models.clj:299) ~[metabase.jar:?]
at toucan.models$do_post_select.invokeStatic(models.clj:349) ~[metabase.jar:?]
at toucan.models$do_post_select.invoke(models.clj:344) ~[metabase.jar:?]
at toucan.db$do_post_select$iter__28532__28536$fn__28537.invoke(db.clj:373) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.lang.LazilyPersistentVector.create(LazilyPersistentVector.java:44) ~[metabase.jar:?]
at clojure.core$vec.invokeStatic(core.clj:377) ~[metabase.jar:?]
at clojure.core$vec.invoke(core.clj:367) ~[metabase.jar:?]
at toucan.db$do_post_select.invokeStatic(db.clj:372) ~[metabase.jar:?]
at toucan.db$do_post_select.invoke(db.clj:363) ~[metabase.jar:?]
at toucan.db$simple_select.invokeStatic(db.clj:394) ~[metabase.jar:?]
at toucan.db$simple_select.invoke(db.clj:383) ~[metabase.jar:?]
at toucan.db$simple_select_one.invokeStatic(db.clj:420) ~[metabase.jar:?]
at toucan.db$simple_select_one.invoke(db.clj:409) ~[metabase.jar:?]
at toucan.db$select_one.invokeStatic(db.clj:627) ~[metabase.jar:?]
at toucan.db$select_one.doInvoke(db.clj:620) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:139) ~[metabase.jar:?]
at clojure.lang.Var.applyTo(Var.java:705) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:671) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at toucan.models$invoke_model.invokeStatic(models.clj:380) ~[metabase.jar:?]
at toucan.models$invoke_model.doInvoke(models.clj:368) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:445) ~[metabase.jar:?]
at toucan.models$invoke_model.invokeStatic(models.clj:378) ~[metabase.jar:?]
at toucan.models$invoke_model.invoke(models.clj:368) ~[metabase.jar:?]
at clojure.lang.AFn.applyToHelper(AFn.java:156) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:132) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:667) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at toucan.models$invoke_model_or_instance.invokeStatic(models.clj:395) ~[metabase.jar:?]
at toucan.models$invoke_model_or_instance.doInvoke(models.clj:391) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:423) ~[metabase.jar:?]
at metabase.models.card.CardInstance.invoke(card.clj:25) ~[metabase.jar:?]
at metabase.api.card$fn__65802.invokeStatic(card.clj:162) ~[metabase.jar:?]
at metabase.api.card$fn__65802.invoke(card.clj:159) ~[metabase.jar:?]
at compojure.core$wrap_response$fn__12050.invoke(core.clj:160) ~[metabase.jar:?]
at compojure.core$wrap_route_middleware$fn__12034.invoke(core.clj:132) ~[metabase.jar:?]
at compojure.core$wrap_route_info$fn__12039.invoke(core.clj:139) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:151) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:153) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at metabase.server.middleware.auth$enforce_authentication$fn__56336.invoke(auth.clj:14) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at clojure.lang.AFn.applyToHelper(AFn.java:160) ~[metabase.jar:?]
at clojure.lang.AFn.applyTo(AFn.java:144) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:665) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at metabase.server.routes$fn__83020$fn__83021.doInvoke(routes.clj:57) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:436) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$make_context$handler__12090.invoke(core.clj:287) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:296) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$wrap_route_matches$fn__12043.invoke(core.clj:152) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:199) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063$respond_SINGLEQUOTE___12064.invoke(core.clj:197) ~[metabase.jar:?]
at compojure.core$make_context$fn__12092.invoke(core.clj:297) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at compojure.core$routes$fn__12062$f__12063.invoke(core.clj:198) ~[metabase.jar:?]
at compojure.core$routes$fn__12062.invoke(core.clj:200) ~[metabase.jar:?]
at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__81060.invoke(exceptions.clj:98) ~[metabase.jar:?]
at metabase.server.middleware.exceptions$catch_api_exceptions$fn__81057.invoke(exceptions.clj:86) ~[metabase.jar:?]
at metabase.server.middleware.log$log_api_call$fn__82936$fn__82937.invoke(log.clj:195) ~[metabase.jar:?]
at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216) ~[metabase.jar:?]
at toucan.db$_do_with_call_counting.invoke(db.clj:209) ~[metabase.jar:?]
at metabase.server.middleware.log$log_api_call$fn__82936.invoke(log.clj:189) ~[metabase.jar:?]
at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__82572.invoke(browser_cookie.clj:30) ~[metabase.jar:?]
at metabase.server.middleware.security$add_security_headers$fn__81022.invoke(security.clj:142) ~[metabase.jar:?]
at metabase.server.middleware.json$wrap_json_body$fn__82717.invoke(json.clj:62) ~[metabase.jar:?]
at metabase.server.middleware.json$wrap_streamed_json_response$fn__82735.invoke(json.clj:98) ~[metabase.jar:?]
at ring.middleware.keyword_params$wrap_keyword_params$fn__83277.invoke(keyword_params.clj:55) ~[metabase.jar:?]
at ring.middleware.params$wrap_params$fn__83293.invoke(params.clj:69) ~[metabase.jar:?]
at metabase.server.middleware.misc$maybe_set_site_url$fn__35809.invoke(misc.clj:58) ~[metabase.jar:?]
at metabase.server.middleware.session$bind_current_user$fn__43391$fn__43392.invoke(session.clj:248) ~[metabase.jar:?]
at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:229) ~[metabase.jar:?]
at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:221) ~[metabase.jar:?]
at metabase.server.middleware.session$bind_current_user$fn__43391.invoke(session.clj:247) ~[metabase.jar:?]
at metabase.server.middleware.session$wrap_current_user_info$fn__43378.invoke(session.clj:207) ~[metabase.jar:?]
at metabase.server.middleware.session$wrap_session_id$fn__43364.invoke(session.clj:153) ~[metabase.jar:?]
at metabase.server.middleware.auth$wrap_api_key$fn__56344.invoke(auth.clj:27) ~[metabase.jar:?]
at ring.middleware.cookies$wrap_cookies$fn__83197.invoke(cookies.clj:216) ~[metabase.jar:?]
at metabase.server.middleware.misc$add_content_type$fn__35792.invoke(misc.clj:27) ~[metabase.jar:?]
at metabase.server.middleware.misc$disable_streaming_buffering$fn__35817.invoke(misc.clj:75) ~[metabase.jar:?]
at ring.middleware.gzip$wrap_gzip$fn__83239.invoke(gzip.clj:86) ~[metabase.jar:?]
at metabase.server.middleware.misc$bind_request$fn__35820.invoke(misc.clj:92) ~[metabase.jar:?]
at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__82953.invoke(ssl.clj:48) ~[metabase.jar:?]
at metabase.server$async_proxy_handler$fn__82516.invoke(server.clj:71) ~[metabase.jar:?]
at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) ~[metabase.jar:?]
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) ~[metabase.jar:?]
at org.eclipse.jetty.server.Server.handle(Server.java:516) ~[metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:383) ~[metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:556) [metabase.jar:?]
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:375) [metabase.jar:?]
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:273) [metabase.jar:?]
at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311) [metabase.jar:?]
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) [metabase.jar:?]
at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:336) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:313) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:171) [metabase.jar:?]
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:129) [metabase.jar:?]
at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:375) [metabase.jar:?]
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:773) [metabase.jar:?]
at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:905) [metabase.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?]
Caused by: clojure.lang.ExceptionInfo: Error canonicalizing query
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:677) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
Caused by: clojure.lang.ExceptionInfo: Invalid MBQL clause
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:555) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?]
at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?]
at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?]
at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?]
at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?]
at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?]
at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?]
at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?]
at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?]
at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?]
at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?]
at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
Caused by: clojure.lang.ExceptionInfo: Output of update-field-options does not match schema:
[nil nil (named (not ("Invalid :temporal-unit for the specified :base-type." a-clojure.lang.PersistentArrayMap)) "options")]
at metabase.mbql.util$fn__20573$update_field_options__20578.doInvoke(util.cljc:581) ~[metabase.jar:?]
at clojure.lang.RestFn.applyTo(RestFn.java:142) ~[metabase.jar:?]
at clojure.core$apply.invokeStatic(core.clj:669) ~[metabase.jar:?]
at clojure.core$apply.invoke(core.clj:660) ~[metabase.jar:?]
at metabase.mbql.util$assoc_field_options.invokeStatic(util.cljc:590) ~[metabase.jar:?]
at metabase.mbql.util$assoc_field_options.doInvoke(util.cljc:587) ~[metabase.jar:?]
at clojure.lang.RestFn.invoke(RestFn.java:439) ~[metabase.jar:?]
at metabase.mbql.util$with_temporal_unit.invokeStatic(util.cljc:595) ~[metabase.jar:?]
at metabase.mbql.util$with_temporal_unit.invoke(util.cljc:592) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20861.invokeStatic(normalize.cljc:404) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20861.invoke(normalize.cljc:398) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_implicit_field_id.invokeStatic(normalize.cljc:363) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_implicit_field_id.invoke(normalize.cljc:360) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_simple_filter_clause.invokeStatic(normalize.cljc:455) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_simple_filter_clause.invoke(normalize.cljc:452) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20918$fn__20919.invoke(normalize.cljc:464) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at clojure.core$map$fn__5866.invoke(core.clj:2753) ~[metabase.jar:?]
at clojure.lang.LazySeq.sval(LazySeq.java:42) ~[metabase.jar:?]
at clojure.lang.LazySeq.seq(LazySeq.java:51) ~[metabase.jar:?]
at clojure.lang.RT.seq(RT.java:535) ~[metabase.jar:?]
at clojure.core$seq__5402.invokeStatic(core.clj:137) ~[metabase.jar:?]
at clojure.core.protocols$seq_reduce.invokeStatic(protocols.clj:24) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8146.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$into.invokeStatic(core.clj:6895) ~[metabase.jar:?]
at clojure.core$into.invoke(core.clj:6887) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_compound_filter_clause.invokeStatic(normalize.cljc:425) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_compound_filter_clause.invoke(normalize.cljc:423) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__20888$fn__20889.invoke(normalize.cljc:433) ~[metabase.jar:?]
at clojure.lang.MultiFn.invoke(MultiFn.java:229) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:552) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at medley.core$map_vals$fn__10074$fn__10075.invoke(core.cljc:110) ~[metabase.jar:?]
at clojure.core$fn__8429$fn__8431.invoke(core.clj:6840) ~[metabase.jar:?]
at clojure.core.protocols$iter_reduce.invokeStatic(protocols.clj:49) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invokeStatic(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8140.invoke(protocols.clj:75) ~[metabase.jar:?]
at clojure.core.protocols$fn__8088$G__8083__8101.invoke(protocols.clj:13) ~[metabase.jar:?]
at clojure.core$reduce.invokeStatic(core.clj:6828) ~[metabase.jar:?]
at clojure.core$fn__8429.invokeStatic(core.clj:6830) ~[metabase.jar:?]
at clojure.core$fn__8429.invoke(core.clj:6830) ~[metabase.jar:?]
at clojure.core.protocols$fn__8167$G__8162__8176.invoke(protocols.clj:175) ~[metabase.jar:?]
at clojure.core$reduce_kv.invokeStatic(core.clj:6856) ~[metabase.jar:?]
at clojure.core$reduce_kv.invoke(core.clj:6847) ~[metabase.jar:?]
at medley.core$reduce_map.invokeStatic(core.cljc:84) ~[metabase.jar:?]
at medley.core$reduce_map.invoke(core.cljc:81) ~[metabase.jar:?]
at medley.core$map_vals.invokeStatic(core.cljc:110) ~[metabase.jar:?]
at medley.core$map_vals.invoke(core.cljc:105) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses$fn__21033.invoke(normalize.cljc:545) ~[metabase.jar:?]
at clojure.walk$prewalk.invokeStatic(walk.clj:65) ~[metabase.jar:?]
at clojure.walk$prewalk.invoke(walk.clj:61) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invokeStatic(normalize.cljc:541) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize_mbql_clauses.invoke(normalize.cljc:538) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$update.invokeStatic(core.clj:6196) ~[metabase.jar:?]
at clojure.core$update.invoke(core.clj:6188) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invokeStatic(normalize.cljc:672) ~[metabase.jar:?]
at metabase.mbql.normalize$canonicalize.invoke(normalize.cljc:665) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at clojure.core$comp$fn__5807.invoke(core.clj:2569) ~[metabase.jar:?]
at metabase.mbql.normalize$fn__21188$fn__21189.invoke(normalize.cljc:793) ~[metabase.jar:?]
... 201 more
```
**To Reproduce**
Not sure - have not found exactly how to reproduce it with the Sample Database.
A similar question complains about having a time filter on a date field, but no "something went wrong"
```
Unsupported temporal bucketing: You can't bucket a :type/Date Field by :minute.
```
Maybe it is about the combination of metric + expression + saved question + filters?
**Expected behavior**
The question should load without error, or at least an error that the user can rectify.
The "something went wrong" page does not show any of the normal Question UI controls.
**Information about your Metabase Installation:**
You can get this information by going to Admin -> Troubleshooting.
- Your browser and the version: (e.x. Chrome 52.1, Firefox 48.0, Safari 11.1, β¦) Firefox 88
- Your operating system: (e.x. OS X 10.10, Windows 10.1809, Ubuntu 16.04, β¦) CentOS 7
- Your databases: (e.x. MySQL, Postgres, MongoDB, β¦) PostgreSQL 10
- Metabase version: (e.x. 0.19.3) Metabase v0.39.2
- Metabase hosting environment: (e.x. Mac app, Elastic Beanstalk, Docker, Heroku, Jar-file on Ubuntu, β¦) CentOS 7
- Metabase internal database: (e.x. H2 (default), Postgres or MySQL) PostgreSQL 10
**Severity**
Blocking with respect to not being able to view or fix this particular question
|
process
|
something went wrong error for card no dataset query returned by api card describe the bug when loading a user created question the generic something went wrong flaming server error page is shown it is not clear why this happened or what the exact repro steps are some notes user interface default language is set to german instance was upgraded from to question created under the get api card endpoint does not include a dataset query object unlike a normal response the question has a join to a saved question as per the question has filters on an expression in the saved question a date field where the filter includes a time the expression is an integer field divided by the count metric calculated in the question if it s not able to be reproduced i would like a pointer on how to fix the question in the backend though i m not sure if the dataset query object looks correct or what should be changed if it s wrong logs the browser firefox console says error typeerror c dataset query is null the metabase database entry for the affected card note leading space select dataset query from report card where id type query query source table card joins alias persons filter minute database when viewing the question the metabase logs say error metabase models interface unable to normalize type query query source table card joins fields all source table condition alias persons filter and minute database clojure lang exceptioninfo error normalizing query at metabase mbql normalize fn fn invoke normalize cljc at metabase models interface maybe normalize invokestatic interface clj at metabase models interface maybe normalize invoke interface clj at metabase models interface catch normalization exceptions fn invoke interface clj at clojure core comp fn invoke core clj at toucan models apply type fns iter fn invoke models clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure core seq invokestatic core clj at clojure core protocols seq reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core into invokestatic core clj at clojure core into invoke core clj at toucan models apply type fns invokestatic models clj at toucan models apply type fns invoke models clj at toucan models do post select invokestatic models clj at toucan models do post select invoke models clj at toucan db do post select iter fn invoke db clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure lang lazilypersistentvector create lazilypersistentvector java at clojure core vec invokestatic core clj at clojure core vec invoke core clj at toucan db do post select invokestatic db clj at toucan db do post select invoke db clj at toucan db simple select invokestatic db clj at toucan db simple select invoke db clj at toucan db simple select one invokestatic db clj at toucan db simple select one invoke db clj at toucan db select one invokestatic db clj at toucan db select one doinvoke db clj at clojure lang restfn applyto restfn java at clojure lang var applyto var java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at toucan models invoke model invokestatic models clj at toucan models invoke model doinvoke models clj at clojure lang restfn invoke restfn java at toucan models invoke model invokestatic models clj at toucan models invoke model invoke models clj at clojure lang afn applytohelper afn java at clojure lang restfn applyto restfn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at toucan models invoke model or instance invokestatic models clj at toucan models invoke model or instance doinvoke models clj at clojure lang restfn invoke restfn java at metabase models card cardinstance invoke card clj at metabase api card fn invokestatic card clj at metabase api card fn invoke card clj at compojure core wrap response fn invoke core clj at compojure core wrap route middleware fn invoke core clj at compojure core wrap route info fn invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware auth enforce authentication fn invoke auth clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at clojure lang afn applytohelper afn java at clojure lang afn applyto afn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at metabase server routes fn fn doinvoke routes clj at clojure lang restfn invoke restfn java at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware exceptions catch uncaught exceptions fn invoke exceptions clj at metabase server middleware exceptions catch api exceptions fn invoke exceptions clj at metabase server middleware log log api call fn fn invoke log clj at toucan db do with call counting invokestatic db clj at toucan db do with call counting invoke db clj at metabase server middleware log log api call fn invoke log clj at metabase server middleware browser cookie ensure browser id cookie fn invoke browser cookie clj at metabase server middleware security add security headers fn invoke security clj at metabase server middleware json wrap json body fn invoke json clj at metabase server middleware json wrap streamed json response fn invoke json clj at ring middleware keyword params wrap keyword params fn invoke keyword params clj at ring middleware params wrap params fn invoke params clj at metabase server middleware misc maybe set site url fn invoke misc clj at metabase server middleware session bind current user fn fn invoke session clj at metabase server middleware session do with current user invokestatic session clj at metabase server middleware session do with current user invoke session clj at metabase server middleware session bind current user fn invoke session clj at metabase server middleware session wrap current user info fn invoke session clj at metabase server middleware session wrap session id fn invoke session clj at metabase server middleware auth wrap api key fn invoke auth clj at ring middleware cookies wrap cookies fn invoke cookies clj at metabase server middleware misc add content type fn invoke misc clj at metabase server middleware misc disable streaming buffering fn invoke misc clj at ring middleware gzip wrap gzip fn invoke gzip clj at metabase server middleware misc bind request fn invoke misc clj at metabase server middleware ssl redirect to https middleware fn invoke ssl clj at metabase server async proxy handler fn invoke server clj at metabase server proxy org eclipse jetty server handler abstracthandler handle unknown source at org eclipse jetty server handler handlerwrapper handle handlerwrapper java at org eclipse jetty server server handle server java at org eclipse jetty server httpchannel lambda handle httpchannel java at org eclipse jetty server httpchannel dispatch httpchannel java at org eclipse jetty server httpchannel handle httpchannel java at org eclipse jetty server httpconnection onfillable httpconnection java at org eclipse jetty io abstractconnection readcallback succeeded abstractconnection java at org eclipse jetty io fillinterest fillable fillinterest java at org eclipse jetty io channelendpoint run channelendpoint java at org eclipse jetty util thread strategy eatwhatyoukill runtask eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill doproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill tryproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill run eatwhatyoukill java at org eclipse jetty util thread reservedthreadexecutor reservedthread run reservedthreadexecutor java at org eclipse jetty util thread queuedthreadpool runjob queuedthreadpool java at org eclipse jetty util thread queuedthreadpool runner run queuedthreadpool java at java lang thread run thread java caused by clojure lang exceptioninfo error canonicalizing query at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more caused by clojure lang exceptioninfo invalid mbql clause at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at medley core map vals fn fn invoke core cljc at clojure core fn fn invoke core clj at clojure core protocols iter reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core fn invokestatic core clj at clojure core fn invoke core clj at clojure core protocols fn g invoke protocols clj at clojure core reduce kv invokestatic core clj at clojure core reduce kv invoke core clj at medley core reduce map invokestatic core cljc at medley core reduce map invoke core cljc at medley core map vals invokestatic core cljc at medley core map vals invoke core cljc at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at clojure core comp fn invoke core clj at clojure core update invokestatic core clj at clojure core update invoke core clj at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more caused by clojure lang exceptioninfo output of update field options does not match schema at metabase mbql util fn update field options doinvoke util cljc at clojure lang restfn applyto restfn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at metabase mbql util assoc field options invokestatic util cljc at metabase mbql util assoc field options doinvoke util cljc at clojure lang restfn invoke restfn java at metabase mbql util with temporal unit invokestatic util cljc at metabase mbql util with temporal unit invoke util cljc at metabase mbql normalize fn invokestatic normalize cljc at metabase mbql normalize fn invoke normalize cljc at clojure lang multifn invoke multifn java at metabase mbql normalize canonicalize implicit field id invokestatic normalize cljc at metabase mbql normalize canonicalize implicit field id invoke normalize cljc at metabase mbql normalize canonicalize simple filter clause invokestatic normalize cljc at metabase mbql normalize canonicalize simple filter clause invoke normalize cljc at metabase mbql normalize fn fn invoke normalize cljc at clojure lang multifn invoke multifn java at clojure core map fn invoke core clj at clojure lang lazyseq sval lazyseq java at clojure lang lazyseq seq lazyseq java at clojure lang rt seq rt java at clojure core seq invokestatic core clj at clojure core protocols seq reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core into invokestatic core clj at clojure core into invoke core clj at metabase mbql normalize canonicalize compound filter clause invokestatic normalize cljc at metabase mbql normalize canonicalize compound filter clause invoke normalize cljc at metabase mbql normalize fn fn invoke normalize cljc at clojure lang multifn invoke multifn java at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at medley core map vals fn fn invoke core cljc at clojure core fn fn invoke core clj at clojure core protocols iter reduce invokestatic protocols clj at clojure core protocols fn invokestatic protocols clj at clojure core protocols fn invoke protocols clj at clojure core protocols fn g invoke protocols clj at clojure core reduce invokestatic core clj at clojure core fn invokestatic core clj at clojure core fn invoke core clj at clojure core protocols fn g invoke protocols clj at clojure core reduce kv invokestatic core clj at clojure core reduce kv invoke core clj at medley core reduce map invokestatic core cljc at medley core reduce map invoke core cljc at medley core map vals invokestatic core cljc at medley core map vals invoke core cljc at metabase mbql normalize canonicalize mbql clauses fn invoke normalize cljc at clojure walk prewalk invokestatic walk clj at clojure walk prewalk invoke walk clj at metabase mbql normalize canonicalize mbql clauses invokestatic normalize cljc at metabase mbql normalize canonicalize mbql clauses invoke normalize cljc at clojure core comp fn invoke core clj at clojure core update invokestatic core clj at clojure core update invoke core clj at metabase mbql normalize canonicalize invokestatic normalize cljc at metabase mbql normalize canonicalize invoke normalize cljc at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase mbql normalize fn fn invoke normalize cljc more to reproduce not sure have not found exactly how to reproduce it with the sample database a similar question complains about having a time filter on a date field but no something went wrong unsupported temporal bucketing you can t bucket a type date field by minute maybe it is about the combination of metric expression saved question filters expected behavior the question should load without error or at least an error that the user can rectify the something went wrong page does not show any of the normal question ui controls information about your metabase installation you can get this information by going to admin troubleshooting your browser and the version e x chrome firefox safari β¦ firefox your operating system e x os x windows ubuntu β¦ centos your databases e x mysql postgres mongodb β¦ postgresql metabase version e x metabase metabase hosting environment e x mac app elastic beanstalk docker heroku jar file on ubuntu β¦ centos metabase internal database e x default postgres or mysql postgresql severity blocking with respect to not being able to view or fix this particular question
| 1
|
17,253
| 23,035,290,332
|
IssuesEvent
|
2022-07-22 17:59:54
|
pycaret/pycaret
|
https://api.github.com/repos/pycaret/pycaret
|
closed
|
[BUG]: Combination of PCA and remove_multicollinearity is causing error
|
bug preprocessing
|
### pycaret version checks
- [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues).
- [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret.
- [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master).
### Issue Description
Not possible to use remove_multicollinearity and pca in combination in setup.
### Reproducible Example
```python
from pycaret.datasets import get_data
from pycaret.classification import *
_data = get_data('satellite')
exp = ClassificationExperiment()
exp.setup(data = _data, remove_multicollinearity = True,
multicollinearity_threshold = 0.9, pca_components= 1.0, pca = True, target = 'Class')
```
### Expected Behavior
Since the nr of pca components is set to 1.0, you would expect 100% of the columns that are left after all the other feature selection methods to be kept to the same nr.
### Actual Results
```python-traceback
Instead, this is throwing the error:
497 raise ValueError(
498 "n_components='mle' is only supported if n_samples >= n_features"
499 )
500 elif not 0 <= n_components <= min(n_samples, n_features):
--> 501 raise ValueError(
502 "n_components=%r must be between 0 and "
503 "min(n_samples, n_features)=%r with "
504 "svd_solver='full'" % (n_components, min(n_samples, n_features))
505 )
506 elif n_components >= 1:
507 if not isinstance(n_components, numbers.Integral):
ValueError: n_components=36 must be between 0 and min(n_samples, n_features)=4 with svd_solver='full'
```
### Installed Versions
<details>
'3.0.0.rc2'
</details>
|
1.0
|
[BUG]: Combination of PCA and remove_multicollinearity is causing error - ### pycaret version checks
- [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues).
- [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret.
- [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master).
### Issue Description
Not possible to use remove_multicollinearity and pca in combination in setup.
### Reproducible Example
```python
from pycaret.datasets import get_data
from pycaret.classification import *
_data = get_data('satellite')
exp = ClassificationExperiment()
exp.setup(data = _data, remove_multicollinearity = True,
multicollinearity_threshold = 0.9, pca_components= 1.0, pca = True, target = 'Class')
```
### Expected Behavior
Since the nr of pca components is set to 1.0, you would expect 100% of the columns that are left after all the other feature selection methods to be kept to the same nr.
### Actual Results
```python-traceback
Instead, this is throwing the error:
497 raise ValueError(
498 "n_components='mle' is only supported if n_samples >= n_features"
499 )
500 elif not 0 <= n_components <= min(n_samples, n_features):
--> 501 raise ValueError(
502 "n_components=%r must be between 0 and "
503 "min(n_samples, n_features)=%r with "
504 "svd_solver='full'" % (n_components, min(n_samples, n_features))
505 )
506 elif n_components >= 1:
507 if not isinstance(n_components, numbers.Integral):
ValueError: n_components=36 must be between 0 and min(n_samples, n_features)=4 with svd_solver='full'
```
### Installed Versions
<details>
'3.0.0.rc2'
</details>
|
process
|
combination of pca and remove multicollinearity is causing error pycaret version checks i have checked that this issue has not already been reported i have confirmed this bug exists on the of pycaret i have confirmed this bug exists on the master branch of pycaret pip install u git issue description not possible to use remove multicollinearity and pca in combination in setup reproducible example python from pycaret datasets import get data from pycaret classification import data get data satellite exp classificationexperiment exp setup data data remove multicollinearity true multicollinearity threshold pca components pca true target class expected behavior since the nr of pca components is set to you would expect of the columns that are left after all the other feature selection methods to be kept to the same nr actual results python traceback instead this is throwing the error raise valueerror n components mle is only supported if n samples n features elif not n components min n samples n features raise valueerror n components r must be between and min n samples n features r with svd solver full n components min n samples n features elif n components if not isinstance n components numbers integral valueerror n components must be between and min n samples n features with svd solver full installed versions
| 1
|
17,525
| 23,336,236,622
|
IssuesEvent
|
2022-08-09 10:11:01
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
System.Diagnostics.Process work planned for .NET 7
|
Epic area-System.Diagnostics.Process Priority:3 Team:Libraries
|
**This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope.**
## Summary
We are not planning any notable investments into the System.Diagnostics.Process area in .NET 7. We will address high-impact issues and resolve test issues. We will consider small community contributions that improve cross-platform compatibility.
## Planned for .NET 7
- [x] #45017
- [x] #63937
- [x] #44453
## Backlog (in roughly priority order)
- [ ] #58492
- [ ] #53095
- [ ] #49107
- [ ] #45685
|
1.0
|
System.Diagnostics.Process work planned for .NET 7 - **This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope.**
## Summary
We are not planning any notable investments into the System.Diagnostics.Process area in .NET 7. We will address high-impact issues and resolve test issues. We will consider small community contributions that improve cross-platform compatibility.
## Planned for .NET 7
- [x] #45017
- [x] #63937
- [x] #44453
## Backlog (in roughly priority order)
- [ ] #58492
- [ ] #53095
- [ ] #49107
- [ ] #45685
|
process
|
system diagnostics process work planned for net this issue captures the planned work for net this list is expected to change throughout the release cycle according to ongoing planning and discussions with possible additions and subtractions to the scope summary we are not planning any notable investments into the system diagnostics process area in net we will address high impact issues and resolve test issues we will consider small community contributions that improve cross platform compatibility planned for net backlog in roughly priority order
| 1
|
17,056
| 22,475,071,639
|
IssuesEvent
|
2022-06-22 11:33:56
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Update Go client to use 1.17
|
kind/toil scope/clients-go team/process-automation scope/zbctl area/maintainability
|
**Description**
Our Go client is now built against 1.15, which is not supported anymore. We should start building against Go 1.17, which is the lowest support version.
|
1.0
|
Update Go client to use 1.17 - **Description**
Our Go client is now built against 1.15, which is not supported anymore. We should start building against Go 1.17, which is the lowest support version.
|
process
|
update go client to use description our go client is now built against which is not supported anymore we should start building against go which is the lowest support version
| 1
|
22,096
| 30,615,780,034
|
IssuesEvent
|
2023-07-24 02:53:37
|
solop-develop/frontend-core
|
https://api.github.com/repos/solop-develop/frontend-core
|
closed
|
[Bug Report] Reporte/Proceso: Parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores
|
bug (PRC) Processes (RPT) Reports (ULP) UI Lookups (UX) User Experience
|
<!--
Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed.
-->
## Bug report
#### Steps to reproduce
1. Abrir el reporte de `Saldos Abiertos`.
2. Mostrar los campos de fecha (todos estΓ‘n marcados como rango).
3. Establecer los valores en las fechas.
#### Screenshot or GifοΌζͺεΎζε¨ζεΎοΌ
https://github.com/solop-develop/frontend-core/assets/20288327/ff36d84e-a97f-40c5-9c83-1b8d39e843ac
#### Expected behavior
Observe como no se puede establecer el valor desde y hasta, solo se establece un valor por campo.
|
1.0
|
[Bug Report] Reporte/Proceso: Parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores - <!--
Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed.
-->
## Bug report
#### Steps to reproduce
1. Abrir el reporte de `Saldos Abiertos`.
2. Mostrar los campos de fecha (todos estΓ‘n marcados como rango).
3. Establecer los valores en las fechas.
#### Screenshot or GifοΌζͺεΎζε¨ζεΎοΌ
https://github.com/solop-develop/frontend-core/assets/20288327/ff36d84e-a97f-40c5-9c83-1b8d39e843ac
#### Expected behavior
Observe como no se puede establecer el valor desde y hasta, solo se establece un valor por campo.
|
process
|
reporte proceso parametro rango tipo fecha no muestra la selecciΓ³n en rango de valores note in order to better solve your problem please refer to the template to provide complete information accurately describe the problem and the incomplete information issue will be closed bug report steps to reproduce abrir el reporte de saldos abiertos mostrar los campos de fecha todos estΓ‘n marcados como rango establecer los valores en las fechas screenshot or gifοΌζͺεΎζε¨ζεΎοΌ expected behavior observe como no se puede establecer el valor desde y hasta solo se establece un valor por campo
| 1
|
14,623
| 17,766,788,285
|
IssuesEvent
|
2021-08-30 08:33:58
|
googleapis/nodejs-pubsub
|
https://api.github.com/repos/googleapis/nodejs-pubsub
|
closed
|
Dependency Dashboard
|
type: process api: pubsub
|
This issue contains a list of Renovate updates and their statuses.
## Awaiting Schedule
These updates are awaiting their schedule. Click on a checkbox to get an update now.
- [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/1123)
- [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/1302) (`mocha`, `@types/mocha`)
- [ ] <!-- recreate-branch=renovate/yargs-17.x -->[chore(deps): update dependency yargs to v17](../pull/1277)
- [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/1273)
- [ ] <!-- recreate-branch=renovate/p-defer-4.x -->[fix(deps): update dependency p-defer to v4](../pull/1250)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Awaiting Schedule
These updates are awaiting their schedule. Click on a checkbox to get an update now.
- [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/1123)
- [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/1302) (`mocha`, `@types/mocha`)
- [ ] <!-- recreate-branch=renovate/yargs-17.x -->[chore(deps): update dependency yargs to v17](../pull/1277)
- [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/1273)
- [ ] <!-- recreate-branch=renovate/p-defer-4.x -->[fix(deps): update dependency p-defer to v4](../pull/1250)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull mocha types mocha pull pull pull check this box to trigger a request for renovate to run again on this repository
| 1
|
10,437
| 13,220,071,105
|
IssuesEvent
|
2020-08-17 11:43:14
|
km4ack/pi-build
|
https://api.github.com/repos/km4ack/pi-build
|
closed
|
Check for location of GPS in script
|
in process
|
My GPS is located at "ttyUSB0" not the default of the script. Could the script check for the location? ls /dev/*USB* and use that to install the GPS. My GPS is a GlobalSat BU-353-S4 USB
|
1.0
|
Check for location of GPS in script - My GPS is located at "ttyUSB0" not the default of the script. Could the script check for the location? ls /dev/*USB* and use that to install the GPS. My GPS is a GlobalSat BU-353-S4 USB
|
process
|
check for location of gps in script my gps is located at not the default of the script could the script check for the location ls dev usb and use that to install the gps my gps is a globalsat bu usb
| 1
|
683,540
| 23,385,858,647
|
IssuesEvent
|
2022-08-11 13:42:35
|
TheYellowArchitect/doubledamnation
|
https://api.github.com/repos/TheYellowArchitect/doubledamnation
|
opened
|
Level Skip - Death Dialogue
|
bug needs more info low priority netcoding dialogue
|
If host is having a dialogue after death, and client joins right there, client cannot change state, and is on a permanently "Idle" state, regardless of movement.
I have not investigated what causes this. Inputs do pass to host (he can magejump properly) but probably disabled locally (as if in cutscene)
Regardless, it is low priority because its easily fixed by simply dying/reviving.
|
1.0
|
Level Skip - Death Dialogue - If host is having a dialogue after death, and client joins right there, client cannot change state, and is on a permanently "Idle" state, regardless of movement.
I have not investigated what causes this. Inputs do pass to host (he can magejump properly) but probably disabled locally (as if in cutscene)
Regardless, it is low priority because its easily fixed by simply dying/reviving.
|
non_process
|
level skip death dialogue if host is having a dialogue after death and client joins right there client cannot change state and is on a permanently idle state regardless of movement i have not investigated what causes this inputs do pass to host he can magejump properly but probably disabled locally as if in cutscene regardless it is low priority because its easily fixed by simply dying reviving
| 0
|
863
| 3,322,369,443
|
IssuesEvent
|
2015-11-09 14:13:40
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
opened
|
Keyref link breaks if target topic contains keyref
|
bug preprocess/keyref
|
`develop` branch. [Sample files](https://github.com/eerohele/dita-ot-issues/tree/master/fixtures/2133).
Easier to show than tell: the first link in `topic1.dita` is broken, the second one works. If you remove `<keyword keyref="keyword"/>` from `topic2.dita`, the first link starts working.
It looks like the `keyref` stage creates a file called `topic2-1.dita` in `temp`. It is an exact clone of `topic2.dita`. The subsequent preprocessing steps then use `topic2-1.dita` instead of `topic2.dita`. Didn't get further than that yet.
|
1.0
|
Keyref link breaks if target topic contains keyref - `develop` branch. [Sample files](https://github.com/eerohele/dita-ot-issues/tree/master/fixtures/2133).
Easier to show than tell: the first link in `topic1.dita` is broken, the second one works. If you remove `<keyword keyref="keyword"/>` from `topic2.dita`, the first link starts working.
It looks like the `keyref` stage creates a file called `topic2-1.dita` in `temp`. It is an exact clone of `topic2.dita`. The subsequent preprocessing steps then use `topic2-1.dita` instead of `topic2.dita`. Didn't get further than that yet.
|
process
|
keyref link breaks if target topic contains keyref develop branch easier to show than tell the first link in dita is broken the second one works if you remove from dita the first link starts working it looks like the keyref stage creates a file called dita in temp it is an exact clone of dita the subsequent preprocessing steps then use dita instead of dita didn t get further than that yet
| 1
|
13,282
| 15,761,497,271
|
IssuesEvent
|
2021-03-31 10:02:48
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
child_process 'close' always (?) emitted after 'exit'
|
child_process
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: master
* **Platform**: Linux 5.4.0-70-generic #78-Ubuntu SMP Fri Mar 19 13:29:52 UTC 2021 x86_64 x86_64 x86_64 GNU/Linuxd
* **Subsystem**: child_process
### What steps will reproduce the bug?
```js
const child_process = require('child_process');
const proc = child_process.spawn('bash', ['-c', 'exec 0>&- 1>&- 2>&-; sleep 5'], {
stdio: ['inherit', 'pipe', 'inherit']
});
proc.stdout.pipe(process.stdout);
proc.on('exit', () => console.log('exit'));
proc.on('close', () => console.log('close'));
```
### How often does it reproduce? Is there a required condition?
Always
### What is the expected behavior?
'close' should be emitted immediately (all stdio streams are closed), 'exit' after 5 seconds
### What do you see instead?
5 seconds pass, then 'exit' is emitted and then 'closed'
|
1.0
|
child_process 'close' always (?) emitted after 'exit' - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: master
* **Platform**: Linux 5.4.0-70-generic #78-Ubuntu SMP Fri Mar 19 13:29:52 UTC 2021 x86_64 x86_64 x86_64 GNU/Linuxd
* **Subsystem**: child_process
### What steps will reproduce the bug?
```js
const child_process = require('child_process');
const proc = child_process.spawn('bash', ['-c', 'exec 0>&- 1>&- 2>&-; sleep 5'], {
stdio: ['inherit', 'pipe', 'inherit']
});
proc.stdout.pipe(process.stdout);
proc.on('exit', () => console.log('exit'));
proc.on('close', () => console.log('close'));
```
### How often does it reproduce? Is there a required condition?
Always
### What is the expected behavior?
'close' should be emitted immediately (all stdio streams are closed), 'exit' after 5 seconds
### What do you see instead?
5 seconds pass, then 'exit' is emitted and then 'closed'
|
process
|
child process close always emitted after exit thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or output of osversion foreach object versionstring if else in powershell console windows subsystem if known please specify affected core module name version master platform linux generic ubuntu smp fri mar utc gnu linuxd subsystem child process what steps will reproduce the bug js const child process require child process const proc child process spawn bash stdio proc stdout pipe process stdout proc on exit console log exit proc on close console log close how often does it reproduce is there a required condition always what is the expected behavior close should be emitted immediately all stdio streams are closed exit after seconds what do you see instead seconds pass then exit is emitted and then closed
| 1
|
406,108
| 27,551,774,582
|
IssuesEvent
|
2023-03-07 15:22:19
|
MentorMate/node-project-cli
|
https://api.github.com/repos/MentorMate/node-project-cli
|
closed
|
Setup swagger generation for express apps
|
documentation express
|
# Description
The application should be able to produce a `swagger.json` file documenting its endpoints
# Tasks
- [x] integrate a `zod-to-openapi` package for generating OpenAPI definitions from zod schemas
- [ ] create a script and/or an endpoint that will produce the [OpenAPI document](https://swagger.io/specification/)
|
1.0
|
Setup swagger generation for express apps - # Description
The application should be able to produce a `swagger.json` file documenting its endpoints
# Tasks
- [x] integrate a `zod-to-openapi` package for generating OpenAPI definitions from zod schemas
- [ ] create a script and/or an endpoint that will produce the [OpenAPI document](https://swagger.io/specification/)
|
non_process
|
setup swagger generation for express apps description the application should be able to produce a swagger json file documenting its endpoints tasks integrate a zod to openapi package for generating openapi definitions from zod schemas create a script and or an endpoint that will produce the
| 0
|
4,529
| 7,371,636,803
|
IssuesEvent
|
2018-03-13 12:28:26
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Broken Link
|
cxp doc-bug in-process media-services triaged
|
The link under https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download
That is linked to: http://southworks.com/blog/2015/05/27/reusing-azure-media-services-locators-to-avoid-facing-the-5-shared-access-policy-limitation/
... is broken.
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: ca2f89cc-d993-eed3-ebf0-c7bfaa7716e0
* Version Independent ID: fd773e3d-6581-c356-ddd0-f9b609d63d7a
* Content: [Get started with delivering content on demand using .NET | Microsoft Docs](https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download)
* Content Source: [articles/media-services/media-services-dotnet-get-started.md](https://github.com/Microsoft/azure-docs/blob/master/articles/media-services/media-services-dotnet-get-started.md)
* Service: **media-services**
* GitHub Login: @Juliako
* Microsoft Alias: **juliako**
|
1.0
|
Broken Link - The link under https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download
That is linked to: http://southworks.com/blog/2015/05/27/reusing-azure-media-services-locators-to-avoid-facing-the-5-shared-access-policy-limitation/
... is broken.
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: ca2f89cc-d993-eed3-ebf0-c7bfaa7716e0
* Version Independent ID: fd773e3d-6581-c356-ddd0-f9b609d63d7a
* Content: [Get started with delivering content on demand using .NET | Microsoft Docs](https://docs.microsoft.com/en-us/azure/media-services/media-services-dotnet-get-started#publish-the-asset-and-get-urls-for-streaming-and-progressive-download)
* Content Source: [articles/media-services/media-services-dotnet-get-started.md](https://github.com/Microsoft/azure-docs/blob/master/articles/media-services/media-services-dotnet-get-started.md)
* Service: **media-services**
* GitHub Login: @Juliako
* Microsoft Alias: **juliako**
|
process
|
broken link the link under that is linked to is broken document details β do not edit this section it is required for docs microsoft com β github issue linking id version independent id content content source service media services github login juliako microsoft alias juliako
| 1
|
11,382
| 14,222,882,162
|
IssuesEvent
|
2020-11-17 17:26:57
|
JuliaParallel/Dagger.jl
|
https://api.github.com/repos/JuliaParallel/Dagger.jl
|
opened
|
Multithread and distribute the scheduler
|
performance processors scheduler
|
The scheduler is quite close to being able to run multithreaded, with many parts running concurrently already. We're also already pushing some extra work onto the worker nodes (mainly argument fetching and processor selection/load balancing), and it would be beneficial for large DAGs to move more work onto each worker. The main blocker is providing a way to split the DAG into multiple domains, where each domain is handled by a given thread on a given worker. With efficient Thunk serialization, we can then send a subgraph to each worker and let them process their own DAG without conflicts. We'll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data; if possible, we can also have workers broadcast and shard their `Chunk`s onto dependent workers as soon as the data is made available.
|
1.0
|
Multithread and distribute the scheduler - The scheduler is quite close to being able to run multithreaded, with many parts running concurrently already. We're also already pushing some extra work onto the worker nodes (mainly argument fetching and processor selection/load balancing), and it would be beneficial for large DAGs to move more work onto each worker. The main blocker is providing a way to split the DAG into multiple domains, where each domain is handled by a given thread on a given worker. With efficient Thunk serialization, we can then send a subgraph to each worker and let them process their own DAG without conflicts. We'll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data; if possible, we can also have workers broadcast and shard their `Chunk`s onto dependent workers as soon as the data is made available.
|
process
|
multithread and distribute the scheduler the scheduler is quite close to being able to run multithreaded with many parts running concurrently already we re also already pushing some extra work onto the worker nodes mainly argument fetching and processor selection load balancing and it would be beneficial for large dags to move more work onto each worker the main blocker is providing a way to split the dag into multiple domains where each domain is handled by a given thread on a given worker with efficient thunk serialization we can then send a subgraph to each worker and let them process their own dag without conflicts we ll need to add a mechanism by which thunks automatically wait on their input thunks to complete before they attempt to download the output data if possible we can also have workers broadcast and shard their chunk s onto dependent workers as soon as the data is made available
| 1
|
148,686
| 13,244,199,643
|
IssuesEvent
|
2020-08-19 12:42:26
|
OraOpenSource/apex-nitro
|
https://api.github.com/repos/OraOpenSource/apex-nitro
|
closed
|
minor: Upate warning message for openApp
|
documentation enhancement
|
When running in headless mode the following warning message is shown:
`[APEX Nitro] Couldn't open browser (if you are using BrowserSync in a headless environment, you might want to set the open option to false)`
The config file has this option as `launch.openApp`. Perhaps update the message with that instead of just `open`.
|
1.0
|
minor: Upate warning message for openApp - When running in headless mode the following warning message is shown:
`[APEX Nitro] Couldn't open browser (if you are using BrowserSync in a headless environment, you might want to set the open option to false)`
The config file has this option as `launch.openApp`. Perhaps update the message with that instead of just `open`.
|
non_process
|
minor upate warning message for openapp when running in headless mode the following warning message is shown couldn t open browser if you are using browsersync in a headless environment you might want to set the open option to false the config file has this option as launch openapp perhaps update the message with that instead of just open
| 0
|
345,815
| 10,373,588,218
|
IssuesEvent
|
2019-09-09 07:40:07
|
jenkins-x/jx
|
https://api.github.com/repos/jenkins-x/jx
|
closed
|
use the Version Stream to manage versions of client side packages l
|
area/versions lifecycle/stale priority/important-soon
|
we should lock down client side binaries like `helm, git, kubectl` so that if folks have ancient or mismatched versions we warn/recommend upgrading/downgrading / fail
https://jenkins-x.io/architecture/version-stream/
|
1.0
|
use the Version Stream to manage versions of client side packages l - we should lock down client side binaries like `helm, git, kubectl` so that if folks have ancient or mismatched versions we warn/recommend upgrading/downgrading / fail
https://jenkins-x.io/architecture/version-stream/
|
non_process
|
use the version stream to manage versions of client side packages l we should lock down client side binaries like helm git kubectl so that if folks have ancient or mismatched versions we warn recommend upgrading downgrading fail
| 0
|
5,480
| 8,355,650,927
|
IssuesEvent
|
2018-10-02 16:13:39
|
pelias/acceptance-tests
|
https://api.github.com/repos/pelias/acceptance-tests
|
closed
|
Add acceptance tests for Eraser Map usage
|
processed
|
- [x] [`/autocomplete + focus.point + text`](https://search.mapzen.com/v1/autocomplete?text=wyt&focus.point.lat=40.744569&focus.point.lon=-73.9902313&api_key=search-BcgTWes)
- [ ] [`/search + focus.viewport + text`](https://search.mapzen.com/v1/search?text=wyth&focus.viewport.min_lat=40.73644229295322&focus.viewport.min_lon=-73.99405878124092&focus.viewport.max_lat=40.74684717132519&focus.viewport.max_lon=-73.9863340187591&api_key=search-BcgTWes)
- [ ] [`/reverse + point`](https://search.mapzen.com/v1/reverse?point.lat=40.74549540292336&point.lon=-73.98920473153952&api_key=search-BcgTWes)
- [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=36.57751244954497&point.lon=-68.0269221649082&api_key=search-BcgTWes)
- [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=31.181260064156664&point.lon=-35.36817462744051&api_key=search-BcgTWes)
- [ ] test for privacy headers: maybe check that there is a `private` flag sent back as part of the response?
|
1.0
|
Add acceptance tests for Eraser Map usage - - [x] [`/autocomplete + focus.point + text`](https://search.mapzen.com/v1/autocomplete?text=wyt&focus.point.lat=40.744569&focus.point.lon=-73.9902313&api_key=search-BcgTWes)
- [ ] [`/search + focus.viewport + text`](https://search.mapzen.com/v1/search?text=wyth&focus.viewport.min_lat=40.73644229295322&focus.viewport.min_lon=-73.99405878124092&focus.viewport.max_lat=40.74684717132519&focus.viewport.max_lon=-73.9863340187591&api_key=search-BcgTWes)
- [ ] [`/reverse + point`](https://search.mapzen.com/v1/reverse?point.lat=40.74549540292336&point.lon=-73.98920473153952&api_key=search-BcgTWes)
- [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=36.57751244954497&point.lon=-68.0269221649082&api_key=search-BcgTWes)
- [ ] [`/reverse + point` away fromNYC](https://search.mapzen.com/v1/reverse?point.lat=31.181260064156664&point.lon=-35.36817462744051&api_key=search-BcgTWes)
- [ ] test for privacy headers: maybe check that there is a `private` flag sent back as part of the response?
|
process
|
add acceptance tests for eraser map usage test for privacy headers maybe check that there is a private flag sent back as part of the response
| 1
|
91,900
| 3,863,516,893
|
IssuesEvent
|
2016-04-08 09:45:42
|
iamxavier/elmah
|
https://api.github.com/repos/iamxavier/elmah
|
closed
|
Use of C#-SQLite library
|
auto-migrated Priority-Medium Type-Enhancement
|
```
What new or enhanced feature are you proposing?
Using the csharp-sqlite library instead of the existing library. C#-Sqlite is
written in C# and is 100% compatible with SQLite (according to the unit tests).
It's SQLite written in C#.
http://code.google.com/p/csharp-sqlite/
I'd be more than happy to make this change.
What goal would this enhancement help you achieve?
It removes the dependence on a 32-bit or 64-bit SQLite DLL. After all, isn't
platform independence one of the reasons we use C#? Dealing with the 32/64 bit
issue is a pain when deploying across that barrier.
The downside is perhaps the code will go stale. This isn't much of an issue
since this is a risk even with the existing library. In fact, the existing one
did go dormant until SQLite took over maintenance. There was even talk about
it using the C#-SQLite library instead of the SQLite DLLs.
```
Original issue reported on code.google.com by `masil...@gmail.com` on 5 Dec 2011 at 5:44
|
1.0
|
Use of C#-SQLite library - ```
What new or enhanced feature are you proposing?
Using the csharp-sqlite library instead of the existing library. C#-Sqlite is
written in C# and is 100% compatible with SQLite (according to the unit tests).
It's SQLite written in C#.
http://code.google.com/p/csharp-sqlite/
I'd be more than happy to make this change.
What goal would this enhancement help you achieve?
It removes the dependence on a 32-bit or 64-bit SQLite DLL. After all, isn't
platform independence one of the reasons we use C#? Dealing with the 32/64 bit
issue is a pain when deploying across that barrier.
The downside is perhaps the code will go stale. This isn't much of an issue
since this is a risk even with the existing library. In fact, the existing one
did go dormant until SQLite took over maintenance. There was even talk about
it using the C#-SQLite library instead of the SQLite DLLs.
```
Original issue reported on code.google.com by `masil...@gmail.com` on 5 Dec 2011 at 5:44
|
non_process
|
use of c sqlite library what new or enhanced feature are you proposing using the csharp sqlite library instead of the existing library c sqlite is written in c and is compatible with sqlite according to the unit tests it s sqlite written in c i d be more than happy to make this change what goal would this enhancement help you achieve it removes the dependence on a bit or bit sqlite dll after all isn t platform independence one of the reasons we use c dealing with the bit issue is a pain when deploying across that barrier the downside is perhaps the code will go stale this isn t much of an issue since this is a risk even with the existing library in fact the existing one did go dormant until sqlite took over maintenance there was even talk about it using the c sqlite library instead of the sqlite dlls original issue reported on code google com by masil gmail com on dec at
| 0
|
444,134
| 12,806,862,475
|
IssuesEvent
|
2020-07-03 10:15:44
|
enso-org/enso
|
https://api.github.com/repos/enso-org/enso
|
closed
|
Implement Undo
|
Category: Tooling Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement
|
### Summary
With the design for undo specified as part of luna/enso#551, we now need to implement the underlying functionality.
### Value
Clients will be able to undo edits.
### Specification
- [ ] Implement the underlying functionality for tracking user action stacks as specified in the [design document](https://github.com/luna/enso/blob/master/doc/design/engine/engine-services.md).
- [ ] Implement the `text/undo` message as specified in the same design document.
- [ ] Document any errors that are added as part of this proposal in the design document.
### Acceptance Criteria & Test Cases
- The undo functionality has been implemented.
- The undo functionality has been rigorously tested.
|
1.0
|
Implement Undo - ### Summary
With the design for undo specified as part of luna/enso#551, we now need to implement the underlying functionality.
### Value
Clients will be able to undo edits.
### Specification
- [ ] Implement the underlying functionality for tracking user action stacks as specified in the [design document](https://github.com/luna/enso/blob/master/doc/design/engine/engine-services.md).
- [ ] Implement the `text/undo` message as specified in the same design document.
- [ ] Document any errors that are added as part of this proposal in the design document.
### Acceptance Criteria & Test Cases
- The undo functionality has been implemented.
- The undo functionality has been rigorously tested.
|
non_process
|
implement undo summary with the design for undo specified as part of luna enso we now need to implement the underlying functionality value clients will be able to undo edits specification implement the underlying functionality for tracking user action stacks as specified in the implement the text undo message as specified in the same design document document any errors that are added as part of this proposal in the design document acceptance criteria test cases the undo functionality has been implemented the undo functionality has been rigorously tested
| 0
|
17,059
| 22,492,218,193
|
IssuesEvent
|
2022-06-23 03:13:07
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Set M Value processing tool does not behave as expected
|
Feedback stale Processing Bug
|
### What is the bug or the crash?
The [documentation and help](https://docs.qgis.org/3.22/en/docs/user_manual/processing_algs/qgis/vectorgeometry.html#qgissetmvalue) for the Set M Value processing tool states that "If no M values exist, the geometry will be upgraded to include M values and the specified value used as the initial M value for all geometries". Existing features have their M value updated according to the value set in the tool. New features created default to a NULL M value regardless of what is set using the processing tool. I've tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool.
### Steps to reproduce the issue
Go to processing toolbox > set m values. Choose the input layer, set the m value to a number. Click Run. With the created temporary layer begin editing and add a new feature. Save your edits and stop editing.
Again using the processing toolbox find the tool "Extract M Values". Set the summaries to first, last and count and extract the M values. Open the attribute table of the resultant layer. Newly added features through editing will have an M Value of NULL. Features existing prior to running the set M values tool will have the M values as set using the set M values tools.
### Versions
3.16.9
3.22.3
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
It may be that this is the desired behaviour in which case I would suggest an update to the documentation to make this clear. I'd be happy to make said update.
|
1.0
|
Set M Value processing tool does not behave as expected - ### What is the bug or the crash?
The [documentation and help](https://docs.qgis.org/3.22/en/docs/user_manual/processing_algs/qgis/vectorgeometry.html#qgissetmvalue) for the Set M Value processing tool states that "If no M values exist, the geometry will be upgraded to include M values and the specified value used as the initial M value for all geometries". Existing features have their M value updated according to the value set in the tool. New features created default to a NULL M value regardless of what is set using the processing tool. I've tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool.
### Steps to reproduce the issue
Go to processing toolbox > set m values. Choose the input layer, set the m value to a number. Click Run. With the created temporary layer begin editing and add a new feature. Save your edits and stop editing.
Again using the processing toolbox find the tool "Extract M Values". Set the summaries to first, last and count and extract the M values. Open the attribute table of the resultant layer. Newly added features through editing will have an M Value of NULL. Features existing prior to running the set M values tool will have the M values as set using the set M values tools.
### Versions
3.16.9
3.22.3
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
It may be that this is the desired behaviour in which case I would suggest an update to the documentation to make this clear. I'd be happy to make said update.
|
process
|
set m value processing tool does not behave as expected what is the bug or the crash the for the set m value processing tool states that if no m values exist the geometry will be upgraded to include m values and the specified value used as the initial m value for all geometries existing features have their m value updated according to the value set in the tool new features created default to a null m value regardless of what is set using the processing tool i ve tested this with geopackages and shapefiles for line and multiline geometry types both with m value capability initially created and added later using the tool steps to reproduce the issue go to processing toolbox set m values choose the input layer set the m value to a number click run with the created temporary layer begin editing and add a new feature save your edits and stop editing again using the processing toolbox find the tool extract m values set the summaries to first last and count and extract the m values open the attribute table of the resultant layer newly added features through editing will have an m value of null features existing prior to running the set m values tool will have the m values as set using the set m values tools versions supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context it may be that this is the desired behaviour in which case i would suggest an update to the documentation to make this clear i d be happy to make said update
| 1
|
12,966
| 15,342,672,656
|
IssuesEvent
|
2021-02-27 17:09:56
|
topcoder-platform/community-app
|
https://api.github.com/repos/topcoder-platform/community-app
|
opened
|
Filter Challenge Type is displayed as Type
|
P4 ShapeupProcess challenge- recommender-tool
|
Filter `Challenge Type` is displayed as `Type`
<img width="1440" alt="Screenshot 2021-02-27 at 10 36 47 PM" src="https://user-images.githubusercontent.com/58783823/109394368-a228f280-794c-11eb-812e-6b46fcb16617.png">
|
1.0
|
Filter Challenge Type is displayed as Type - Filter `Challenge Type` is displayed as `Type`
<img width="1440" alt="Screenshot 2021-02-27 at 10 36 47 PM" src="https://user-images.githubusercontent.com/58783823/109394368-a228f280-794c-11eb-812e-6b46fcb16617.png">
|
process
|
filter challenge type is displayed as type filter challenge type is displayed as type img width alt screenshot at pm src
| 1
|
108,689
| 16,812,562,592
|
IssuesEvent
|
2021-06-17 01:01:58
|
RG4421/ampere-centos-kernel
|
https://api.github.com/repos/RG4421/ampere-centos-kernel
|
opened
|
CVE-2021-0512 (High) detected in linuxv5.2
|
security vulnerability
|
## CVE-2021-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in android kernel common. The HID subsystem allows an "HID report field" to have a different
number of "values" and "usages" when it is allocated. When a field struct is created, the size of the usage array is guaranteed to be at least as large as the values array, but it may be larger. This leads to a potential out-of-bounds write in.
<p>Publish Date: 2020-11-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0512>CVE-2021-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-06-01">https://source.android.com/security/bulletin/2021-06-01</a></p>
<p>Release Date: 2020-11-07</p>
<p>Fix Resolution: ASB-2021-02-05_mainline</p>
</p>
</details>
<p></p>
|
True
|
CVE-2021-0512 (High) detected in linuxv5.2 - ## CVE-2021-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in android kernel common. The HID subsystem allows an "HID report field" to have a different
number of "values" and "usages" when it is allocated. When a field struct is created, the size of the usage array is guaranteed to be at least as large as the values array, but it may be larger. This leads to a potential out-of-bounds write in.
<p>Publish Date: 2020-11-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0512>CVE-2021-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-06-01">https://source.android.com/security/bulletin/2021-06-01</a></p>
<p>Release Date: 2020-11-07</p>
<p>Fix Resolution: ASB-2021-02-05_mainline</p>
</p>
</details>
<p></p>
|
non_process
|
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files vulnerability details an issue was found in android kernel common the hid subsystem allows an hid report field to have a different number of values and usages when it is allocated when a field struct is created the size of the usage array is guaranteed to be at least as large as the values array but it may be larger this leads to a potential out of bounds write in publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution asb mainline
| 0
|
16,531
| 21,558,572,967
|
IssuesEvent
|
2022-04-30 21:01:35
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
Auto TLS Beta
|
kind/feature area/networking kind/process triage/accepted
|
<!-- If you need to report a security issue with Knative, send an email to knative-security@googlegroups.com. -->
<!--
## In what area(s)?
Remove the '> ' to select
> /area API
> /area autoscale
> /area build
> /area monitoring
/area networking
> /area test-and-release
Other classifications:
> /kind good-first-issue
/kind process
> /kind spec
-->
## Describe the feature
This is the issue to tracking the work for Auto TLS Beta.
Below are the work items I think we need for Auto TLS Beta:
1. Support cert-manager 0.11 and above with their v1alpha2 APIs https://github.com/knative/serving/issues/6011
2. E2E tests for Auto TLS feature https://github.com/knative/serving/issues/4066, including testing the following cases:
- Certificate provision per Knative Service
- Certificate provision per namespace
- Certificate provision with HTTP01 challenge
3. Conformance test for Certificate: https://github.com/knative/serving/issues/6445
4. Detailed documentation about Auto TLS feature and the modes it supports (DNS challenge/HTTP challenge, per ksvc/per namespace). Related issue: https://github.com/knative/docs/issues/1949
Feel free to comment in this thread if anyone has thoughts about the work items needed for Beta launch.
/cc @mattmoor @tcnghia
|
1.0
|
Auto TLS Beta - <!-- If you need to report a security issue with Knative, send an email to knative-security@googlegroups.com. -->
<!--
## In what area(s)?
Remove the '> ' to select
> /area API
> /area autoscale
> /area build
> /area monitoring
/area networking
> /area test-and-release
Other classifications:
> /kind good-first-issue
/kind process
> /kind spec
-->
## Describe the feature
This is the issue to tracking the work for Auto TLS Beta.
Below are the work items I think we need for Auto TLS Beta:
1. Support cert-manager 0.11 and above with their v1alpha2 APIs https://github.com/knative/serving/issues/6011
2. E2E tests for Auto TLS feature https://github.com/knative/serving/issues/4066, including testing the following cases:
- Certificate provision per Knative Service
- Certificate provision per namespace
- Certificate provision with HTTP01 challenge
3. Conformance test for Certificate: https://github.com/knative/serving/issues/6445
4. Detailed documentation about Auto TLS feature and the modes it supports (DNS challenge/HTTP challenge, per ksvc/per namespace). Related issue: https://github.com/knative/docs/issues/1949
Feel free to comment in this thread if anyone has thoughts about the work items needed for Beta launch.
/cc @mattmoor @tcnghia
|
process
|
auto tls beta in what area s remove the to select area api area autoscale area build area monitoring area networking area test and release other classifications kind good first issue kind process kind spec describe the feature this is the issue to tracking the work for auto tls beta below are the work items i think we need for auto tls beta support cert manager and above with their apis tests for auto tls feature including testing the following cases certificate provision per knative service certificate provision per namespace certificate provision with challenge conformance test for certificate detailed documentation about auto tls feature and the modes it supports dns challenge http challenge per ksvc per namespace related issue feel free to comment in this thread if anyone has thoughts about the work items needed for beta launch cc mattmoor tcnghia
| 1
|
8,494
| 2,993,742,531
|
IssuesEvent
|
2015-07-22 07:04:29
|
sunlightlabs/the-phantom-mask
|
https://api.github.com/repos/sunlightlabs/the-phantom-mask
|
opened
|
Improved copy needed in failure notifications
|
bug user testing
|
Failure message (in italics) is lengthy and awkwardly phrased.
Links to congresspersons' contact forms are unclear (and what to do with them). These links can be confusingβwhen a contact page is down or has moved, the user will seem to gets a dead/broken link.
|
1.0
|
Improved copy needed in failure notifications - Failure message (in italics) is lengthy and awkwardly phrased.
Links to congresspersons' contact forms are unclear (and what to do with them). These links can be confusingβwhen a contact page is down or has moved, the user will seem to gets a dead/broken link.
|
non_process
|
improved copy needed in failure notifications failure message in italics is lengthy and awkwardly phrased links to congresspersons contact forms are unclear and what to do with them these links can be confusingβwhen a contact page is down or has moved the user will seem to gets a dead broken link
| 0
|
34,941
| 4,599,883,598
|
IssuesEvent
|
2016-09-22 01:08:52
|
MozillaFoundation/Mozfest2016_production
|
https://api.github.com/repos/MozillaFoundation/Mozfest2016_production
|
closed
|
Test and Finalize One Web Theme Colours
|
Design Production
|
<img width="1027" alt="screen shot 2016-09-07 at 4 32 10 pm" src="https://cloud.githubusercontent.com/assets/2492510/18332095/ac28b5e6-7518-11e6-9788-08d12bdfc1a1.png">
We are using a bold palette with neons this year and we need to test:
- for print in CMYK
- for print in Pantone
- How the palette works as a whole (example: what if yellow was neon and the rest were not?)
- finalize the palette with referencing a pantone book or printed samples
We should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do.
|
1.0
|
Test and Finalize One Web Theme Colours - <img width="1027" alt="screen shot 2016-09-07 at 4 32 10 pm" src="https://cloud.githubusercontent.com/assets/2492510/18332095/ac28b5e6-7518-11e6-9788-08d12bdfc1a1.png">
We are using a bold palette with neons this year and we need to test:
- for print in CMYK
- for print in Pantone
- How the palette works as a whole (example: what if yellow was neon and the rest were not?)
- finalize the palette with referencing a pantone book or printed samples
We should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do.
|
non_process
|
test and finalize one web theme colours img width alt screen shot at pm src we are using a bold palette with neons this year and we need to test for print in cmyk for print in pantone how the palette works as a whole example what if yellow was neon and the rest were not finalize the palette with referencing a pantone book or printed samples we should identify in spreadsheet what key pieces have a a higher budget for special printing and connect with what the printers can do
| 0
|
20,339
| 26,994,633,445
|
IssuesEvent
|
2023-02-09 23:16:59
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
multiprocessing's "spawn" doesn't actually use spawn
|
type-bug stdlib 3.11 3.10 3.9 expert-multiprocessing
|
BPO | [46367](https://bugs.python.org/issue46367)
--- | :---
Nosy | @vstinner, @jakirkham
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2022-01-13.19:20:59.356>
labels = ['type-bug', 'library', '3.9', '3.10', '3.11']
title = 'multiprocessing\'s "spawn" doesn\'t actually use spawn'
updated_at = <Date 2022-01-15.00:10:58.327>
user = 'https://github.com/jakirkham'
```
bugs.python.org fields:
```python
activity = <Date 2022-01-15.00:10:58.327>
actor = 'vstinner'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Library (Lib)']
creation = <Date 2022-01-13.19:20:59.356>
creator = 'jakirkham'
dependencies = []
files = []
hgrepos = []
issue_num = 46367
keywords = []
message_count = 2.0
messages = ['410512', '410612']
nosy_count = 2.0
nosy_names = ['vstinner', 'jakirkham']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue46367'
versions = ['Python 3.9', 'Python 3.10', 'Python 3.11']
```
</p></details>
|
1.0
|
multiprocessing's "spawn" doesn't actually use spawn - BPO | [46367](https://bugs.python.org/issue46367)
--- | :---
Nosy | @vstinner, @jakirkham
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2022-01-13.19:20:59.356>
labels = ['type-bug', 'library', '3.9', '3.10', '3.11']
title = 'multiprocessing\'s "spawn" doesn\'t actually use spawn'
updated_at = <Date 2022-01-15.00:10:58.327>
user = 'https://github.com/jakirkham'
```
bugs.python.org fields:
```python
activity = <Date 2022-01-15.00:10:58.327>
actor = 'vstinner'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Library (Lib)']
creation = <Date 2022-01-13.19:20:59.356>
creator = 'jakirkham'
dependencies = []
files = []
hgrepos = []
issue_num = 46367
keywords = []
message_count = 2.0
messages = ['410512', '410612']
nosy_count = 2.0
nosy_names = ['vstinner', 'jakirkham']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue46367'
versions = ['Python 3.9', 'Python 3.10', 'Python 3.11']
```
</p></details>
|
process
|
multiprocessing s spawn doesn t actually use spawn bpo nosy vstinner jakirkham note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing s spawn doesn t actually use spawn updated at user bugs python org fields python activity actor vstinner assignee none closed false closed date none closer none components creation creator jakirkham dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions
| 1
|
196,143
| 22,440,849,364
|
IssuesEvent
|
2022-06-21 01:05:52
|
kedacore/test-tools
|
https://api.github.com/repos/kedacore/test-tools
|
opened
|
CVE-2022-33987 (Medium) detected in got-11.8.2.tgz
|
security vulnerability
|
## CVE-2022-33987 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>got-11.8.2.tgz</b></p></summary>
<p>Human-friendly and powerful HTTP request library for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-11.8.2.tgz">https://registry.npmjs.org/got/-/got-11.8.2.tgz</a></p>
<p>Path to dependency file: /e2e/images/selenium-grid/package.json</p>
<p>Path to vulnerable library: /e2e/images/selenium-grid/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- cli-7.8.0.tgz (Root Library)
- webdriverio-7.8.0.tgz
- webdriver-7.8.0.tgz
- :x: **got-11.8.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-33987 (Medium) detected in got-11.8.2.tgz - ## CVE-2022-33987 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>got-11.8.2.tgz</b></p></summary>
<p>Human-friendly and powerful HTTP request library for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-11.8.2.tgz">https://registry.npmjs.org/got/-/got-11.8.2.tgz</a></p>
<p>Path to dependency file: /e2e/images/selenium-grid/package.json</p>
<p>Path to vulnerable library: /e2e/images/selenium-grid/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- cli-7.8.0.tgz (Root Library)
- webdriverio-7.8.0.tgz
- webdriver-7.8.0.tgz
- :x: **got-11.8.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in got tgz cve medium severity vulnerability vulnerable library got tgz human friendly and powerful http request library for node js library home page a href path to dependency file images selenium grid package json path to vulnerable library images selenium grid node modules got package json dependency hierarchy cli tgz root library webdriverio tgz webdriver tgz x got tgz vulnerable library found in head commit a href found in base branch main vulnerability details the got package before for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend
| 0
|
14,342
| 17,369,298,841
|
IssuesEvent
|
2021-07-30 11:49:51
|
encode/uvicorn
|
https://api.github.com/repos/encode/uvicorn
|
closed
|
Sending SIGTERM to parent process when running with --workers hangs indefinitely
|
bug multiprocessing
|
### Checklist
<!-- Please make sure you check all these items before submitting your bug report. -->
- [x] The bug is reproducible against the latest release and/or `master`.
- [x] There are no similar issues or pull requests to fix it yet.
### Describe the bug
<!-- A clear and concise description of what the bug is. -->
When running uvicorn with multiple workers, sending a SIGTERM ie `kill -15 ppid` there's no graceful shutdown and the process hangs indefinitely on https://github.com/encode/uvicorn/blob/ff4af12d6902bc9d535fe2a948d1df3ffa02b0d3/uvicorn/supervisors/multiprocess.py#L57
### To reproduce
<!-- Provide a *minimal* example with steps to reproduce the bug locally.
1. Run `uvicorn app:app --workers 2 --log-level=debug`
2. Get the ppid in the logs `INFO: Started parent process [38237]`
3. Send the SIGTERM `kill -15 38237`
NOTE: try to keep any external dependencies *at an absolute minimum* .
In other words, remove anything that doesn't make the bug go away.
-->
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
A graceful shutdown
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
### Debugging material
<!-- Any tracebacks, screenshots, etc. that can help understanding the problem.
NOTE:
- Please list tracebacks in full (don't truncate them).
- If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag).
- Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d).
-->
### Environment
- OS / Python / Uvicorn version: just run `uvicorn --version`
`Running uvicorn 0.12.2 with CPython 3.8.6 on Linux`
- The exact command you're running uvicorn with, all flags you passed included. If you run it with gunicorn please do the same. If there is a reverse-proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce.
### Additional context
<!-- Any additional information that can help understanding the problem.
Eg. linked issues, or a description of what you were trying to achieve. -->
originally discovered in https://github.com/encode/uvicorn/issues/364 and mistakenly taken for a docker issue, it's in fact a SIGTERM issue afaiu
|
1.0
|
Sending SIGTERM to parent process when running with --workers hangs indefinitely - ### Checklist
<!-- Please make sure you check all these items before submitting your bug report. -->
- [x] The bug is reproducible against the latest release and/or `master`.
- [x] There are no similar issues or pull requests to fix it yet.
### Describe the bug
<!-- A clear and concise description of what the bug is. -->
When running uvicorn with multiple workers, sending a SIGTERM ie `kill -15 ppid` there's no graceful shutdown and the process hangs indefinitely on https://github.com/encode/uvicorn/blob/ff4af12d6902bc9d535fe2a948d1df3ffa02b0d3/uvicorn/supervisors/multiprocess.py#L57
### To reproduce
<!-- Provide a *minimal* example with steps to reproduce the bug locally.
1. Run `uvicorn app:app --workers 2 --log-level=debug`
2. Get the ppid in the logs `INFO: Started parent process [38237]`
3. Send the SIGTERM `kill -15 38237`
NOTE: try to keep any external dependencies *at an absolute minimum* .
In other words, remove anything that doesn't make the bug go away.
-->
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
A graceful shutdown
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
### Debugging material
<!-- Any tracebacks, screenshots, etc. that can help understanding the problem.
NOTE:
- Please list tracebacks in full (don't truncate them).
- If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag).
- Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d).
-->
### Environment
- OS / Python / Uvicorn version: just run `uvicorn --version`
`Running uvicorn 0.12.2 with CPython 3.8.6 on Linux`
- The exact command you're running uvicorn with, all flags you passed included. If you run it with gunicorn please do the same. If there is a reverse-proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce.
### Additional context
<!-- Any additional information that can help understanding the problem.
Eg. linked issues, or a description of what you were trying to achieve. -->
originally discovered in https://github.com/encode/uvicorn/issues/364 and mistakenly taken for a docker issue, it's in fact a SIGTERM issue afaiu
|
process
|
sending sigterm to parent process when running with workers hangs indefinitely checklist the bug is reproducible against the latest release and or master there are no similar issues or pull requests to fix it yet describe the bug when running uvicorn with multiple workers sending a sigterm ie kill ppid there s no graceful shutdown and the process hangs indefinitely on to reproduce provide a minimal example with steps to reproduce the bug locally run uvicorn app app workers log level debug get the ppid in the logs info started parent process send the sigterm kill note try to keep any external dependencies at an absolute minimum in other words remove anything that doesn t make the bug go away expected behavior a graceful shutdown actual behavior debugging material any tracebacks screenshots etc that can help understanding the problem note please list tracebacks in full don t truncate them if relevant consider turning on debug or trace logs for additional details see the logging section on specifically the log level flag consider using to make tracebacks logs collapsible if they re very large see environment os python uvicorn version just run uvicorn version running uvicorn with cpython on linux the exact command you re running uvicorn with all flags you passed included if you run it with gunicorn please do the same if there is a reverse proxy involved and you cannot reproduce without it please give the minimal config of it to reproduce additional context any additional information that can help understanding the problem eg linked issues or a description of what you were trying to achieve originally discovered in and mistakenly taken for a docker issue it s in fact a sigterm issue afaiu
| 1
|
19,574
| 25,894,937,892
|
IssuesEvent
|
2022-12-14 21:27:44
|
googleapis/python-monitoring-dashboards
|
https://api.github.com/repos/googleapis/python-monitoring-dashboards
|
closed
|
Your .repo-metadata.json file has a problem π€
|
type: process api: monitoring repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan π:
* api_shortname 'monitoring-dashboards' invalid in .repo-metadata.json
βοΈ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem π€ - You have a problem with your .repo-metadata.json file:
Result of scan π:
* api_shortname 'monitoring-dashboards' invalid in .repo-metadata.json
βοΈ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem π€ you have a problem with your repo metadata json file result of scan π api shortname monitoring dashboards invalid in repo metadata json βοΈ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
7,136
| 10,278,674,188
|
IssuesEvent
|
2019-08-25 16:18:20
|
xethya/framework
|
https://api.github.com/repos/xethya/framework
|
opened
|
[publish process] Add "files" property to all packages
|
monorepo publish process
|
Packages are being published with folders such as `.rpt2_cache` or `coverage`.
|
1.0
|
[publish process] Add "files" property to all packages - Packages are being published with folders such as `.rpt2_cache` or `coverage`.
|
process
|
add files property to all packages packages are being published with folders such as cache or coverage
| 1
|
5,103
| 7,882,665,732
|
IssuesEvent
|
2018-06-27 00:00:29
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
Grammar - Mid statement is not correct
|
bug parse-tree-processing
|
Per https://msdn.microsoft.com/en-us/library/ee156918.aspx, this parser rule is not correct:
https://github.com/rubberduck-vba/Rubberduck/blob/e2468806b7da88f20ca28a4ea0f08edb92cacd78/Rubberduck.Parsing/Grammar/VBAParser.g4#L467
`ASSIGN` should be `EQ`. This should be fixed, and also a test added.
Related: Not sure if it's desirable to have `ASSIGN` defined in the lexer. The only other place it appears to be used is for named arguments, which could equally well be parsed with `COLON EQ`. IMO, having a lexer rule named `ASSIGN` suggests that `:=` is a general-use token in VBA, which it is not.
|
1.0
|
Grammar - Mid statement is not correct - Per https://msdn.microsoft.com/en-us/library/ee156918.aspx, this parser rule is not correct:
https://github.com/rubberduck-vba/Rubberduck/blob/e2468806b7da88f20ca28a4ea0f08edb92cacd78/Rubberduck.Parsing/Grammar/VBAParser.g4#L467
`ASSIGN` should be `EQ`. This should be fixed, and also a test added.
Related: Not sure if it's desirable to have `ASSIGN` defined in the lexer. The only other place it appears to be used is for named arguments, which could equally well be parsed with `COLON EQ`. IMO, having a lexer rule named `ASSIGN` suggests that `:=` is a general-use token in VBA, which it is not.
|
process
|
grammar mid statement is not correct per this parser rule is not correct assign should be eq this should be fixed and also a test added related not sure if it s desirable to have assign defined in the lexer the only other place it appears to be used is for named arguments which could equally well be parsed with colon eq imo having a lexer rule named assign suggests that is a general use token in vba which it is not
| 1
|
109,942
| 13,879,680,220
|
IssuesEvent
|
2020-10-17 15:29:38
|
kambiz-frounchi/Foodie
|
https://api.github.com/repos/kambiz-frounchi/Foodie
|
opened
|
Define Express routes
|
design
|
These are the routes that come to mind:
1. **HTML routes**:
- '/'
- '/feed'
- '/recipe'
- '/postrecipe'
- `/findrecipe'
- 'userprofile'
2. **API routes**:
-
|
1.0
|
Define Express routes - These are the routes that come to mind:
1. **HTML routes**:
- '/'
- '/feed'
- '/recipe'
- '/postrecipe'
- `/findrecipe'
- 'userprofile'
2. **API routes**:
-
|
non_process
|
define express routes these are the routes that come to mind html routes feed recipe postrecipe findrecipe userprofile api routes
| 0
|
4,744
| 7,603,030,396
|
IssuesEvent
|
2018-04-29 09:51:12
|
brucemiller/LaTeXML
|
https://api.github.com/repos/brucemiller/LaTeXML
|
opened
|
HTML subfigures via flexbox rather than tables?
|
enhancement postprocessing
|
I was inspecting an arXiv document (1802.06832) which has 4 subfigures a-d in its Results section and noticed they don't reflow in the HTML.
They are currently marked as 4 `td` table cells with class `ltx_subfigure` in the same `tr` row.
The fixed single-row display looks quite bad even on large displays, as it is significantly wider than the main article width. And on small displays you end up with a huge horizontal scroll, which is just as bad.
As I recently did some flexbox work with the moderncv binding, I'm wondering if @brucemiller would see that as a better mechanism for presenting the subfigures. That would give us decent reflow, and we can contain the max-width to a reasonable responsive width.
|
1.0
|
HTML subfigures via flexbox rather than tables? - I was inspecting an arXiv document (1802.06832) which has 4 subfigures a-d in its Results section and noticed they don't reflow in the HTML.
They are currently marked as 4 `td` table cells with class `ltx_subfigure` in the same `tr` row.
The fixed single-row display looks quite bad even on large displays, as it is significantly wider than the main article width. And on small displays you end up with a huge horizontal scroll, which is just as bad.
As I recently did some flexbox work with the moderncv binding, I'm wondering if @brucemiller would see that as a better mechanism for presenting the subfigures. That would give us decent reflow, and we can contain the max-width to a reasonable responsive width.
|
process
|
html subfigures via flexbox rather than tables i was inspecting an arxiv document which has subfigures a d in its results section and noticed they don t reflow in the html they are currently marked as td table cells with class ltx subfigure in the same tr row the fixed single row display looks quite bad even on large displays as it is significantly wider than the main article width and on small displays you end up with a huge horizontal scroll which is just as bad as i recently did some flexbox work with the moderncv binding i m wondering if brucemiller would see that as a better mechanism for presenting the subfigures that would give us decent reflow and we can contain the max width to a reasonable responsive width
| 1
|
57,900
| 6,560,279,965
|
IssuesEvent
|
2017-09-07 08:42:48
|
LiskHQ/lisk-js
|
https://api.github.com/repos/LiskHQ/lisk-js
|
closed
|
[178] Add tests for constants
|
easy test
|
Part of #178
Not the values themselves, but the shape of the exported object so we don't e.g. accidentally delete a constant we need.
|
1.0
|
[178] Add tests for constants - Part of #178
Not the values themselves, but the shape of the exported object so we don't e.g. accidentally delete a constant we need.
|
non_process
|
add tests for constants part of not the values themselves but the shape of the exported object so we don t e g accidentally delete a constant we need
| 0
|
20,546
| 27,195,718,281
|
IssuesEvent
|
2023-02-20 04:58:02
|
AvaloniaUI/Avalonia
|
https://api.github.com/repos/AvaloniaUI/Avalonia
|
closed
|
An error occurs when a font with a Chinese name is used
|
bug area-textprocessing
|
``` csharp
System.InvalidOperationException: Could not create glyph typeface for: 216-δΈι¦ζθ΄’η«δ½.
at Avalonia.Skia.FontManagerImpl.CreateGlyphTypeface(Typeface typeface) in /_/src/Skia/Avalonia.Skia/FontManagerImpl.cs:line 146
at Avalonia.Media.GlyphTypeface..ctor(Typeface typeface) in /_/src/Avalonia.Visuals/Media/GlyphTypeface.cs:line 9
at Avalonia.Media.FontManager.GetOrAddGlyphTypeface(Typeface typeface) in /_/src/Avalonia.Visuals/Media/FontManager.cs:line 99
at Avalonia.Media.Typeface.get_GlyphTypeface() in /_/src/Avalonia.Visuals/Media/Typeface.cs:line 69
at Avalonia.Media.TextFormatting.TextCharacters.TryGetRunProperties(ReadOnlySlice`1 text, Typeface typeface, Typeface defaultTypeface, Int32& count) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 132
at Avalonia.Media.TextFormatting.TextCharacters.CreateShapeableRun(ReadOnlySlice`1 text, TextRunProperties defaultProperties) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 62
at Avalonia.Media.TextFormatting.TextCharacters.GetShapeableCharacters() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 40
at Avalonia.Media.TextFormatting.TextFormatterImpl.FetchTextRuns(ITextSource textSource, Int32 firstTextSourceIndex, TextLineBreak previousLineBreak, TextLineBreak& nextLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 286
at Avalonia.Media.TextFormatting.TextFormatterImpl.FormatLine(ITextSource textSource, Int32 firstTextSourceIndex, Double paragraphWidth, TextParagraphProperties paragraphProperties, TextLineBreak previousLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 15
at Avalonia.Media.TextFormatting.TextLayout.UpdateLayout() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 225
at Avalonia.Media.TextFormatting.TextLayout..ctor(String text, Typeface typeface, Double fontSize, IBrush foreground, TextAlignment textAlignment, TextWrapping textWrapping, TextTrimming textTrimming, TextDecorationCollection textDecorations, Double maxWidth, Double maxHeight, Double lineHeight, Int32 maxLines, IReadOnlyList`1 textStyleOverrides) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 71
at Avalonia.Controls.TextBlock.CreateTextLayout(Size constraint, String text) in /_/src/Avalonia.Controls/TextBlock.cs:line 475
at Avalonia.Controls.TextBlock.get_TextLayout() in /_/src/Avalonia.Controls/TextBlock.cs:line 166
at Avalonia.Controls.TextBlock.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/TextBlock.cs:line 521
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Controls.StackPanel.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/StackPanel.cs:line 207
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding, Thickness borderThickness) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 39
at Avalonia.Controls.Presenters.ContentPresenter.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Presenters/ContentPresenter.cs:line 366
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46
at Avalonia.Controls.Decorator.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Decorator.cs:line 54
at Avalonia.Controls.Primitives.VisualLayerManager.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Primitives/VisualLayerManager.cs:line 133
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625
at Avalonia.Controls.Window.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Window.cs:line 937
at Avalonia.Controls.WindowBase.MeasureCore(Size availableSize) in /_/src/Avalonia.Controls/WindowBase.cs:line 247
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutManager.Measure(ILayoutable control) in /_/src/Avalonia.Layout/LayoutManager.cs:line 297
at Avalonia.Layout.LayoutManager.ExecuteInitialLayoutPass() in /_/src/Avalonia.Layout/LayoutManager.cs:line 184
at Avalonia.Controls.Window.ShowCore(Window parent) in /_/src/Avalonia.Controls/Window.cs:line 693
at Avalonia.Controls.Window.Show() in /_/src/Avalonia.Controls/Window.cs:line 631
at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.ShowMainWindow() in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 129
at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.Start(String[] args) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 118
at Avalonia.ClassicDesktopStyleApplicationLifetimeExtensions.StartWithClassicDesktopLifetime[T](T builder, String[] args, ShutdownMode shutdownMode) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 209
at LearnAvalonia.Program.Main(String[] args) in D:\Desktop\LearnAvalonia\LearnAvalonia\Program.cs:line 13
```
# That's my way of quoting
```xaml
FontFamily="resm:LearnAvalonia.Assets.Fonts.No.216-ShangShouZhaoCaiMaoTi-2.ttf?assembly=LearnAvalonia#216-δΈι¦ζθ΄’η«δ½"
```
<img width="911" alt="image" src="https://user-images.githubusercontent.com/108921607/219945857-7588a20d-c66f-4fca-8930-9dea9e9b22ea.png">
|
1.0
|
An error occurs when a font with a Chinese name is used - ``` csharp
System.InvalidOperationException: Could not create glyph typeface for: 216-δΈι¦ζθ΄’η«δ½.
at Avalonia.Skia.FontManagerImpl.CreateGlyphTypeface(Typeface typeface) in /_/src/Skia/Avalonia.Skia/FontManagerImpl.cs:line 146
at Avalonia.Media.GlyphTypeface..ctor(Typeface typeface) in /_/src/Avalonia.Visuals/Media/GlyphTypeface.cs:line 9
at Avalonia.Media.FontManager.GetOrAddGlyphTypeface(Typeface typeface) in /_/src/Avalonia.Visuals/Media/FontManager.cs:line 99
at Avalonia.Media.Typeface.get_GlyphTypeface() in /_/src/Avalonia.Visuals/Media/Typeface.cs:line 69
at Avalonia.Media.TextFormatting.TextCharacters.TryGetRunProperties(ReadOnlySlice`1 text, Typeface typeface, Typeface defaultTypeface, Int32& count) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 132
at Avalonia.Media.TextFormatting.TextCharacters.CreateShapeableRun(ReadOnlySlice`1 text, TextRunProperties defaultProperties) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 62
at Avalonia.Media.TextFormatting.TextCharacters.GetShapeableCharacters() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextCharacters.cs:line 40
at Avalonia.Media.TextFormatting.TextFormatterImpl.FetchTextRuns(ITextSource textSource, Int32 firstTextSourceIndex, TextLineBreak previousLineBreak, TextLineBreak& nextLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 286
at Avalonia.Media.TextFormatting.TextFormatterImpl.FormatLine(ITextSource textSource, Int32 firstTextSourceIndex, Double paragraphWidth, TextParagraphProperties paragraphProperties, TextLineBreak previousLineBreak) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextFormatterImpl.cs:line 15
at Avalonia.Media.TextFormatting.TextLayout.UpdateLayout() in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 225
at Avalonia.Media.TextFormatting.TextLayout..ctor(String text, Typeface typeface, Double fontSize, IBrush foreground, TextAlignment textAlignment, TextWrapping textWrapping, TextTrimming textTrimming, TextDecorationCollection textDecorations, Double maxWidth, Double maxHeight, Double lineHeight, Int32 maxLines, IReadOnlyList`1 textStyleOverrides) in /_/src/Avalonia.Visuals/Media/TextFormatting/TextLayout.cs:line 71
at Avalonia.Controls.TextBlock.CreateTextLayout(Size constraint, String text) in /_/src/Avalonia.Controls/TextBlock.cs:line 475
at Avalonia.Controls.TextBlock.get_TextLayout() in /_/src/Avalonia.Controls/TextBlock.cs:line 166
at Avalonia.Controls.TextBlock.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/TextBlock.cs:line 521
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Controls.StackPanel.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/StackPanel.cs:line 207
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding, Thickness borderThickness) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 39
at Avalonia.Controls.Presenters.ContentPresenter.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Presenters/ContentPresenter.cs:line 366
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutHelper.MeasureChild(ILayoutable control, Size availableSize, Thickness padding) in /_/src/Avalonia.Layout/LayoutHelper.cs:line 46
at Avalonia.Controls.Decorator.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Decorator.cs:line 54
at Avalonia.Controls.Primitives.VisualLayerManager.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Primitives/VisualLayerManager.cs:line 133
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625
at Avalonia.Layout.Layoutable.MeasureCore(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 559
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.Layoutable.MeasureOverride(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 625
at Avalonia.Controls.Window.MeasureOverride(Size availableSize) in /_/src/Avalonia.Controls/Window.cs:line 937
at Avalonia.Controls.WindowBase.MeasureCore(Size availableSize) in /_/src/Avalonia.Controls/WindowBase.cs:line 247
at Avalonia.Layout.Layoutable.Measure(Size availableSize) in /_/src/Avalonia.Layout/Layoutable.cs:line 364
at Avalonia.Layout.LayoutManager.Measure(ILayoutable control) in /_/src/Avalonia.Layout/LayoutManager.cs:line 297
at Avalonia.Layout.LayoutManager.ExecuteInitialLayoutPass() in /_/src/Avalonia.Layout/LayoutManager.cs:line 184
at Avalonia.Controls.Window.ShowCore(Window parent) in /_/src/Avalonia.Controls/Window.cs:line 693
at Avalonia.Controls.Window.Show() in /_/src/Avalonia.Controls/Window.cs:line 631
at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.ShowMainWindow() in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 129
at Avalonia.Controls.ApplicationLifetimes.ClassicDesktopStyleApplicationLifetime.Start(String[] args) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 118
at Avalonia.ClassicDesktopStyleApplicationLifetimeExtensions.StartWithClassicDesktopLifetime[T](T builder, String[] args, ShutdownMode shutdownMode) in /_/src/Avalonia.Controls/ApplicationLifetimes/ClassicDesktopStyleApplicationLifetime.cs:line 209
at LearnAvalonia.Program.Main(String[] args) in D:\Desktop\LearnAvalonia\LearnAvalonia\Program.cs:line 13
```
# That's my way of quoting
```xaml
FontFamily="resm:LearnAvalonia.Assets.Fonts.No.216-ShangShouZhaoCaiMaoTi-2.ttf?assembly=LearnAvalonia#216-δΈι¦ζθ΄’η«δ½"
```
<img width="911" alt="image" src="https://user-images.githubusercontent.com/108921607/219945857-7588a20d-c66f-4fca-8930-9dea9e9b22ea.png">
|
process
|
an error occurs when a font with a chinese name is used csharp system invalidoperationexception could not create glyph typeface for δΈι¦ζθ΄’η«δ½ at avalonia skia fontmanagerimpl createglyphtypeface typeface typeface in src skia avalonia skia fontmanagerimpl cs line at avalonia media glyphtypeface ctor typeface typeface in src avalonia visuals media glyphtypeface cs line at avalonia media fontmanager getoraddglyphtypeface typeface typeface in src avalonia visuals media fontmanager cs line at avalonia media typeface get glyphtypeface in src avalonia visuals media typeface cs line at avalonia media textformatting textcharacters trygetrunproperties readonlyslice text typeface typeface typeface defaulttypeface count in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textcharacters createshapeablerun readonlyslice text textrunproperties defaultproperties in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textcharacters getshapeablecharacters in src avalonia visuals media textformatting textcharacters cs line at avalonia media textformatting textformatterimpl fetchtextruns itextsource textsource firsttextsourceindex textlinebreak previouslinebreak textlinebreak nextlinebreak in src avalonia visuals media textformatting textformatterimpl cs line at avalonia media textformatting textformatterimpl formatline itextsource textsource firsttextsourceindex double paragraphwidth textparagraphproperties paragraphproperties textlinebreak previouslinebreak in src avalonia visuals media textformatting textformatterimpl cs line at avalonia media textformatting textlayout updatelayout in src avalonia visuals media textformatting textlayout cs line at avalonia media textformatting textlayout ctor string text typeface typeface double fontsize ibrush foreground textalignment textalignment textwrapping textwrapping texttrimming texttrimming textdecorationcollection textdecorations double maxwidth double maxheight double lineheight maxlines ireadonlylist textstyleoverrides in src avalonia visuals media textformatting textlayout cs line at avalonia controls textblock createtextlayout size constraint string text in src avalonia controls textblock cs line at avalonia controls textblock get textlayout in src avalonia controls textblock cs line at avalonia controls textblock measureoverride size availablesize in src avalonia controls textblock cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia controls stackpanel measureoverride size availablesize in src avalonia controls stackpanel cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding in src avalonia layout layouthelper cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding thickness borderthickness in src avalonia layout layouthelper cs line at avalonia controls presenters contentpresenter measureoverride size availablesize in src avalonia controls presenters contentpresenter cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layouthelper measurechild ilayoutable control size availablesize thickness padding in src avalonia layout layouthelper cs line at avalonia controls decorator measureoverride size availablesize in src avalonia controls decorator cs line at avalonia controls primitives visuallayermanager measureoverride size availablesize in src avalonia controls primitives visuallayermanager cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measureoverride size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measurecore size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutable measureoverride size availablesize in src avalonia layout layoutable cs line at avalonia controls window measureoverride size availablesize in src avalonia controls window cs line at avalonia controls windowbase measurecore size availablesize in src avalonia controls windowbase cs line at avalonia layout layoutable measure size availablesize in src avalonia layout layoutable cs line at avalonia layout layoutmanager measure ilayoutable control in src avalonia layout layoutmanager cs line at avalonia layout layoutmanager executeinitiallayoutpass in src avalonia layout layoutmanager cs line at avalonia controls window showcore window parent in src avalonia controls window cs line at avalonia controls window show in src avalonia controls window cs line at avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime showmainwindow in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime start string args in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at avalonia classicdesktopstyleapplicationlifetimeextensions startwithclassicdesktoplifetime t builder string args shutdownmode shutdownmode in src avalonia controls applicationlifetimes classicdesktopstyleapplicationlifetime cs line at learnavalonia program main string args in d desktop learnavalonia learnavalonia program cs line that s my way of quoting xaml fontfamily resm learnavalonia assets fonts no shangshouzhaocaimaoti ttf assembly learnavalonia δΈι¦ζθ΄’η«δ½ img width alt image src
| 1
|
8,624
| 11,779,265,573
|
IssuesEvent
|
2020-03-16 17:42:01
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
PH - metric Heat Balance Operations Tab
|
Process Heating Quick Fix
|
In the process heating tool if you set the units to metric in setup the operations assessment tab still shows $/MMBtu. I think the calculations are ok, its just that the units don't change correctly.
I *think* they should say GJ. Look at the unit conversion to double check.

|
1.0
|
PH - metric Heat Balance Operations Tab - In the process heating tool if you set the units to metric in setup the operations assessment tab still shows $/MMBtu. I think the calculations are ok, its just that the units don't change correctly.
I *think* they should say GJ. Look at the unit conversion to double check.

|
process
|
ph metric heat balance operations tab in the process heating tool if you set the units to metric in setup the operations assessment tab still shows mmbtu i think the calculations are ok its just that the units don t change correctly i think they should say gj look at the unit conversion to double check
| 1
|
6,663
| 9,782,053,081
|
IssuesEvent
|
2019-06-07 21:48:04
|
googleapis/google-cloud-java
|
https://api.github.com/repos/googleapis/google-cloud-java
|
closed
|
Perform snapshot releases after successful builds to master
|
type: process
|
We lost releases of snapshot versions when we turned off the oraclejdk7 build on Travis (since the after_success.sh script only performed releases for that build variant).
Now that we are using Circle, we should add a step in the CircleCI workflow to perform a snapshot release after all jobs finish (operating systems + integration tests).
|
1.0
|
Perform snapshot releases after successful builds to master - We lost releases of snapshot versions when we turned off the oraclejdk7 build on Travis (since the after_success.sh script only performed releases for that build variant).
Now that we are using Circle, we should add a step in the CircleCI workflow to perform a snapshot release after all jobs finish (operating systems + integration tests).
|
process
|
perform snapshot releases after successful builds to master we lost releases of snapshot versions when we turned off the build on travis since the after success sh script only performed releases for that build variant now that we are using circle we should add a step in the circleci workflow to perform a snapshot release after all jobs finish operating systems integration tests
| 1
|
17,413
| 23,228,628,556
|
IssuesEvent
|
2022-08-03 04:44:38
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Processing history rework (Request in QGIS)
|
Processing 3.24
|
### Request for documentation
From pull request QGIS/qgis#46525
Author: @nyalldawson
QGIS version: 3.24
**Processing history rework**
### PR Description:
This PR partially implements the framework described in https://github.com/qgis/QGIS-Enhancement-Proposals/issues/130, and ports the processing history log over to this framework.
While there's no user visible changes which arise from this, it's a much more flexible framework as it allows us to store a full variant map in the history. So we can use this in future PRs to store more than just the python command in the history (e.g. storing the qgis_process command, the environment of the algorithm execution (ellipsoid settings, distance/area units, etc), AND any outputs given by the algorithm execution and the full output log.
History log storage in now stored in a sqlite database instead of the text file previously used, which will allow for more efficient storage/retrieval/filtering.
(This framework extends to use for more than just Processing -- it could also be used to store the history of other user operations, such as project edits, layout exports, ...)
Also now includes new right-click actions in the Processing history dialog to copy entries as JSON or qgis_process commands
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] As 'copy as json' and 'copy as qgis_process\ncommand' actions to right click menu in history dialog"
|
1.0
|
Processing history rework (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#46525
Author: @nyalldawson
QGIS version: 3.24
**Processing history rework**
### PR Description:
This PR partially implements the framework described in https://github.com/qgis/QGIS-Enhancement-Proposals/issues/130, and ports the processing history log over to this framework.
While there's no user visible changes which arise from this, it's a much more flexible framework as it allows us to store a full variant map in the history. So we can use this in future PRs to store more than just the python command in the history (e.g. storing the qgis_process command, the environment of the algorithm execution (ellipsoid settings, distance/area units, etc), AND any outputs given by the algorithm execution and the full output log.
History log storage in now stored in a sqlite database instead of the text file previously used, which will allow for more efficient storage/retrieval/filtering.
(This framework extends to use for more than just Processing -- it could also be used to store the history of other user operations, such as project edits, layout exports, ...)
Also now includes new right-click actions in the Processing history dialog to copy entries as JSON or qgis_process commands
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] As 'copy as json' and 'copy as qgis_process\ncommand' actions to right click menu in history dialog"
|
process
|
processing history rework request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version processing history rework pr description this pr partially implements the framework described in and ports the processing history log over to this framework while there s no user visible changes which arise from this it s a much more flexible framework as it allows us to store a full variant map in the history so we can use this in future prs to store more than just the python command in the history e g storing the qgis process command the environment of the algorithm execution ellipsoid settings distance area units etc and any outputs given by the algorithm execution and the full output log history log storage in now stored in a sqlite database instead of the text file previously used which will allow for more efficient storage retrieval filtering this framework extends to use for more than just processing it could also be used to store the history of other user operations such as project edits layout exports also now includes new right click actions in the processing history dialog to copy entries as json or qgis process commands sponsored by the research institute for nature and forest flemish govt commits tagged with or as copy as json and copy as qgis process ncommand actions to right click menu in history dialog
| 1
|
19,611
| 25,962,622,691
|
IssuesEvent
|
2022-12-19 01:57:49
|
AbdElAziz333/SmoothBoot-Reloaded
|
https://api.github.com/repos/AbdElAziz333/SmoothBoot-Reloaded
|
closed
|
version check data
|
in processing
|
It spams my logs with
[Forge Version Check/DEBUG] [net.minecraftforge.fml.VersionChecker/]: [smoothboot] Received version check data:
```
<!DOCTYPE html>
<html lang="en" data-color-mode="auto" data-light-theme="light" data-dark-theme="dark" data-a11y-animated-images="system">
<head>
<meta charset="utf-8">
<link rel="dns-prefetch" href="https://github.githubassets.com">
<link rel="dns-prefetch" href="https://avatars.githubusercontent.com">
<link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com">
<link rel="dns-prefetch" href="https://user-images.githubusercontent.com/">
<link rel="preconnect" href="https://github.githubassets.com" crossorigin>
<link rel="preconnect" href="https://avatars.githubusercontent.com">
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/light-719f1193e0c0.css" /><link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/dark-0c343b529849.css" /><link data-color-theme="dark_dimmed" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_dimmed-f22da508b62a.css" /><link data-color-theme="dark_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_high_contrast-188ef1de59e6.css" /><link data-color-theme="dark_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_colorblind-bc6bf4eea850.css" /><link data-color-theme="light_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_colorblind-527658dec362.css" /><link data-color-theme="light_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_high_contrast-c7a7fe0cd8ec.css" /><link data-color-theme="light_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_tritanopia-6aa855bdae0f.css" /><link data-color-theme="dark_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_tritanopia-6aa5e25aacc0.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/primer-ef7ca334db1f.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/global-86610d5db319.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/github-6d0bfca6b879.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/code-fa476fb5c04d.css" />
<meta name="optimizely-datafile" content="{"groups": [], "environmentKey": "production", "rollouts": [], "typedAudiences": [], "projectId": "16737760170", "variables": [], "featureFlags": [], "experiments": [], "version": "4", "audiences": [{"conditions": "[\"or\", {\"match\": \"exact\", \"name\": \"$opt_dummy_attribute\", \"type\": \"custom_attribute\", \"value\": \"$opt_dummy_value\"}]", "id": "$opt_dummy_audience", "name": "Optimizely-Generated Audience for Backwards Compatibility"}], "anonymizeIP": true, "sdkKey": "WTc6awnGuYDdG98CYRban", "attributes": [{"id": "16822470375", "key": "user_id"}, {"id": "17143601254", "key": "spammy"}, {"id": "18175660309", "key":
[...]```
had to truncate
|
1.0
|
version check data - It spams my logs with
[Forge Version Check/DEBUG] [net.minecraftforge.fml.VersionChecker/]: [smoothboot] Received version check data:
```
<!DOCTYPE html>
<html lang="en" data-color-mode="auto" data-light-theme="light" data-dark-theme="dark" data-a11y-animated-images="system">
<head>
<meta charset="utf-8">
<link rel="dns-prefetch" href="https://github.githubassets.com">
<link rel="dns-prefetch" href="https://avatars.githubusercontent.com">
<link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com">
<link rel="dns-prefetch" href="https://user-images.githubusercontent.com/">
<link rel="preconnect" href="https://github.githubassets.com" crossorigin>
<link rel="preconnect" href="https://avatars.githubusercontent.com">
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/light-719f1193e0c0.css" /><link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/dark-0c343b529849.css" /><link data-color-theme="dark_dimmed" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_dimmed-f22da508b62a.css" /><link data-color-theme="dark_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_high_contrast-188ef1de59e6.css" /><link data-color-theme="dark_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_colorblind-bc6bf4eea850.css" /><link data-color-theme="light_colorblind" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_colorblind-527658dec362.css" /><link data-color-theme="light_high_contrast" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_high_contrast-c7a7fe0cd8ec.css" /><link data-color-theme="light_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/light_tritanopia-6aa855bdae0f.css" /><link data-color-theme="dark_tritanopia" crossorigin="anonymous" media="all" rel="stylesheet" data-href="https://github.githubassets.com/assets/dark_tritanopia-6aa5e25aacc0.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/primer-ef7ca334db1f.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/global-86610d5db319.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/github-6d0bfca6b879.css" />
<link crossorigin="anonymous" media="all" rel="stylesheet" href="https://github.githubassets.com/assets/code-fa476fb5c04d.css" />
<meta name="optimizely-datafile" content="{"groups": [], "environmentKey": "production", "rollouts": [], "typedAudiences": [], "projectId": "16737760170", "variables": [], "featureFlags": [], "experiments": [], "version": "4", "audiences": [{"conditions": "[\"or\", {\"match\": \"exact\", \"name\": \"$opt_dummy_attribute\", \"type\": \"custom_attribute\", \"value\": \"$opt_dummy_value\"}]", "id": "$opt_dummy_audience", "name": "Optimizely-Generated Audience for Backwards Compatibility"}], "anonymizeIP": true, "sdkKey": "WTc6awnGuYDdG98CYRban", "attributes": [{"id": "16822470375", "key": "user_id"}, {"id": "17143601254", "key": "spammy"}, {"id": "18175660309", "key":
[...]```
had to truncate
|
process
|
version check data it spams my logs with received version check data link rel dns prefetch href link rel dns prefetch href link rel dns prefetch href link rel dns prefetch href link rel preconnect href meta name optimizely datafile content quot groups quot quot environmentkey quot quot production quot quot rollouts quot quot typedaudiences quot quot projectid quot quot quot quot variables quot quot featureflags quot quot experiments quot quot version quot quot quot quot audiences quot quot quot id quot quot opt dummy audience quot quot name quot quot optimizely generated audience for backwards compatibility quot quot anonymizeip quot true quot sdkkey quot quot quot quot attributes quot quot id quot quot quot quot key quot quot user id quot quot id quot quot quot quot key quot quot spammy quot quot id quot quot quot quot key quot had to truncate
| 1
|
4,659
| 7,496,694,231
|
IssuesEvent
|
2018-04-08 12:17:10
|
ODiogoSilva/assemblerflow
|
https://api.github.com/repos/ODiogoSilva/assemblerflow
|
closed
|
Module to collect reports
|
enhancement process
|
Create a terminal module that collects the `.report.json` files generated throughout the pipeline and adds the fields required for parsing by [report-nf](https://github.com/ODiogoSilva/report-nf)
|
1.0
|
Module to collect reports - Create a terminal module that collects the `.report.json` files generated throughout the pipeline and adds the fields required for parsing by [report-nf](https://github.com/ODiogoSilva/report-nf)
|
process
|
module to collect reports create a terminal module that collects the report json files generated throughout the pipeline and adds the fields required for parsing by
| 1
|
12,860
| 15,252,209,726
|
IssuesEvent
|
2021-02-20 01:58:15
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Support drag and drop in a processing model
|
Feature Request Feedback Processing stale
|
Author Name: **Magnus Nilsson** (Magnus Nilsson)
Original Redmine Issue: [21009](https://issues.qgis.org/issues/21009)
Redmine category:processing/modeller
---
I would like to see support for drag and drop in Processing models. When I drop a file (CSV, GeoPackage etc.) from my Windows deskop on my model, QGIS should be able to accept it as a data source.
|
1.0
|
Support drag and drop in a processing model - Author Name: **Magnus Nilsson** (Magnus Nilsson)
Original Redmine Issue: [21009](https://issues.qgis.org/issues/21009)
Redmine category:processing/modeller
---
I would like to see support for drag and drop in Processing models. When I drop a file (CSV, GeoPackage etc.) from my Windows deskop on my model, QGIS should be able to accept it as a data source.
|
process
|
support drag and drop in a processing model author name magnus nilsson magnus nilsson original redmine issue redmine category processing modeller i would like to see support for drag and drop in processing models when i drop a file csv geopackage etc from my windows deskop on my model qgis should be able to accept it as a data source
| 1
|
465,300
| 13,382,298,460
|
IssuesEvent
|
2020-09-02 08:36:30
|
oppia/oppia-android
|
https://api.github.com/repos/oppia/oppia-android
|
opened
|
Trying to input answers in landscape is impossible due to keyboard automatically dismissing.
|
Priority: Essential Status: In implementation Status: Not started Type: Bug Where: Exploration player Where: Skills/Questions
|
**Describe the bug**
When you try to input anything in landscape, the keyboard automatically dismisses.
**To Reproduce**
Steps to reproduce the behavior:
1. Open the app
2. Navigate to the prototype exploration
3. Rotate to landscape
4. Continue to the second state (fractions)
5. Select the fraction input box
6. Observe the soft keyboard open
7. Observe the soft keyboard immediately close
**Expected behavior**
The soft keyboard shouldn't be automatically closing.
**Screenshots**
-
**Device**
- Device: Pixel XL
- SDK version: SDK 29
**Additional context**
This is definitely a regression since StateFragmentTest was passing 2 weeks ago (the last time I tried these scenarios).
|
1.0
|
Trying to input answers in landscape is impossible due to keyboard automatically dismissing. - **Describe the bug**
When you try to input anything in landscape, the keyboard automatically dismisses.
**To Reproduce**
Steps to reproduce the behavior:
1. Open the app
2. Navigate to the prototype exploration
3. Rotate to landscape
4. Continue to the second state (fractions)
5. Select the fraction input box
6. Observe the soft keyboard open
7. Observe the soft keyboard immediately close
**Expected behavior**
The soft keyboard shouldn't be automatically closing.
**Screenshots**
-
**Device**
- Device: Pixel XL
- SDK version: SDK 29
**Additional context**
This is definitely a regression since StateFragmentTest was passing 2 weeks ago (the last time I tried these scenarios).
|
non_process
|
trying to input answers in landscape is impossible due to keyboard automatically dismissing describe the bug when you try to input anything in landscape the keyboard automatically dismisses to reproduce steps to reproduce the behavior open the app navigate to the prototype exploration rotate to landscape continue to the second state fractions select the fraction input box observe the soft keyboard open observe the soft keyboard immediately close expected behavior the soft keyboard shouldn t be automatically closing screenshots device device pixel xl sdk version sdk additional context this is definitely a regression since statefragmenttest was passing weeks ago the last time i tried these scenarios
| 0
|
72,855
| 31,769,565,512
|
IssuesEvent
|
2023-09-12 10:53:11
|
gauravrs18/issue_onboarding
|
https://api.github.com/repos/gauravrs18/issue_onboarding
|
closed
|
dev-angular-style-account-services-new-connection-component-activate-component
-consumer-details-component
-application-component
-address-component
|
CX-account-services
|
dev-angular-style-account-services-new-connection-component-activate-component
-consumer-details-component
-application-component
-address-component
|
1.0
|
dev-angular-style-account-services-new-connection-component-activate-component
-consumer-details-component
-application-component
-address-component - dev-angular-style-account-services-new-connection-component-activate-component
-consumer-details-component
-application-component
-address-component
|
non_process
|
dev angular style account services new connection component activate component consumer details component application component address component dev angular style account services new connection component activate component consumer details component application component address component
| 0
|
55,086
| 13,521,887,011
|
IssuesEvent
|
2020-09-15 07:44:38
|
spring-projects/spring-session
|
https://api.github.com/repos/spring-projects/spring-session
|
closed
|
spring-session-bom `2020.0.0-SNAPSHOT` refers to milestones
|
in: build type: dependency-upgrade
|
The latest [spring-session-bom](https://repo.spring.io/snapshot/org/springframework/session/spring-session-bom/2020.0.0-SNAPSHOT/spring-session-bom-2020.0.0-20200915.060532-123.pom) on repo.spring.io has references to milestone 1 artifacts.
I believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot.
|
1.0
|
spring-session-bom `2020.0.0-SNAPSHOT` refers to milestones - The latest [spring-session-bom](https://repo.spring.io/snapshot/org/springframework/session/spring-session-bom/2020.0.0-SNAPSHOT/spring-session-bom-2020.0.0-20200915.060532-123.pom) on repo.spring.io has references to milestone 1 artifacts.
I believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot.
|
non_process
|
spring session bom snapshot refers to milestones the latest on repo spring io has references to milestone artifacts i believe this should be snapshots so that integration tests can be applied on the latest state when using the snapshot
| 0
|
298,632
| 22,540,886,050
|
IssuesEvent
|
2022-06-26 00:22:46
|
apache/camel-k
|
https://api.github.com/repos/apache/camel-k
|
closed
|
In troubleshooting documentation section, provide information on which logs to look at
|
area/documentation status/stale
|
In troubleshooting documentation section provide information on which log to look at
operator log/kit log, others?
using kubectl/VS Code?
|
1.0
|
In troubleshooting documentation section, provide information on which logs to look at - In troubleshooting documentation section provide information on which log to look at
operator log/kit log, others?
using kubectl/VS Code?
|
non_process
|
in troubleshooting documentation section provide information on which logs to look at in troubleshooting documentation section provide information on which log to look at operator log kit log others using kubectl vs code
| 0
|
4,614
| 7,459,552,181
|
IssuesEvent
|
2018-03-30 15:47:58
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
bigquery parquet loading code not available on pypi, please push dev release
|
api: bigquery type: process
|
I'm evaluating the BigQuery new support for loading from parquet files and following the code samples from:
https://cloud.google.com/bigquery/docs/loading-data-cloud-storage-parquet
The above uses code that is not yet on pypi. I can build it locally but for rolling out our real tests it would be convenient if a release could be made. It doesn't have to be an offical release, just pushing the dev release would be enough: https://github.com/GoogleCloudPlatform/google-cloud-python/commit/fc91c877eb6aee2a8edfd5a8950bc5ca51e8255c
/cc @tswast
|
1.0
|
bigquery parquet loading code not available on pypi, please push dev release - I'm evaluating the BigQuery new support for loading from parquet files and following the code samples from:
https://cloud.google.com/bigquery/docs/loading-data-cloud-storage-parquet
The above uses code that is not yet on pypi. I can build it locally but for rolling out our real tests it would be convenient if a release could be made. It doesn't have to be an offical release, just pushing the dev release would be enough: https://github.com/GoogleCloudPlatform/google-cloud-python/commit/fc91c877eb6aee2a8edfd5a8950bc5ca51e8255c
/cc @tswast
|
process
|
bigquery parquet loading code not available on pypi please push dev release i m evaluating the bigquery new support for loading from parquet files and following the code samples from the above uses code that is not yet on pypi i can build it locally but for rolling out our real tests it would be convenient if a release could be made it doesn t have to be an offical release just pushing the dev release would be enough cc tswast
| 1
|
13,715
| 16,480,396,106
|
IssuesEvent
|
2021-05-24 10:50:33
|
New-Time-Development/OmeCord
|
https://api.github.com/repos/New-Time-Development/OmeCord
|
closed
|
Chat system does work not correct
|
Big issue Known bug bug work in process
|
I fix it today. Everybody can write in the chat. Thats bad
|
1.0
|
Chat system does work not correct - I fix it today. Everybody can write in the chat. Thats bad
|
process
|
chat system does work not correct i fix it today everybody can write in the chat thats bad
| 1
|
3,282
| 6,369,676,148
|
IssuesEvent
|
2017-08-01 12:32:05
|
gatewayapps/ims-feedback
|
https://api.github.com/repos/gatewayapps/ims-feedback
|
closed
|
It shouldn't be scoring the activities, it should be scoring the aspects.??
|
bug ims.packages.process
|
Version: 1.1.1-master.53(9e63209)
User: CHRISSIE GRAY(chrissie.gray@nucor-yamato.com)
|
1.0
|
It shouldn't be scoring the activities, it should be scoring the aspects.?? - Version: 1.1.1-master.53(9e63209)
User: CHRISSIE GRAY(chrissie.gray@nucor-yamato.com)
|
process
|
it shouldn t be scoring the activities it should be scoring the aspects version master user chrissie gray chrissie gray nucor yamato com
| 1
|
29,625
| 11,764,607,048
|
IssuesEvent
|
2020-03-14 13:31:45
|
Molizo/FTC-Scouting-App-Skystone
|
https://api.github.com/repos/Molizo/FTC-Scouting-App-Skystone
|
opened
|
CVE-2019-0820 (High) detected in msbump.2.3.2.nupkg
|
security vulnerability
|
## CVE-2019-0820 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>msbump.2.3.2.nupkg</b></p></summary>
<p>MSBuild task that bumps the version of a Visual Studio 2017 project before build and pack.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/msbump.2.3.2.nupkg">https://api.nuget.org/packages/msbump.2.3.2.nupkg</a></p>
<p>Path to dependency file: /tmp/ws-scm/FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p>
<p>Path to vulnerable library: /FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p>
<p>
Dependency Hierarchy:
- :x: **msbump.2.3.2.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Molizo/FTC-Scouting-App-Skystone/commit/7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f">7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.
<p>Publish Date: 2019-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820</a></p>
<p>Release Date: 2019-05-16</p>
<p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-0820 (High) detected in msbump.2.3.2.nupkg - ## CVE-2019-0820 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>msbump.2.3.2.nupkg</b></p></summary>
<p>MSBuild task that bumps the version of a Visual Studio 2017 project before build and pack.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/msbump.2.3.2.nupkg">https://api.nuget.org/packages/msbump.2.3.2.nupkg</a></p>
<p>Path to dependency file: /tmp/ws-scm/FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p>
<p>Path to vulnerable library: /FTC-Scouting-App-Skystone/SkystoneScouting/SkystoneScouting.csproj</p>
<p>
Dependency Hierarchy:
- :x: **msbump.2.3.2.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Molizo/FTC-Scouting-App-Skystone/commit/7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f">7f8d4baf9caf6174e8aa37b79cb92b7ba6d5658f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.
<p>Publish Date: 2019-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0820</a></p>
<p>Release Date: 2019-05-16</p>
<p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in msbump nupkg cve high severity vulnerability vulnerable library msbump nupkg msbuild task that bumps the version of a visual studio project before build and pack library home page a href path to dependency file tmp ws scm ftc scouting app skystone skystonescouting skystonescouting csproj path to vulnerable library ftc scouting app skystone skystonescouting skystonescouting csproj dependency hierarchy x msbump nupkg vulnerable library found in head commit a href vulnerability details a denial of service vulnerability exists when net framework and net core improperly process regex strings aka net framework and net core denial of service vulnerability this cve id is unique from cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text regularexpressions step up your open source security game with whitesource
| 0
|
18,461
| 24,549,627,108
|
IssuesEvent
|
2022-10-12 11:34:00
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Admins > Add new admin > UI issue in add new admin screen
|
Bug P2 Participant manager Process: Fixed Process: Tested dev
|
**Steps:**
1. Login to PM
2. Click on 'Admins' tab
3. Click on 'Add new admin' button
4. Verify the email field
**AR:** Dropdown icon is getting displayed in the email field
**ER:** Dropdown icon should not get displayed in the email field

|
2.0
|
[PM] Admins > Add new admin > UI issue in add new admin screen - **Steps:**
1. Login to PM
2. Click on 'Admins' tab
3. Click on 'Add new admin' button
4. Verify the email field
**AR:** Dropdown icon is getting displayed in the email field
**ER:** Dropdown icon should not get displayed in the email field

|
process
|
admins add new admin ui issue in add new admin screen steps login to pm click on admins tab click on add new admin button verify the email field ar dropdown icon is getting displayed in the email field er dropdown icon should not get displayed in the email field
| 1
|
300,114
| 25,945,916,576
|
IssuesEvent
|
2022-12-17 00:57:54
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Retain LGTM through squashes
|
sig/contributor-experience kind/feature help wanted sig/testing lifecycle/frozen triage/accepted
|
#### What would you like to be added:
Configure this repo ([k/kubernetes](https://github.com/kubernetes/kubernetes)) to retain LGTM labels when people squash commits. PR https://github.com/kubernetes/test-infra/pull/9138 made this possible.
#### Why is this needed:
This feature will make it easier for people to squash commits, and should give reviewers confidence to ask other contributors to squash when appropriate.
To implement this change, make a PR against ([k/test-infra](https://github.com/kubernetes/test-infra)) that is similar to https://github.com/kubernetes/test-infra/pull/19304.
|
1.0
|
Retain LGTM through squashes - #### What would you like to be added:
Configure this repo ([k/kubernetes](https://github.com/kubernetes/kubernetes)) to retain LGTM labels when people squash commits. PR https://github.com/kubernetes/test-infra/pull/9138 made this possible.
#### Why is this needed:
This feature will make it easier for people to squash commits, and should give reviewers confidence to ask other contributors to squash when appropriate.
To implement this change, make a PR against ([k/test-infra](https://github.com/kubernetes/test-infra)) that is similar to https://github.com/kubernetes/test-infra/pull/19304.
|
non_process
|
retain lgtm through squashes what would you like to be added configure this repo to retain lgtm labels when people squash commits pr made this possible why is this needed this feature will make it easier for people to squash commits and should give reviewers confidence to ask other contributors to squash when appropriate to implement this change make a pr against that is similar to
| 0
|
150,677
| 11,980,556,182
|
IssuesEvent
|
2020-04-07 09:32:24
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
Can't find web3 in browser (on simpledapp.eth) when refreshing the page
|
bug e2e test blocker fix them all webview
|
# Bug Report
## Problem
When refreshing browser page with simpledapp.eth then it's often "..fails to find web3 library"
While I could not reproduce it with usual opening Dapp in browser, it's reproduced 2 of 3 times when refreshing the page
#### Expected behavior
Refreshing the page and finds web3 fine
#### Actual behavior
<img width="303" alt="Screenshot 2020-03-26 at 12 34 46" src="https://user-images.githubusercontent.com/8749671/77637368-3bda6980-6f5e-11ea-8105-272544503246.png">
### Reproduction
- Open Status and create new account
- Open `simpledapp.eth` in browser
- Tap `Refresh` button (top-right)
### Additional Information
- Status version: develop nightly 26th of March
- Operating System: Android
#### Logs
```
03-26 12:28:37.445 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 0
03-26 12:28:37.542 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 1
03-26 12:28:37.719 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in
03-26 12:28:37.937 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/loading-started
03-26 12:28:37.939 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in
03-26 12:28:38.456 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/navigation-state-changed
03-26 12:28:42.128 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :mailserver/check-connection-timeout
```
|
1.0
|
Can't find web3 in browser (on simpledapp.eth) when refreshing the page - # Bug Report
## Problem
When refreshing browser page with simpledapp.eth then it's often "..fails to find web3 library"
While I could not reproduce it with usual opening Dapp in browser, it's reproduced 2 of 3 times when refreshing the page
#### Expected behavior
Refreshing the page and finds web3 fine
#### Actual behavior
<img width="303" alt="Screenshot 2020-03-26 at 12 34 46" src="https://user-images.githubusercontent.com/8749671/77637368-3bda6980-6f5e-11ea-8105-272544503246.png">
### Reproduction
- Open Status and create new account
- Open `simpledapp.eth` in browser
- Tap `Refresh` button (top-right)
### Additional Information
- Status version: develop nightly 26th of March
- Operating System: Android
#### Logs
```
03-26 12:28:37.445 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 0
03-26 12:28:37.542 1543 1543 D ViewRootImpl@e3f57af[MainActivity]: ViewPostIme pointer 1
03-26 12:28:37.719 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in
03-26 12:28:37.937 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/loading-started
03-26 12:28:37.939 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :set-in
03-26 12:28:38.456 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :browser/navigation-state-changed
03-26 12:28:42.128 1543 1583 D ReactNativeJS: DEBUG [status-im.utils.handlers:32] - Handling re-frame event: :mailserver/check-connection-timeout
```
|
non_process
|
can t find in browser on simpledapp eth when refreshing the page bug report problem when refreshing browser page with simpledapp eth then it s often fails to find library while i could not reproduce it with usual opening dapp in browser it s reproduced of times when refreshing the page expected behavior refreshing the page and finds fine actual behavior img width alt screenshot at src reproduction open status and create new account open simpledapp eth in browser tap refresh button top right additional information status version develop nightly of march operating system android logs d viewrootimpl viewpostime pointer d viewrootimpl viewpostime pointer d reactnativejs debug handling re frame event set in d reactnativejs debug handling re frame event browser loading started d reactnativejs debug handling re frame event set in d reactnativejs debug handling re frame event browser navigation state changed d reactnativejs debug handling re frame event mailserver check connection timeout
| 0
|
1,534
| 4,120,220,867
|
IssuesEvent
|
2016-06-08 17:12:25
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
execFile arguments appear to be in the wrong order
|
child_process
|
* **Version**: 6.0.0
* **Platform**: MacOS
Reading the docs for execFile, I expected the invocation with environment variables to look something like:
`var ls = childProcess.execFile(path, args, {env: {SOME_ENV_VARIABLE: "somethign"}});`
However, this signature did not work. Instead, we found both:
`var ls = childProcess.execFile(path, args, function() {}, {env: {SOME_ENV_VARIABLE: "somethign"}});`
and
`var ls = childProcess.execFile(path, args, [], {env: {SOME_ENV_VARIABLE: "somethign"}});`
worked. Am I misreading the docs, or is the behavior misdocumented?
|
1.0
|
execFile arguments appear to be in the wrong order - * **Version**: 6.0.0
* **Platform**: MacOS
Reading the docs for execFile, I expected the invocation with environment variables to look something like:
`var ls = childProcess.execFile(path, args, {env: {SOME_ENV_VARIABLE: "somethign"}});`
However, this signature did not work. Instead, we found both:
`var ls = childProcess.execFile(path, args, function() {}, {env: {SOME_ENV_VARIABLE: "somethign"}});`
and
`var ls = childProcess.execFile(path, args, [], {env: {SOME_ENV_VARIABLE: "somethign"}});`
worked. Am I misreading the docs, or is the behavior misdocumented?
|
process
|
execfile arguments appear to be in the wrong order version platform macos reading the docs for execfile i expected the invocation with environment variables to look something like var ls childprocess execfile path args env some env variable somethign however this signature did not work instead we found both var ls childprocess execfile path args function env some env variable somethign and var ls childprocess execfile path args env some env variable somethign worked am i misreading the docs or is the behavior misdocumented
| 1
|
19,754
| 26,123,717,478
|
IssuesEvent
|
2022-12-28 15:39:29
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_fs (__main__.TestMultiprocessing)
|
module: multiprocessing triaged module: flaky-tests skipped
|
Platforms: asan, linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_fs%2C%20TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6297446390).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green.
cc @VitalyFedyunin
|
1.0
|
DISABLED test_fs (__main__.TestMultiprocessing) - Platforms: asan, linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_fs%2C%20TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6297446390).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green.
cc @VitalyFedyunin
|
process
|
disabled test fs main testmultiprocessing platforms asan linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with red and green cc vitalyfedyunin
| 1
|
604,349
| 18,682,315,818
|
IssuesEvent
|
2021-11-01 07:54:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
enewspaper.latimes.com - site is not usable
|
browser-firefox priority-normal engine-gecko
|
<!-- @browser: Firefox 93.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/91731 -->
**URL**: https://enewspaper.latimes.com/desktop/latimes/default.aspx?token=42e23962a5d74614be16bae3d62d13e7&utm_id=39150&sfmc_id=3035949&edid=95390162-66b6-4b40-9414-680b46c0eb7a
**Browser / Version**: Firefox 93.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The site used to load with Firefox with no problem, now it does not but does if I switch to Chrome.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/5eb1c0fb-f144-462f-906c-aca31fbba112.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with β€οΈ_
|
1.0
|
enewspaper.latimes.com - site is not usable - <!-- @browser: Firefox 93.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/91731 -->
**URL**: https://enewspaper.latimes.com/desktop/latimes/default.aspx?token=42e23962a5d74614be16bae3d62d13e7&utm_id=39150&sfmc_id=3035949&edid=95390162-66b6-4b40-9414-680b46c0eb7a
**Browser / Version**: Firefox 93.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The site used to load with Firefox with no problem, now it does not but does if I switch to Chrome.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/5eb1c0fb-f144-462f-906c-aca31fbba112.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with β€οΈ_
|
non_process
|
enewspaper latimes com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce the site used to load with firefox with no problem now it does not but does if i switch to chrome view the screenshot img alt screenshot src browser configuration none from with β€οΈ
| 0
|
9,940
| 12,974,088,184
|
IssuesEvent
|
2020-07-21 14:56:43
|
kubeflow/kubeflow
|
https://api.github.com/repos/kubeflow/kubeflow
|
opened
|
Update OWNERs of central dashboard
|
area/community area/front-end kind/process priority/p1
|
/kind process
The OWNERs file of the centraldashboard was last updated over a year ago.
https://github.com/kubeflow/kubeflow/blob/master/components/centraldashboard/OWNERS
It looks like at least @prodonjs has indicated [here](https://github.com/kubeflow/kubeflow/issues/3576#issuecomment-660055405) that he is no longer actively working on this.
I think @avdaredevil is also primarily focused on other projects but ocassionally pitches in.
@swiftdiaries are you still actively working on centraldashboard or willing to?
It looks like we may need to find an appropriate set of owners in order to continue to drive this forward.
|
1.0
|
Update OWNERs of central dashboard - /kind process
The OWNERs file of the centraldashboard was last updated over a year ago.
https://github.com/kubeflow/kubeflow/blob/master/components/centraldashboard/OWNERS
It looks like at least @prodonjs has indicated [here](https://github.com/kubeflow/kubeflow/issues/3576#issuecomment-660055405) that he is no longer actively working on this.
I think @avdaredevil is also primarily focused on other projects but ocassionally pitches in.
@swiftdiaries are you still actively working on centraldashboard or willing to?
It looks like we may need to find an appropriate set of owners in order to continue to drive this forward.
|
process
|
update owners of central dashboard kind process the owners file of the centraldashboard was last updated over a year ago it looks like at least prodonjs has indicated that he is no longer actively working on this i think avdaredevil is also primarily focused on other projects but ocassionally pitches in swiftdiaries are you still actively working on centraldashboard or willing to it looks like we may need to find an appropriate set of owners in order to continue to drive this forward
| 1
|
518,376
| 15,027,646,564
|
IssuesEvent
|
2021-02-02 01:15:42
|
Conteo-Inc/conteo
|
https://api.github.com/repos/Conteo-Inc/conteo
|
closed
|
FR-M2: Create Match List
|
0_Frontend 0_Priority 1
|
Communicate with @sb3958 about the design requirements.
While the list itself does not need to be a separate component, the list items need at least one separate component.
Do not include the request, fill with dummy data until <Linked Issue> is completed.
|
1.0
|
FR-M2: Create Match List - Communicate with @sb3958 about the design requirements.
While the list itself does not need to be a separate component, the list items need at least one separate component.
Do not include the request, fill with dummy data until <Linked Issue> is completed.
|
non_process
|
fr create match list communicate with about the design requirements while the list itself does not need to be a separate component the list items need at least one separate component do not include the request fill with dummy data until is completed
| 0
|
161,896
| 25,415,093,448
|
IssuesEvent
|
2022-11-22 22:56:38
|
WordPress/pattern-directory
|
https://api.github.com/repos/WordPress/pattern-directory
|
opened
|
Create a new "featured" view
|
[Component] Theme [Status] Needs Design
|
As mentioned in https://github.com/WordPress/gutenberg/issues/44501, there should be a "featured" tag for the pattern directory. Currently we have a "featured" category, which could work for this, or we could take the opportunity to switch this over to an internal keyword (since it's not _really_ meant to be used in the same way as the other categories), as you should be able to filter featured by category.
- [ ] Create a new keyword term "featured", and apply the new term to patterns by wordpressdotorg
- [ ] Add a βfeaturedβ toggle to the filter navigation (needs design)
- [ ] Support showing only featured posts in categories, homepage, etc
Tech noteβ currently, we've done a lot of permalink rewriting to allow complex filtering of patterns with "pretty" urls. Adding in this filter will make it even more complex, so I would stick with a query parameter instead of trying to make the URLs pretty (ex, leave it as `/patterns/categories/header/?featured=true` rather than trying to inject `/featured/` in there somewhere).
@jasmussen I've added this issue separately as a "needs design", but you can combine it with the design for #395 if that's easier. There are two other design/experience questions here though: Should the front page show all patterns or only featured by default? Is there a use case for showing just βcommunityβ, or should it be βAll | Featuredβ?
|
1.0
|
Create a new "featured" view - As mentioned in https://github.com/WordPress/gutenberg/issues/44501, there should be a "featured" tag for the pattern directory. Currently we have a "featured" category, which could work for this, or we could take the opportunity to switch this over to an internal keyword (since it's not _really_ meant to be used in the same way as the other categories), as you should be able to filter featured by category.
- [ ] Create a new keyword term "featured", and apply the new term to patterns by wordpressdotorg
- [ ] Add a βfeaturedβ toggle to the filter navigation (needs design)
- [ ] Support showing only featured posts in categories, homepage, etc
Tech noteβ currently, we've done a lot of permalink rewriting to allow complex filtering of patterns with "pretty" urls. Adding in this filter will make it even more complex, so I would stick with a query parameter instead of trying to make the URLs pretty (ex, leave it as `/patterns/categories/header/?featured=true` rather than trying to inject `/featured/` in there somewhere).
@jasmussen I've added this issue separately as a "needs design", but you can combine it with the design for #395 if that's easier. There are two other design/experience questions here though: Should the front page show all patterns or only featured by default? Is there a use case for showing just βcommunityβ, or should it be βAll | Featuredβ?
|
non_process
|
create a new featured view as mentioned in there should be a featured tag for the pattern directory currently we have a featured category which could work for this or we could take the opportunity to switch this over to an internal keyword since it s not really meant to be used in the same way as the other categories as you should be able to filter featured by category create a new keyword term featured and apply the new term to patterns by wordpressdotorg add a βfeaturedβ toggle to the filter navigation needs design support showing only featured posts in categories homepage etc tech noteβ currently we ve done a lot of permalink rewriting to allow complex filtering of patterns with pretty urls adding in this filter will make it even more complex so i would stick with a query parameter instead of trying to make the urls pretty ex leave it as patterns categories header featured true rather than trying to inject featured in there somewhere jasmussen i ve added this issue separately as a needs design but you can combine it with the design for if that s easier there are two other design experience questions here though should the front page show all patterns or only featured by default is there a use case for showing just βcommunityβ or should it be βall featuredβ
| 0
|
4,166
| 7,107,918,947
|
IssuesEvent
|
2018-01-16 21:45:53
|
18F/product-guide
|
https://api.github.com/repos/18F/product-guide
|
closed
|
UPDATE SECTION (Project Comms) - Dashboard
|
process change
|
Additional guidance to come from @gboone re: how PMs should use the dashboard per comment in googledoc -
@awfrancisco: Any additional guidance you want to add for how PMs should use the Dashboard?
@gboone: I am sure we will eventually but since we're up in the air on redesigning it ATM I'll hold off and update this when we have a real answer.
|
1.0
|
UPDATE SECTION (Project Comms) - Dashboard - Additional guidance to come from @gboone re: how PMs should use the dashboard per comment in googledoc -
@awfrancisco: Any additional guidance you want to add for how PMs should use the Dashboard?
@gboone: I am sure we will eventually but since we're up in the air on redesigning it ATM I'll hold off and update this when we have a real answer.
|
process
|
update section project comms dashboard additional guidance to come from gboone re how pms should use the dashboard per comment in googledoc awfrancisco any additional guidance you want to add for how pms should use the dashboard gboone i am sure we will eventually but since we re up in the air on redesigning it atm i ll hold off and update this when we have a real answer
| 1
|
14,895
| 18,291,056,115
|
IssuesEvent
|
2021-10-05 15:17:28
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Change Public Access Prevention "unspecified" to "inherited"
|
api: storage priority: p2 type: process
|
GCS is changing the `unspecified` public access prevention value to `inherited`.
This will require the following changes:
1. Modify the values in the library. Constants/enum values for unspecified: these should be marked as deprecated. They can stay in the library if deletion would constitute a breaking change
2. Update the system tests.
3. Create a new sample with region tag public_access_prevention_inherited. So we don't break any documentation links, please do not delete the unspecified sample or sample test from your repo at this point in time.
Please see nodejs example here: https://github.com/googleapis/nodejs-storage/pull/1614/files
Please wait for approval from myself or @tritone before merging or releasing this change so we can coordinate around release dates. New inherited value should be rolled out to allowlisted testing projects this coming week; should be at 100% by 9/30
|
1.0
|
Change Public Access Prevention "unspecified" to "inherited" - GCS is changing the `unspecified` public access prevention value to `inherited`.
This will require the following changes:
1. Modify the values in the library. Constants/enum values for unspecified: these should be marked as deprecated. They can stay in the library if deletion would constitute a breaking change
2. Update the system tests.
3. Create a new sample with region tag public_access_prevention_inherited. So we don't break any documentation links, please do not delete the unspecified sample or sample test from your repo at this point in time.
Please see nodejs example here: https://github.com/googleapis/nodejs-storage/pull/1614/files
Please wait for approval from myself or @tritone before merging or releasing this change so we can coordinate around release dates. New inherited value should be rolled out to allowlisted testing projects this coming week; should be at 100% by 9/30
|
process
|
change public access prevention unspecified to inherited gcs is changing the unspecified public access prevention value to inherited this will require the following changes modify the values in the library constants enum values for unspecified these should be marked as deprecated they can stay in the library if deletion would constitute a breaking change update the system tests create a new sample with region tag public access prevention inherited so we don t break any documentation links please do not delete the unspecified sample or sample test from your repo at this point in time please see nodejs example here please wait for approval from myself or tritone before merging or releasing this change so we can coordinate around release dates new inherited value should be rolled out to allowlisted testing projects this coming week should be at by
| 1
|
21,024
| 27,969,912,736
|
IssuesEvent
|
2023-03-25 00:17:32
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
rotated, mirrored cloning
|
feature: enhancement scope: image processing no-issue-activity
|
**Is your feature request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is, e.g. "I'm always frustrated when [...]" -->
Cloning is not easy if we have only a limited amount of sources to pick from, and a bigger area to cover / repair
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
my idea is, to manipulate the output of the cloned source to be more flexible:
- mirroring the source input along the x/y axis,
- rotating of the source input
In this way, we avoid to see "repetitive patterns" when we have to pick multiple times from the same source
The only problem which i see is, light and shadows may influence the cloned object, and the cloned object might look fake. But, if our cloned output needs only a slight rotation, the end result should look acceptable, imo.
**Alternatives**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
|
1.0
|
rotated, mirrored cloning - **Is your feature request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is, e.g. "I'm always frustrated when [...]" -->
Cloning is not easy if we have only a limited amount of sources to pick from, and a bigger area to cover / repair
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
my idea is, to manipulate the output of the cloned source to be more flexible:
- mirroring the source input along the x/y axis,
- rotating of the source input
In this way, we avoid to see "repetitive patterns" when we have to pick multiple times from the same source
The only problem which i see is, light and shadows may influence the cloned object, and the cloned object might look fake. But, if our cloned output needs only a slight rotation, the end result should look acceptable, imo.
**Alternatives**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
|
process
|
rotated mirrored cloning is your feature request related to a problem please describe cloning is not easy if we have only a limited amount of sources to pick from and a bigger area to cover repair describe the solution you d like my idea is to manipulate the output of the cloned source to be more flexible mirroring the source input along the x y axis rotating of the source input in this way we avoid to see repetitive patterns when we have to pick multiple times from the same source the only problem which i see is light and shadows may influence the cloned object and the cloned object might look fake but if our cloned output needs only a slight rotation the end result should look acceptable imo alternatives
| 1
|
493,418
| 14,231,703,352
|
IssuesEvent
|
2020-11-18 09:55:53
|
kubermatic/kubermatic
|
https://api.github.com/repos/kubermatic/kubermatic
|
closed
|
Make pod and service networks configurable
|
customer-request lifecycle/rotten priority/normal team/lifecycle
|
**User Story**
The pod and service network cidr is part of the cluster CRD. Since the networks cannot be change for existing clusters, there is not effective way to configure the networks. Customers requested to use a different cidr as problems are expected when the pod / service networks overlap with other infrastructure.
**Acceptance criteria**
In best case, the pod and service cidr should be configurable per cluster. The option to change the defaults globally would also help.
|
1.0
|
Make pod and service networks configurable - **User Story**
The pod and service network cidr is part of the cluster CRD. Since the networks cannot be change for existing clusters, there is not effective way to configure the networks. Customers requested to use a different cidr as problems are expected when the pod / service networks overlap with other infrastructure.
**Acceptance criteria**
In best case, the pod and service cidr should be configurable per cluster. The option to change the defaults globally would also help.
|
non_process
|
make pod and service networks configurable user story the pod and service network cidr is part of the cluster crd since the networks cannot be change for existing clusters there is not effective way to configure the networks customers requested to use a different cidr as problems are expected when the pod service networks overlap with other infrastructure acceptance criteria in best case the pod and service cidr should be configurable per cluster the option to change the defaults globally would also help
| 0
|
96,878
| 28,038,744,115
|
IssuesEvent
|
2023-03-28 16:50:41
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
reopened
|
evdi broken with linux 6.0
|
6.topic: kernel 6.topic: hardware 0.kind: build failure
|
### Steps To Reproduce
```nix
boot.kernelPackages = pkgs.linuxPackages_6_0;
services.xserver.videoDrivers = [ "modesetting" "displaylink" ];
```
### Build log
```
error: builder for '/nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv' failed with exit code 2;
last 10 log lines:
> 1060 | drm_framebuffer_get(&efb->base);
> | ^~~~~~~~~~~~~~~~~~~
> make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_fb.o] Error 1
> cc1: some warnings being treated as errors
> make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_painter.o] Error 1
> make[2]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/Makefile:1852: /build/source/module] Error 2
> make[2]: Leaving directory '/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/build'
> make[1]: *** [Makefile:81: module] Error 2
> make[1]: Leaving directory '/build/source/module'
> make: *** [Makefile:8: all] Error 2
For full logs, run 'nix log /nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv'.
error: 1 dependencies of derivation '/nix/store/49lczfqfpp17y4vmpr04pkss0cwrb4pf-kernel-modules.drv' failed to build
error: 1 dependencies of derivation '/nix/store/q6p42l3r50xv4jphgpd9bx81y32r7m6y-linux-6.0.7-modules.drv' failed to build
error: 1 dependencies of derivation '/nix/store/pw5f02p73pvpj03a60pm0iysl88csdmn-nixos-system-Ryan-Laptop-22.11.20221108.0932685.drv' failed to build
error: 1 dependencies of derivation '/nix/store/zs7ma4qvymdrb999cqk1hyqpziadb0yc-nixops-machines.drv' failed to build
```
### Additional context
### Notify maintainers
@eyjhb
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
[user@system:~]$ nix-shell -p nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 5.19.16, NixOS, 22.11 (Raccoon), 22.11.20221031.d40fea9`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.0`
- channels(root): `"nixos"`
- nixpkgs: `/nix/store/mrpx61vyrmhqhr2ms9svscyc1h523gf1-source`
```
|
1.0
|
evdi broken with linux 6.0 - ### Steps To Reproduce
```nix
boot.kernelPackages = pkgs.linuxPackages_6_0;
services.xserver.videoDrivers = [ "modesetting" "displaylink" ];
```
### Build log
```
error: builder for '/nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv' failed with exit code 2;
last 10 log lines:
> 1060 | drm_framebuffer_get(&efb->base);
> | ^~~~~~~~~~~~~~~~~~~
> make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_fb.o] Error 1
> cc1: some warnings being treated as errors
> make[3]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/scripts/Makefile.build:249: /build/source/module/evdi_painter.o] Error 1
> make[2]: *** [/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/source/Makefile:1852: /build/source/module] Error 2
> make[2]: Leaving directory '/nix/store/5vjwh5n4a6ajcdg5sbmx9xpijzk0lvn4-linux-6.0.7-dev/lib/modules/6.0.7/build'
> make[1]: *** [Makefile:81: module] Error 2
> make[1]: Leaving directory '/build/source/module'
> make: *** [Makefile:8: all] Error 2
For full logs, run 'nix log /nix/store/pzmj5xfl0pb4f9f3d1lc9ylvdkr41kc6-evdi-1.12.0.drv'.
error: 1 dependencies of derivation '/nix/store/49lczfqfpp17y4vmpr04pkss0cwrb4pf-kernel-modules.drv' failed to build
error: 1 dependencies of derivation '/nix/store/q6p42l3r50xv4jphgpd9bx81y32r7m6y-linux-6.0.7-modules.drv' failed to build
error: 1 dependencies of derivation '/nix/store/pw5f02p73pvpj03a60pm0iysl88csdmn-nixos-system-Ryan-Laptop-22.11.20221108.0932685.drv' failed to build
error: 1 dependencies of derivation '/nix/store/zs7ma4qvymdrb999cqk1hyqpziadb0yc-nixops-machines.drv' failed to build
```
### Additional context
### Notify maintainers
@eyjhb
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
[user@system:~]$ nix-shell -p nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 5.19.16, NixOS, 22.11 (Raccoon), 22.11.20221031.d40fea9`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.0`
- channels(root): `"nixos"`
- nixpkgs: `/nix/store/mrpx61vyrmhqhr2ms9svscyc1h523gf1-source`
```
|
non_process
|
evdi broken with linux steps to reproduce nix boot kernelpackages pkgs linuxpackages services xserver videodrivers build log error builder for nix store evdi drv failed with exit code last log lines drm framebuffer get efb base make error some warnings being treated as errors make error make error make leaving directory nix store linux dev lib modules build make error make leaving directory build source module make error for full logs run nix log nix store evdi drv error dependencies of derivation nix store kernel modules drv failed to build error dependencies of derivation nix store linux modules drv failed to build error dependencies of derivation nix store nixos system ryan laptop drv failed to build error dependencies of derivation nix store nixops machines drv failed to build additional context notify maintainers eyjhb metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels root nixos nixpkgs nix store source
| 0
|
98,957
| 16,389,571,236
|
IssuesEvent
|
2021-05-17 14:34:39
|
Thanraj/linux-1
|
https://api.github.com/repos/Thanraj/linux-1
|
opened
|
CVE-2019-19241 (High) detected in linuxv5.0
|
security vulnerability
|
## CVE-2019-19241 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.4.2, the io_uring feature leads to requests that inadvertently have UID 0 and full capabilities, aka CID-181e448d8709. This is related to fs/io-wq.c, fs/io_uring.c, and net/socket.c. For example, an attacker can bypass intended restrictions on adding an IPv4 address to the loopback interface. This occurs because IORING_OP_SENDMSG operations, although requested in the context of an unprivileged user, are sometimes performed by a kernel worker thread without considering that context.
<p>Publish Date: 2019-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19241>CVE-2019-19241</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-19241 (High) detected in linuxv5.0 - ## CVE-2019-19241 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.4.2, the io_uring feature leads to requests that inadvertently have UID 0 and full capabilities, aka CID-181e448d8709. This is related to fs/io-wq.c, fs/io_uring.c, and net/socket.c. For example, an attacker can bypass intended restrictions on adding an IPv4 address to the loopback interface. This occurs because IORING_OP_SENDMSG operations, although requested in the context of an unprivileged user, are sometimes performed by a kernel worker thread without considering that context.
<p>Publish Date: 2019-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19241>CVE-2019-19241</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19241</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details in the linux kernel before the io uring feature leads to requests that inadvertently have uid and full capabilities aka cid this is related to fs io wq c fs io uring c and net socket c for example an attacker can bypass intended restrictions on adding an address to the loopback interface this occurs because ioring op sendmsg operations although requested in the context of an unprivileged user are sometimes performed by a kernel worker thread without considering that context publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
667,319
| 22,465,441,359
|
IssuesEvent
|
2022-06-22 00:56:43
|
pusharray/new-collections
|
https://api.github.com/repos/pusharray/new-collections
|
closed
|
Reduce popup width and use standard layout for list items
|
feature priority: high effort: low
|
- In popup or narrow layout use same style for list item as in tab selector
- When there is only 1 item display either hosname or title from first tab to prevent empty looking panel list
- ...
|
1.0
|
Reduce popup width and use standard layout for list items - - In popup or narrow layout use same style for list item as in tab selector
- When there is only 1 item display either hosname or title from first tab to prevent empty looking panel list
- ...
|
non_process
|
reduce popup width and use standard layout for list items in popup or narrow layout use same style for list item as in tab selector when there is only item display either hosname or title from first tab to prevent empty looking panel list
| 0
|
16,595
| 21,647,948,965
|
IssuesEvent
|
2022-05-06 05:51:01
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Hybrid Worker Certificate Installation
|
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
|
Reviewing the .\Export-RunAsCertificateToHybridWorker, it is not clear as to how it would identify the hybrid worker where the certificate is to be installed. Given that is the case, it makes me hesitate to run the script since I do not know where it is going to try to install the certificate. Plus, what if you have multiple hybrid workers? Wouldn't you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly?
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: a21ca143-2f33-5cea-94a8-ace7e9de5f9c
* Version Independent ID: d7f2ef01-8c25-770e-dfd9-37b98dc7ba29
* Content: [Run Azure Automation runbooks on a Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hrw-run-runbooks#runbook-auth-managed-identities)
* Content Source: [articles/automation/automation-hrw-run-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-hrw-run-runbooks.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **sudhirsneha**
|
1.0
|
Hybrid Worker Certificate Installation -
Reviewing the .\Export-RunAsCertificateToHybridWorker, it is not clear as to how it would identify the hybrid worker where the certificate is to be installed. Given that is the case, it makes me hesitate to run the script since I do not know where it is going to try to install the certificate. Plus, what if you have multiple hybrid workers? Wouldn't you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly?
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: a21ca143-2f33-5cea-94a8-ace7e9de5f9c
* Version Independent ID: d7f2ef01-8c25-770e-dfd9-37b98dc7ba29
* Content: [Run Azure Automation runbooks on a Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hrw-run-runbooks#runbook-auth-managed-identities)
* Content Source: [articles/automation/automation-hrw-run-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-hrw-run-runbooks.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **sudhirsneha**
|
process
|
hybrid worker certificate installation reviewing the export runascertificatetohybridworker it is not clear as to how it would identify the hybrid worker where the certificate is to be installed given that is the case it makes me hesitate to run the script since i do not know where it is going to try to install the certificate plus what if you have multiple hybrid workers wouldn t you want to run this against each one so they all have the certificate and where in the script would you make them known to install correctly document details β do not edit this section it is required for docs microsoft com β github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias sudhirsneha
| 1
|
1,949
| 4,771,091,234
|
IssuesEvent
|
2016-10-26 16:59:05
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Complete System.Diagnostics namespace
|
area-System.Diagnostics.Process netstandard2.0
|
These types are either missing or incomplete.
- [x] T:System.Diagnostics.CorrelationManager
- [x] T:System.Diagnostics.DebuggerStepperBoundaryAttribute
- [x] T:System.Diagnostics.DebuggerVisualizerAttribute
- [x] T:System.Diagnostics.MonitoringDescriptionAttribute
- [x] T:System.Diagnostics.Process
- [x] T:System.Diagnostics.ProcessModule
- [x] T:System.Diagnostics.ProcessModuleCollection
- [x] T:System.Diagnostics.ProcessThread
- [x] T:System.Diagnostics.ProcessThreadCollection
- [x] T:System.Diagnostics.ProcessWindowStyle
- [x] T:System.Diagnostics.StackFrame
- [x] T:System.Diagnostics.StackTrace
- [x] T:System.Diagnostics.SwitchAttribute
- [x] T:System.Diagnostics.SwitchLevelAttribute
- [x] T:System.Diagnostics.TraceListener
Some of these will need to throw PNSE, eg., many of the missing members of Process.
|
1.0
|
Complete System.Diagnostics namespace - These types are either missing or incomplete.
- [x] T:System.Diagnostics.CorrelationManager
- [x] T:System.Diagnostics.DebuggerStepperBoundaryAttribute
- [x] T:System.Diagnostics.DebuggerVisualizerAttribute
- [x] T:System.Diagnostics.MonitoringDescriptionAttribute
- [x] T:System.Diagnostics.Process
- [x] T:System.Diagnostics.ProcessModule
- [x] T:System.Diagnostics.ProcessModuleCollection
- [x] T:System.Diagnostics.ProcessThread
- [x] T:System.Diagnostics.ProcessThreadCollection
- [x] T:System.Diagnostics.ProcessWindowStyle
- [x] T:System.Diagnostics.StackFrame
- [x] T:System.Diagnostics.StackTrace
- [x] T:System.Diagnostics.SwitchAttribute
- [x] T:System.Diagnostics.SwitchLevelAttribute
- [x] T:System.Diagnostics.TraceListener
Some of these will need to throw PNSE, eg., many of the missing members of Process.
|
process
|
complete system diagnostics namespace these types are either missing or incomplete t system diagnostics correlationmanager t system diagnostics debuggerstepperboundaryattribute t system diagnostics debuggervisualizerattribute t system diagnostics monitoringdescriptionattribute t system diagnostics process t system diagnostics processmodule t system diagnostics processmodulecollection t system diagnostics processthread t system diagnostics processthreadcollection t system diagnostics processwindowstyle t system diagnostics stackframe t system diagnostics stacktrace t system diagnostics switchattribute t system diagnostics switchlevelattribute t system diagnostics tracelistener some of these will need to throw pnse eg many of the missing members of process
| 1
|
10,214
| 13,079,062,103
|
IssuesEvent
|
2020-08-01 01:46:38
|
kubeflow/kfserving
|
https://api.github.com/repos/kubeflow/kfserving
|
closed
|
Define KFServing related area labels
|
kind/process
|
It would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels.
Labels should be added here:
https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml
In order to get added to all the Kubeflow repos automatically.
Owners (people responsible for issue triage) can be listed in
https://github.com/kubeflow/community/blob/master/labels-owners.yaml
It would probably be good to define a top level label e.g. "area/kfserving".
As well as possible subareas e.g. "area/kfserving/modelservers" as appropriate.
|
1.0
|
Define KFServing related area labels - It would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels.
Labels should be added here:
https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml
In order to get added to all the Kubeflow repos automatically.
Owners (people responsible for issue triage) can be listed in
https://github.com/kubeflow/community/blob/master/labels-owners.yaml
It would probably be good to define a top level label e.g. "area/kfserving".
As well as possible subareas e.g. "area/kfserving/modelservers" as appropriate.
|
process
|
define kfserving related area labels it would be good to define an an appropriate set of area labels for kfserving and add them to our list of labels labels should be added here in order to get added to all the kubeflow repos automatically owners people responsible for issue triage can be listed in it would probably be good to define a top level label e g area kfserving as well as possible subareas e g area kfserving modelservers as appropriate
| 1
|
5,637
| 2,957,575,958
|
IssuesEvent
|
2015-07-08 17:01:11
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
closed
|
Embed doc URLs in swagger documentation
|
area/cluster-api area/usability kind/documentation priority/P2 team/UX
|
We should embed links to github docs in the API documentation. I think there's a swagger field for that.
/cc @nikhiljindal
|
1.0
|
Embed doc URLs in swagger documentation - We should embed links to github docs in the API documentation. I think there's a swagger field for that.
/cc @nikhiljindal
|
non_process
|
embed doc urls in swagger documentation we should embed links to github docs in the api documentation i think there s a swagger field for that cc nikhiljindal
| 0
|
17,040
| 22,420,243,723
|
IssuesEvent
|
2022-06-20 01:42:26
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Lauren is Explorin'
|
suggested title in process
|
**Is your feature request related to a problem? Please describe.**
[Watch classic episodes like Lauren is Explorin' is learning about Colors.]
**Describe the solution you'd like**
[TV-Y - 2 seasons - tv series, comedy]
**Describe alternatives you've considered**
Director: Craig McCracken
As seen in: Foster's Home for Imaginary Friends (2004-2004)
**Additional context**


|
1.0
|
Lauren is Explorin' - **Is your feature request related to a problem? Please describe.**
[Watch classic episodes like Lauren is Explorin' is learning about Colors.]
**Describe the solution you'd like**
[TV-Y - 2 seasons - tv series, comedy]
**Describe alternatives you've considered**
Director: Craig McCracken
As seen in: Foster's Home for Imaginary Friends (2004-2004)
**Additional context**


|
process
|
lauren is explorin is your feature request related to a problem please describe describe the solution you d like describe alternatives you ve considered director craig mccracken as seen in foster s home for imaginary friends additional context
| 1
|
89,949
| 15,856,029,081
|
IssuesEvent
|
2021-04-08 01:20:50
|
akshat702/cart-ionic
|
https://api.github.com/repos/akshat702/cart-ionic
|
opened
|
CVE-2020-36048 (High) detected in engine.io-3.2.1.tgz
|
security vulnerability
|
## CVE-2020-36048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-3.2.1.tgz</b></p></summary>
<p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p>
<p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz</a></p>
<p>Path to dependency file: cart-ionic/cart/package.json</p>
<p>Path to vulnerable library: cart-ionic/cart/e2e/node_modules/engine.io/package.json,cart-ionic/cart/e2e/node_modules/engine.io/package.json</p>
<p>
Dependency Hierarchy:
- karma-3.1.4.tgz (Root Library)
- socket.io-2.1.1.tgz
- :x: **engine.io-3.2.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport.
<p>Publish Date: 2021-01-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p>
<p>Release Date: 2021-01-08</p>
<p>Fix Resolution: engine.io - 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-36048 (High) detected in engine.io-3.2.1.tgz - ## CVE-2020-36048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-3.2.1.tgz</b></p></summary>
<p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p>
<p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-3.2.1.tgz</a></p>
<p>Path to dependency file: cart-ionic/cart/package.json</p>
<p>Path to vulnerable library: cart-ionic/cart/e2e/node_modules/engine.io/package.json,cart-ionic/cart/e2e/node_modules/engine.io/package.json</p>
<p>
Dependency Hierarchy:
- karma-3.1.4.tgz (Root Library)
- socket.io-2.1.1.tgz
- :x: **engine.io-3.2.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport.
<p>Publish Date: 2021-01-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p>
<p>Release Date: 2021-01-08</p>
<p>Fix Resolution: engine.io - 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in engine io tgz cve high severity vulnerability vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file cart ionic cart package json path to vulnerable library cart ionic cart node modules engine io package json cart ionic cart node modules engine io package json dependency hierarchy karma tgz root library socket io tgz x engine io tgz vulnerable library vulnerability details engine io before allows attackers to cause a denial of service resource consumption via a post request to the long polling transport publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io step up your open source security game with whitesource
| 0
|
15,069
| 18,765,576,505
|
IssuesEvent
|
2021-11-05 23:15:01
|
esmero/strawberryfield
|
https://api.github.com/repos/esmero/strawberryfield
|
closed
|
Add SOLR SBF Flavor Document removal when File gets removed
|
JSON Preprocessors Events and Subscriber Typed Data and Search
|
# What?
When we remove a file, old SBR documents could end being left in Solr. We may want to remove them automatically (using the Solr tracker) when a node and/or file gets removed.
- Different ways of removing depending if the Node is removed v/s a File from a Node
Important: Revision deletion triggers the same event as a full Node removal. Check that we do not remove everything if a revision is deleted.
|
1.0
|
Add SOLR SBF Flavor Document removal when File gets removed - # What?
When we remove a file, old SBR documents could end being left in Solr. We may want to remove them automatically (using the Solr tracker) when a node and/or file gets removed.
- Different ways of removing depending if the Node is removed v/s a File from a Node
Important: Revision deletion triggers the same event as a full Node removal. Check that we do not remove everything if a revision is deleted.
|
process
|
add solr sbf flavor document removal when file gets removed what when we remove a file old sbr documents could end being left in solr we may want to remove them automatically using the solr tracker when a node and or file gets removed different ways of removing depending if the node is removed v s a file from a node important revision deletion triggers the same event as a full node removal check that we do not remove everything if a revision is deleted
| 1
|
20,554
| 27,211,219,303
|
IssuesEvent
|
2023-02-20 16:41:54
|
TUM-Dev/NavigaTUM
|
https://api.github.com/repos/TUM-Dev/NavigaTUM
|
closed
|
[Entry] [5506.03.618]: Edit coordinate
|
entry webform delete-after-processing
|
Hello, I would like to add this coordinate to the roomfinder:
```yaml
"5506.03.618": { lat: 48.266430538130976, lon: 11.668311542516562 }
```
|
1.0
|
[Entry] [5506.03.618]: Edit coordinate - Hello, I would like to add this coordinate to the roomfinder:
```yaml
"5506.03.618": { lat: 48.266430538130976, lon: 11.668311542516562 }
```
|
process
|
edit coordinate hello i would like to add this coordinate to the roomfinder yaml lat lon
| 1
|
98,141
| 16,360,163,061
|
IssuesEvent
|
2021-05-14 08:13:26
|
Galaxy-Software-Service/Express-Demo
|
https://api.github.com/repos/Galaxy-Software-Service/Express-Demo
|
opened
|
CVE-2021-29469 (High) detected in redis-2.8.0.tgz
|
security vulnerability
|
## CVE-2021-29469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis-2.8.0.tgz</b></p></summary>
<p>Redis client library</p>
<p>Library home page: <a href="https://registry.npmjs.org/redis/-/redis-2.8.0.tgz">https://registry.npmjs.org/redis/-/redis-2.8.0.tgz</a></p>
<p>Path to dependency file: Express-Demo/package.json</p>
<p>Path to vulnerable library: Express-Demo/node_modules/redis/package.json</p>
<p>
Dependency Hierarchy:
- connect-redis-3.4.2.tgz (Root Library)
- :x: **redis-2.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Express-Demo/commit/41acf582b4ea1792c35f65c02df57132590f1702">41acf582b4ea1792c35f65c02df57132590f1702</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.
<p>Publish Date: 2021-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469>CVE-2021-29469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3">https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3</a></p>
<p>Release Date: 2021-04-23</p>
<p>Fix Resolution: redis - 3.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"redis","packageVersion":"2.8.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect-redis:3.4.2;redis:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"redis - 3.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29469","vulnerabilityDetails":"Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-29469 (High) detected in redis-2.8.0.tgz - ## CVE-2021-29469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis-2.8.0.tgz</b></p></summary>
<p>Redis client library</p>
<p>Library home page: <a href="https://registry.npmjs.org/redis/-/redis-2.8.0.tgz">https://registry.npmjs.org/redis/-/redis-2.8.0.tgz</a></p>
<p>Path to dependency file: Express-Demo/package.json</p>
<p>Path to vulnerable library: Express-Demo/node_modules/redis/package.json</p>
<p>
Dependency Hierarchy:
- connect-redis-3.4.2.tgz (Root Library)
- :x: **redis-2.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Express-Demo/commit/41acf582b4ea1792c35f65c02df57132590f1702">41acf582b4ea1792c35f65c02df57132590f1702</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.
<p>Publish Date: 2021-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469>CVE-2021-29469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3">https://github.com/NodeRedis/node-redis/security/advisories/GHSA-35q2-47q7-3pc3</a></p>
<p>Release Date: 2021-04-23</p>
<p>Fix Resolution: redis - 3.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"redis","packageVersion":"2.8.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect-redis:3.4.2;redis:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"redis - 3.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29469","vulnerabilityDetails":"Node-redis is a Node.js Redis client. Before version 3.1.1, when a client is in monitoring mode, the regex begin used to detected monitor messages could cause exponential backtracking on some strings. This issue could lead to a denial of service. The issue is patched in version 3.1.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in redis tgz cve high severity vulnerability vulnerable library redis tgz redis client library library home page a href path to dependency file express demo package json path to vulnerable library express demo node modules redis package json dependency hierarchy connect redis tgz root library x redis tgz vulnerable library found in head commit a href found in base branch master vulnerability details node redis is a node js redis client before version when a client is in monitoring mode the regex begin used to detected monitor messages could cause exponential backtracking on some strings this issue could lead to a denial of service the issue is patched in version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution redis isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree connect redis redis isminimumfixversionavailable true minimumfixversion redis basebranches vulnerabilityidentifier cve vulnerabilitydetails node redis is a node js redis client before version when a client is in monitoring mode the regex begin used to detected monitor messages could cause exponential backtracking on some strings this issue could lead to a denial of service the issue is patched in version vulnerabilityurl
| 0
|
15,171
| 18,942,926,941
|
IssuesEvent
|
2021-11-18 06:33:49
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371>
|
high priority triage review module: multiprocessing
|
## π Bug
<!-- A clear and concise description of what the bug is. -->
## To Reproduce
Steps to reproduce the behavior:
run a bert model based simcse traing.
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
67%|ββββββββββββββββββββββββ | 11156/16617 [1:04:38<31:12, 2.92it/s]Traceback (most recent call last):
File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/queues.py", line 234, in _feed
obj = _ForkingPickler.dumps(obj)
File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/reduction.py", line 51, in dumps
cls(buf, protocol).dump(obj)
File "/root/miniconda3/envs/se/lib/python3.6/site-packages/torch/multiprocessing/reductions.py", line 328, in reduce_storage
fd, size = storage._share_fd_()
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371>
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
In fact, the error only appeared today. Before, I could train and save the model normally.
## Environment
Collecting environment information...
PyTorch version: 1.10.0+cu102
Is debug build: False
CUDA used to build PyTorch: 10.2
ROCM used to build PyTorch: N/A
OS: Ubuntu 18.04.3 LTS (x86_64)
GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0
Clang version: Could not collect
CMake version: version 3.10.2
Libc version: glibc-2.17
Python version: 3.6.13 |Anaconda, Inc.| (default, Jun 4 2021, 14:25:59) [GCC 7.5.0] (64-bit runtime)
Python platform: Linux-3.10.0-957.el7.x86_64-x86_64-with-debian-buster-sid
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: GeForce RTX 2080 Ti
GPU 1: GeForce RTX 2080 Ti
GPU 2: GeForce RTX 2080 Ti
GPU 3: GeForce RTX 2080 Ti
GPU 4: GeForce RTX 2080 Ti
GPU 5: GeForce RTX 2080 Ti
GPU 6: GeForce RTX 2080 Ti
GPU 7: GeForce RTX 2080 Ti
Nvidia driver version: 440.33.01
cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.5
[pip3] torch==1.10.0
[pip3] torchaudio==0.10.0
[pip3] torchvision==0.11.1
[conda] cudatoolkit 10.1.243 h6bb024c_0
[conda] faiss-gpu 1.7.1 py3.6_hf4b0e32_1_cuda10.1 pytorch
[conda] libfaiss 1.7.1 hbf11de5_1_cuda10.1 pytorch
[conda] mkl 2021.3.0 h06a4308_520
[conda] numpy 1.19.5 pypi_0 pypi
[conda] numpy-base 1.17.0 py36h2f8d375_0
[conda] torch 1.10.0 pypi_0 pypi
[conda] torchaudio 0.10.0 pypi_0 pypi
[conda] torchvision 0.11.1 pypi_0 pypi
## Additional context
<!-- Add any other context about the problem here. -->
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @VitalyFedyunin
|
1.0
|
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371> - ## π Bug
<!-- A clear and concise description of what the bug is. -->
## To Reproduce
Steps to reproduce the behavior:
run a bert model based simcse traing.
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
67%|ββββββββββββββββββββββββ | 11156/16617 [1:04:38<31:12, 2.92it/s]Traceback (most recent call last):
File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/queues.py", line 234, in _feed
obj = _ForkingPickler.dumps(obj)
File "/root/miniconda3/envs/se/lib/python3.6/multiprocessing/reduction.py", line 51, in dumps
cls(buf, protocol).dump(obj)
File "/root/miniconda3/envs/se/lib/python3.6/site-packages/torch/multiprocessing/reductions.py", line 328, in reduce_storage
fd, size = storage._share_fd_()
RuntimeError: falseINTERNAL ASSERT FAILED at "../aten/src/ATen/MapAllocator.cpp":300, please report a bug to PyTorch. unable to write to file </torch_85217_8371>
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
In fact, the error only appeared today. Before, I could train and save the model normally.
## Environment
Collecting environment information...
PyTorch version: 1.10.0+cu102
Is debug build: False
CUDA used to build PyTorch: 10.2
ROCM used to build PyTorch: N/A
OS: Ubuntu 18.04.3 LTS (x86_64)
GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0
Clang version: Could not collect
CMake version: version 3.10.2
Libc version: glibc-2.17
Python version: 3.6.13 |Anaconda, Inc.| (default, Jun 4 2021, 14:25:59) [GCC 7.5.0] (64-bit runtime)
Python platform: Linux-3.10.0-957.el7.x86_64-x86_64-with-debian-buster-sid
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: GeForce RTX 2080 Ti
GPU 1: GeForce RTX 2080 Ti
GPU 2: GeForce RTX 2080 Ti
GPU 3: GeForce RTX 2080 Ti
GPU 4: GeForce RTX 2080 Ti
GPU 5: GeForce RTX 2080 Ti
GPU 6: GeForce RTX 2080 Ti
GPU 7: GeForce RTX 2080 Ti
Nvidia driver version: 440.33.01
cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.5
[pip3] torch==1.10.0
[pip3] torchaudio==0.10.0
[pip3] torchvision==0.11.1
[conda] cudatoolkit 10.1.243 h6bb024c_0
[conda] faiss-gpu 1.7.1 py3.6_hf4b0e32_1_cuda10.1 pytorch
[conda] libfaiss 1.7.1 hbf11de5_1_cuda10.1 pytorch
[conda] mkl 2021.3.0 h06a4308_520
[conda] numpy 1.19.5 pypi_0 pypi
[conda] numpy-base 1.17.0 py36h2f8d375_0
[conda] torch 1.10.0 pypi_0 pypi
[conda] torchaudio 0.10.0 pypi_0 pypi
[conda] torchvision 0.11.1 pypi_0 pypi
## Additional context
<!-- Add any other context about the problem here. -->
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @VitalyFedyunin
|
process
|
runtimeerror falseinternal assert failed at aten src aten mapallocator cpp please report a bug to pytorch unable to write to file π bug to reproduce steps to reproduce the behavior run a bert model based simcse traing ββββββββββββββββββββββββ traceback most recent call last file root envs se lib multiprocessing queues py line in feed obj forkingpickler dumps obj file root envs se lib multiprocessing reduction py line in dumps cls buf protocol dump obj file root envs se lib site packages torch multiprocessing reductions py line in reduce storage fd size storage share fd runtimeerror falseinternal assert failed at aten src aten mapallocator cpp please report a bug to pytorch unable to write to file expected behavior in fact the error only appeared today before i could train and save the model normally environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version anaconda inc default jun bit runtime python platform linux with debian buster sid is cuda available true cuda runtime version could not collect gpu models and configuration gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti gpu geforce rtx ti nvidia driver version cudnn version usr lib linux gnu libcudnn so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy torch torchaudio torchvision cudatoolkit faiss gpu pytorch libfaiss pytorch mkl numpy pypi pypi numpy base torch pypi pypi torchaudio pypi pypi torchvision pypi pypi additional context cc ezyang gchanan bdhirsh jbschlosser vitalyfedyunin
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.