Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
14,864
18,273,429,697
IssuesEvent
2021-10-04 15:59:29
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Re-introspection support using `db pull` with MongoDB
kind/feature process/candidate topic: re-introspection tech/engines tech/typescript topic: mongodb
Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection.
1.0
Re-introspection support using `db pull` with MongoDB - Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection.
process
re introspection support using db pull with mongodb introspection support for mongodb preview was shipped with prisma version without re introspection
1
766
3,252,741,960
IssuesEvent
2015-10-19 16:02:30
beesmart-it/trend-hrm
https://api.github.com/repos/beesmart-it/trend-hrm
closed
View selection process
client company requirement selection process
Access selection process view to see candidates by step, add candidates, manage process
1.0
View selection process - Access selection process view to see candidates by step, add candidates, manage process
process
view selection process access selection process view to see candidates by step add candidates manage process
1
638,064
20,712,178,022
IssuesEvent
2022-03-12 03:57:03
aitos-io/BoAT-X-Framework
https://api.github.com/repos/aitos-io/BoAT-X-Framework
closed
An error occurs when test case InitEthWalletGenerationKey is run.
bug Severity/moderate Priority/P2
`01Wallet.c:496:F:wallet_api:test_002InitWallet_0013InitEthWalletGenerationKey:0: Assertion 'rtnVal == NULL' failed: rtnVal == 0x80007a870`
1.0
An error occurs when test case InitEthWalletGenerationKey is run. - `01Wallet.c:496:F:wallet_api:test_002InitWallet_0013InitEthWalletGenerationKey:0: Assertion 'rtnVal == NULL' failed: rtnVal == 0x80007a870`
non_process
an error occurs when test case initethwalletgenerationkey is run c f wallet api test assertion rtnval null failed rtnval
0
15,348
2,850,649,105
IssuesEvent
2015-05-31 19:10:11
damonkohler/sl4a
https://api.github.com/repos/damonkohler/sl4a
opened
ERROR/sl4a.FileUtils:121(2049): Failed to create directory.
auto-migrated Priority-Medium Type-Defect
_From @GoogleCodeExporter on May 31, 2015 11:29_ ``` What device(s) are you experiencing the problem on? miumiu 9 (android 2.1) What firmware version are you running on the device? What steps will reproduce the problem? 1. install sl4a_r3.apk using adb 2. try to run it 3. What is the expected output? What do you see instead? Logcat 03-17 20:19:58.676: VERBOSE/sl4a.FileUtils:119(2049): Creating directory: scripts 03-17 20:19:58.676: ERROR/sl4a.FileUtils:121(2049): Failed to create directory. 03-17 20:19:58.676: DEBUG/AndroidRuntime(2049): Shutting down VM 03-17 20:19:58.676: WARN/dalvikvm(2049): threadid=3: thread exiting with uncaught exception (group=0x4001b168) 03-17 20:19:58.686: ERROR/AndroidRuntime(2049): Uncaught handler: thread main exiting due to uncaught exception 03-17 20:19:58.696: ERROR/AndroidRuntime(2049): java.lang.RuntimeException: Unable to start activity ComponentInfo{com.googlecode.android_scripting/com.googlecode.android_scripting. activity.ScriptManager}: java.lang.RuntimeException: Failed to create scripts directory. at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2496) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2512) at android.app.ActivityThread.access$2200(ActivityThread.java:119) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1863) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:123) at android.app.ActivityThread.main(ActivityThread.java:4363) at java.lang.reflect.Method.invokeNative(Native Method) at java.lang.reflect.Method.invoke(Method.java:521) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:860) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:618) at dalvik.system.NativeStart.main(Native Method) Caused by: java.lang.RuntimeException: Failed to create scripts directory. at com.googlecode.android_scripting.activity.ScriptManager.onCreate(ScriptManager.java:112) at android.app.Instrumentation.callActivityOnCreate(Instrumentation.java:1047) at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2459) ... 11 more What version of the product are you using? On what operating system? Please provide any additional information below. I unmounted the SD just in case ``` Original issue reported on code.google.com by `scottbur...@gmail.com` on 17 Mar 2011 at 8:23 _Copied from original issue: damonkohler/android-scripting#532_
1.0
ERROR/sl4a.FileUtils:121(2049): Failed to create directory. - _From @GoogleCodeExporter on May 31, 2015 11:29_ ``` What device(s) are you experiencing the problem on? miumiu 9 (android 2.1) What firmware version are you running on the device? What steps will reproduce the problem? 1. install sl4a_r3.apk using adb 2. try to run it 3. What is the expected output? What do you see instead? Logcat 03-17 20:19:58.676: VERBOSE/sl4a.FileUtils:119(2049): Creating directory: scripts 03-17 20:19:58.676: ERROR/sl4a.FileUtils:121(2049): Failed to create directory. 03-17 20:19:58.676: DEBUG/AndroidRuntime(2049): Shutting down VM 03-17 20:19:58.676: WARN/dalvikvm(2049): threadid=3: thread exiting with uncaught exception (group=0x4001b168) 03-17 20:19:58.686: ERROR/AndroidRuntime(2049): Uncaught handler: thread main exiting due to uncaught exception 03-17 20:19:58.696: ERROR/AndroidRuntime(2049): java.lang.RuntimeException: Unable to start activity ComponentInfo{com.googlecode.android_scripting/com.googlecode.android_scripting. activity.ScriptManager}: java.lang.RuntimeException: Failed to create scripts directory. at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2496) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2512) at android.app.ActivityThread.access$2200(ActivityThread.java:119) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1863) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:123) at android.app.ActivityThread.main(ActivityThread.java:4363) at java.lang.reflect.Method.invokeNative(Native Method) at java.lang.reflect.Method.invoke(Method.java:521) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:860) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:618) at dalvik.system.NativeStart.main(Native Method) Caused by: java.lang.RuntimeException: Failed to create scripts directory. at com.googlecode.android_scripting.activity.ScriptManager.onCreate(ScriptManager.java:112) at android.app.Instrumentation.callActivityOnCreate(Instrumentation.java:1047) at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2459) ... 11 more What version of the product are you using? On what operating system? Please provide any additional information below. I unmounted the SD just in case ``` Original issue reported on code.google.com by `scottbur...@gmail.com` on 17 Mar 2011 at 8:23 _Copied from original issue: damonkohler/android-scripting#532_
non_process
error fileutils failed to create directory from googlecodeexporter on may what device s are you experiencing the problem on miumiu android what firmware version are you running on the device what steps will reproduce the problem install apk using adb try to run it what is the expected output what do you see instead logcat verbose fileutils creating directory scripts error fileutils failed to create directory debug androidruntime shutting down vm warn dalvikvm threadid thread exiting with uncaught exception group error androidruntime uncaught handler thread main exiting due to uncaught exception error androidruntime java lang runtimeexception unable to start activity componentinfo com googlecode android scripting com googlecode android scripting activity scriptmanager java lang runtimeexception failed to create scripts directory at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative native method at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main native method caused by java lang runtimeexception failed to create scripts directory at com googlecode android scripting activity scriptmanager oncreate scriptmanager java at android app instrumentation callactivityoncreate instrumentation java at android app activitythread performlaunchactivity activitythread java more what version of the product are you using on what operating system please provide any additional information below i unmounted the sd just in case original issue reported on code google com by scottbur gmail com on mar at copied from original issue damonkohler android scripting
0
217,209
16,682,218,015
IssuesEvent
2021-06-08 02:17:15
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
closed
Bevy ECS crate needs a readme
documentation ecs
Many people would like to use bevy_ecs outside of the context of Bevy, but it isn't immediately clear that it can be used standalone (#2002). The crates.io page is pretty bare: ![image](https://user-images.githubusercontent.com/2694663/116002644-9570f580-a5af-11eb-93d6-3cb6eec0e51b.png) We should give `bevy_ecs` a nice readme that introduces the project (in an ECS-only context) and lets people know that it can be used by itself.
1.0
Bevy ECS crate needs a readme - Many people would like to use bevy_ecs outside of the context of Bevy, but it isn't immediately clear that it can be used standalone (#2002). The crates.io page is pretty bare: ![image](https://user-images.githubusercontent.com/2694663/116002644-9570f580-a5af-11eb-93d6-3cb6eec0e51b.png) We should give `bevy_ecs` a nice readme that introduces the project (in an ECS-only context) and lets people know that it can be used by itself.
non_process
bevy ecs crate needs a readme many people would like to use bevy ecs outside of the context of bevy but it isn t immediately clear that it can be used standalone the crates io page is pretty bare we should give bevy ecs a nice readme that introduces the project in an ecs only context and lets people know that it can be used by itself
0
37,771
6,641,629,904
IssuesEvent
2017-09-27 02:26:17
apache/incubator-openwhisk-wskdeploy
https://api.github.com/repos/apache/incubator-openwhisk-wskdeploy
closed
Integration Test - Dependency (package binding)
documentation priority: high
Add a working example of package binding and dependency section specified in [dependency under usecases](https://github.com/apache/incubator-openwhisk-wskdeploy/tree/master/tests/usecases/dependency). Delete dependency folder from use cases folder.
1.0
Integration Test - Dependency (package binding) - Add a working example of package binding and dependency section specified in [dependency under usecases](https://github.com/apache/incubator-openwhisk-wskdeploy/tree/master/tests/usecases/dependency). Delete dependency folder from use cases folder.
non_process
integration test dependency package binding add a working example of package binding and dependency section specified in delete dependency folder from use cases folder
0
113,645
9,659,884,918
IssuesEvent
2019-05-20 14:22:59
kcigeospatial/Fred_Co_Land-Management
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
closed
R4C-ResUse Detail Page, Setbacks
Ready For Retest
The Setbacks should have "in ft." next to them just as the Building Height does. -aw ![image](https://user-images.githubusercontent.com/47611580/58024384-0cf29c00-7ae0-11e9-8f92-9176300a4a97.png)
1.0
R4C-ResUse Detail Page, Setbacks - The Setbacks should have "in ft." next to them just as the Building Height does. -aw ![image](https://user-images.githubusercontent.com/47611580/58024384-0cf29c00-7ae0-11e9-8f92-9176300a4a97.png)
non_process
resuse detail page setbacks the setbacks should have in ft next to them just as the building height does aw
0
6,217
9,126,909,042
IssuesEvent
2019-02-25 01:13:32
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
Unable to interrupt inferior process
bug process
I haven't had time to look at this in much detail yet, but I'm not able to interrupt the inferior process with e.g. C-c C-c. The simplest way I could reproduce is with this R script: ``` cat("\n");Sys.sleep(100) ``` then C-c C-c to evaluate, switch to the inferior process and try to C-c C-c. I get a message about the text being read-only.
1.0
Unable to interrupt inferior process - I haven't had time to look at this in much detail yet, but I'm not able to interrupt the inferior process with e.g. C-c C-c. The simplest way I could reproduce is with this R script: ``` cat("\n");Sys.sleep(100) ``` then C-c C-c to evaluate, switch to the inferior process and try to C-c C-c. I get a message about the text being read-only.
process
unable to interrupt inferior process i haven t had time to look at this in much detail yet but i m not able to interrupt the inferior process with e g c c c c the simplest way i could reproduce is with this r script cat n sys sleep then c c c c to evaluate switch to the inferior process and try to c c c c i get a message about the text being read only
1
7,285
9,544,285,414
IssuesEvent
2019-05-01 13:45:28
ORelio/Minecraft-Console-Client
https://api.github.com/repos/ORelio/Minecraft-Console-Client
closed
Implement 1.14 support
a:enhancement in:protocol-compatibility in:terrain-handling resolved
Hi! I'm just curious if the latest version of MinecraftClient.exe supports Minecraft 1.14. Thanks!
True
Implement 1.14 support - Hi! I'm just curious if the latest version of MinecraftClient.exe supports Minecraft 1.14. Thanks!
non_process
implement support hi i m just curious if the latest version of minecraftclient exe supports minecraft thanks
0
8,329
11,490,371,586
IssuesEvent
2020-02-11 16:58:33
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Preprocessing overlapping data not implemented yet
preprocessor
Hi, I am having an issue with the preprocessing when trying to run a diagnostic which uses two variables from a single model: https://github.com/ESMValGroup/ESMValTool/blob/MAGIC_BSC/esmvaltool/recipes/recipe_diurnal_temperature_index_wp7.yml An unexpected problem prevented concatenation. Expected only a single cube, found 2. I guess this is just a minor error in our .yml file. Do you have an example of any other diagnostics which load two (or more) variables from a single model for use as input for a single R script (or Python script).
1.0
Preprocessing overlapping data not implemented yet - Hi, I am having an issue with the preprocessing when trying to run a diagnostic which uses two variables from a single model: https://github.com/ESMValGroup/ESMValTool/blob/MAGIC_BSC/esmvaltool/recipes/recipe_diurnal_temperature_index_wp7.yml An unexpected problem prevented concatenation. Expected only a single cube, found 2. I guess this is just a minor error in our .yml file. Do you have an example of any other diagnostics which load two (or more) variables from a single model for use as input for a single R script (or Python script).
process
preprocessing overlapping data not implemented yet hi i am having an issue with the preprocessing when trying to run a diagnostic which uses two variables from a single model an unexpected problem prevented concatenation expected only a single cube found i guess this is just a minor error in our yml file do you have an example of any other diagnostics which load two or more variables from a single model for use as input for a single r script or python script
1
2,884
5,848,169,551
IssuesEvent
2017-05-10 20:15:45
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Update copy-html to work like copy-image, with parameter for destination directory
feature P3 preprocess
In preprocess we have the `copy-image` task, which copies images from the source directory. For X/HTML this goes straight to the output directory. The task sets a property to figure out the destination directory. I've taken advantage of this by initializing the `copy-image.todir` property in a plugin: for one output type the images are placed in a generated directory in the temp dir, which is later zipped together with the rendered content. I'd like to be able to do the same with HTML topics referenced from the map - I should be able to set the copy destination in the same way as the images, so that they become part of my zip. Any objection to updating the `copy-html` task so that it works the same way as `copy-image`? I suspect the main reason for the property in `copy-image` is to handle the uplevels variable. I suspect the same uplevels support is probably missing today in `copy-html`, so that may be another reason to make this update. Current `copy-html` task: ``` <target name="copy-html" dita:depends="{depend.preprocess.copy-html.pre},copy-html-check" dita:extension="depends org.dita.dost.platform.InsertDependsAction" unless="preprocess.copy-html.skip" description="Copy html files"> <copy todir="${output.dir}"> <fileset dir="${user.input.dir}" includesfile="${dita.temp.dir}/${htmlfile}"/> </copy> </target> ``` Suggested revision, based on `copy-image`: ``` <target name="copy-html" dita:depends="{depend.preprocess.copy-html.pre},copy-html-check" dita:extension="depends org.dita.dost.platform.InsertDependsAction" unless="preprocess.copy-html.skip" description="Copy html files"> <condition property="copy-html.todir" value="${output.dir}/${uplevels}" else="${output.dir}"> <equals arg1="${generate.copy.outer}" arg2="1"/> </condition> <copy todir="${copy-html.todir}"> <fileset dir="${user.input.dir}" includesfile="${dita.temp.dir}/${htmlfile}" /> </copy> </target> ```
1.0
Update copy-html to work like copy-image, with parameter for destination directory - In preprocess we have the `copy-image` task, which copies images from the source directory. For X/HTML this goes straight to the output directory. The task sets a property to figure out the destination directory. I've taken advantage of this by initializing the `copy-image.todir` property in a plugin: for one output type the images are placed in a generated directory in the temp dir, which is later zipped together with the rendered content. I'd like to be able to do the same with HTML topics referenced from the map - I should be able to set the copy destination in the same way as the images, so that they become part of my zip. Any objection to updating the `copy-html` task so that it works the same way as `copy-image`? I suspect the main reason for the property in `copy-image` is to handle the uplevels variable. I suspect the same uplevels support is probably missing today in `copy-html`, so that may be another reason to make this update. Current `copy-html` task: ``` <target name="copy-html" dita:depends="{depend.preprocess.copy-html.pre},copy-html-check" dita:extension="depends org.dita.dost.platform.InsertDependsAction" unless="preprocess.copy-html.skip" description="Copy html files"> <copy todir="${output.dir}"> <fileset dir="${user.input.dir}" includesfile="${dita.temp.dir}/${htmlfile}"/> </copy> </target> ``` Suggested revision, based on `copy-image`: ``` <target name="copy-html" dita:depends="{depend.preprocess.copy-html.pre},copy-html-check" dita:extension="depends org.dita.dost.platform.InsertDependsAction" unless="preprocess.copy-html.skip" description="Copy html files"> <condition property="copy-html.todir" value="${output.dir}/${uplevels}" else="${output.dir}"> <equals arg1="${generate.copy.outer}" arg2="1"/> </condition> <copy todir="${copy-html.todir}"> <fileset dir="${user.input.dir}" includesfile="${dita.temp.dir}/${htmlfile}" /> </copy> </target> ```
process
update copy html to work like copy image with parameter for destination directory in preprocess we have the copy image task which copies images from the source directory for x html this goes straight to the output directory the task sets a property to figure out the destination directory i ve taken advantage of this by initializing the copy image todir property in a plugin for one output type the images are placed in a generated directory in the temp dir which is later zipped together with the rendered content i d like to be able to do the same with html topics referenced from the map i should be able to set the copy destination in the same way as the images so that they become part of my zip any objection to updating the copy html task so that it works the same way as copy image i suspect the main reason for the property in copy image is to handle the uplevels variable i suspect the same uplevels support is probably missing today in copy html so that may be another reason to make this update current copy html task target name copy html dita depends depend preprocess copy html pre copy html check dita extension depends org dita dost platform insertdependsaction unless preprocess copy html skip description copy html files suggested revision based on copy image target name copy html dita depends depend preprocess copy html pre copy html check dita extension depends org dita dost platform insertdependsaction unless preprocess copy html skip description copy html files
1
404,520
27,489,496,480
IssuesEvent
2023-03-04 12:46:14
ra3xdh/qucs_s
https://api.github.com/repos/ra3xdh/qucs_s
closed
Legacy Qucs equations versus Nutmeg (ngspice )
question documentation
I have 700 Qucs schematics with Equation blocks that have been "ported" over to Qucs-S. The vast majority of Equation blocks worked without converting them to Nutmeg equations. The 3rd example uses v(Output) versus Output.v. Is there a way to know what Qucs functions don't work in Qucs-S (ngspice)? The first three examples work but the last Equation block using "diff" does not. Trying to reconcile why it doesn't work when many other functions do. ![2022-10-31_175702](https://user-images.githubusercontent.com/102838604/199119698-72bdcb2b-5353-4c92-b66e-e04b0a8ad1aa.jpg) ![2022-10-31_180231](https://user-images.githubusercontent.com/102838604/199119700-16bcde3f-8ac8-472a-96d9-a52408ca08c6.jpg) ![2022-10-31_212718](https://user-images.githubusercontent.com/102838604/199138763-b2d8f06f-4d29-4baf-9aa0-6bc0cffc9cf1.jpg) ![2022-10-31_190321](https://user-images.githubusercontent.com/102838604/199138229-a409ee53-071a-4d7a-af14-0ac99cb68014.jpg)
1.0
Legacy Qucs equations versus Nutmeg (ngspice ) - I have 700 Qucs schematics with Equation blocks that have been "ported" over to Qucs-S. The vast majority of Equation blocks worked without converting them to Nutmeg equations. The 3rd example uses v(Output) versus Output.v. Is there a way to know what Qucs functions don't work in Qucs-S (ngspice)? The first three examples work but the last Equation block using "diff" does not. Trying to reconcile why it doesn't work when many other functions do. ![2022-10-31_175702](https://user-images.githubusercontent.com/102838604/199119698-72bdcb2b-5353-4c92-b66e-e04b0a8ad1aa.jpg) ![2022-10-31_180231](https://user-images.githubusercontent.com/102838604/199119700-16bcde3f-8ac8-472a-96d9-a52408ca08c6.jpg) ![2022-10-31_212718](https://user-images.githubusercontent.com/102838604/199138763-b2d8f06f-4d29-4baf-9aa0-6bc0cffc9cf1.jpg) ![2022-10-31_190321](https://user-images.githubusercontent.com/102838604/199138229-a409ee53-071a-4d7a-af14-0ac99cb68014.jpg)
non_process
legacy qucs equations versus nutmeg ngspice i have qucs schematics with equation blocks that have been ported over to qucs s the vast majority of equation blocks worked without converting them to nutmeg equations the example uses v output versus output v is there a way to know what qucs functions don t work in qucs s ngspice the first three examples work but the last equation block using diff does not trying to reconcile why it doesn t work when many other functions do
0
15,399
19,591,866,935
IssuesEvent
2022-01-05 13:52:05
remnoteio/remnote-issues
https://api.github.com/repos/remnoteio/remnote-issues
closed
Added images not shown in native MacOS app, but shown within browser
in-replication-process image-uploading to explore checked waiting for response check for duplicates
**Describe the bug** When reading through the document within the native MacOS app, I can't see my images at all, see here: ![Screen Shot 2021-02-05 at 14 07 32](https://user-images.githubusercontent.com/23134021/107032101-ee549d00-67bb-11eb-85a8-986d0cdddb3f.png) But the same document within the browser works fine with images and all the info, as shown here: ![Screen Shot 2021-02-05 at 14 07 45](https://user-images.githubusercontent.com/23134021/107032154-06c4b780-67bc-11eb-9aad-c2d74cd5f2a2.png) To be clear, images were originally uploaded within the Desktop app, copied from Word, or via screenshot. Screenshots were deleted after use. **To Reproduce** Steps to reproduce the behavior: 1. Add an image to remnote desktop app 2. Make sure all info is synced 3. Quit the app completely 4. Reopen and the image is blank 5. Images are shown within the browser app (Brave, in this case) **Expected behavior** Images are shown in all apps (Desktop, mobile, and web) **Screenshots** Added above **Desktop (please complete the following information):** - OS: [MacOS] - Browser [e.g. Brave] - Version [Desktop 1.1.12]
1.0
Added images not shown in native MacOS app, but shown within browser - **Describe the bug** When reading through the document within the native MacOS app, I can't see my images at all, see here: ![Screen Shot 2021-02-05 at 14 07 32](https://user-images.githubusercontent.com/23134021/107032101-ee549d00-67bb-11eb-85a8-986d0cdddb3f.png) But the same document within the browser works fine with images and all the info, as shown here: ![Screen Shot 2021-02-05 at 14 07 45](https://user-images.githubusercontent.com/23134021/107032154-06c4b780-67bc-11eb-9aad-c2d74cd5f2a2.png) To be clear, images were originally uploaded within the Desktop app, copied from Word, or via screenshot. Screenshots were deleted after use. **To Reproduce** Steps to reproduce the behavior: 1. Add an image to remnote desktop app 2. Make sure all info is synced 3. Quit the app completely 4. Reopen and the image is blank 5. Images are shown within the browser app (Brave, in this case) **Expected behavior** Images are shown in all apps (Desktop, mobile, and web) **Screenshots** Added above **Desktop (please complete the following information):** - OS: [MacOS] - Browser [e.g. Brave] - Version [Desktop 1.1.12]
process
added images not shown in native macos app but shown within browser describe the bug when reading through the document within the native macos app i can t see my images at all see here but the same document within the browser works fine with images and all the info as shown here to be clear images were originally uploaded within the desktop app copied from word or via screenshot screenshots were deleted after use to reproduce steps to reproduce the behavior add an image to remnote desktop app make sure all info is synced quit the app completely reopen and the image is blank images are shown within the browser app brave in this case expected behavior images are shown in all apps desktop mobile and web screenshots added above desktop please complete the following information os browser version
1
55,464
30,761,981,488
IssuesEvent
2023-07-29 20:35:36
TheGameCreators/GameGuruRepo
https://api.github.com/repos/TheGameCreators/GameGuruRepo
closed
MAX - Huge performance hit/freeze from exploding barrels
Fixed but need confirmation. Max High Performance Physics
Was playing Canyon Offensive to test out some of my combat improvements I was working on, and I noticed that the engine comes to a screeching halt whenever a barrel explodes. (Especially the clustered up groups of barrels, when they all go off it causes a 1-2 second freeze.) To replicate, play the Canyon Offensive demo, and look for one of the explosive barrel clusters. Shooting them will cause a severe nosedive in performance, and often a full-on brief freeze. It seems to be more drastic/pronounced the more entities that are sitting near the barrel(s), so my guess is it might have something to do with the way explosions are looking for nearby entities to apply damage(?)
True
MAX - Huge performance hit/freeze from exploding barrels - Was playing Canyon Offensive to test out some of my combat improvements I was working on, and I noticed that the engine comes to a screeching halt whenever a barrel explodes. (Especially the clustered up groups of barrels, when they all go off it causes a 1-2 second freeze.) To replicate, play the Canyon Offensive demo, and look for one of the explosive barrel clusters. Shooting them will cause a severe nosedive in performance, and often a full-on brief freeze. It seems to be more drastic/pronounced the more entities that are sitting near the barrel(s), so my guess is it might have something to do with the way explosions are looking for nearby entities to apply damage(?)
non_process
max huge performance hit freeze from exploding barrels was playing canyon offensive to test out some of my combat improvements i was working on and i noticed that the engine comes to a screeching halt whenever a barrel explodes especially the clustered up groups of barrels when they all go off it causes a second freeze to replicate play the canyon offensive demo and look for one of the explosive barrel clusters shooting them will cause a severe nosedive in performance and often a full on brief freeze it seems to be more drastic pronounced the more entities that are sitting near the barrel s so my guess is it might have something to do with the way explosions are looking for nearby entities to apply damage
0
358,970
25,210,890,687
IssuesEvent
2022-11-14 03:42:17
risingwavelabs/risingwave-docs
https://api.github.com/repos/risingwavelabs/risingwave-docs
closed
Document where the data is stored in the tutorials
documentation
### Related code PR _No response_ ### Which part(s) of the docs might be affected or should be updated? And how? All of our tutorials, we do not mention where the data of the MVs is stored. For our demo cluster, the data is stored in the MinIO instance that is running in the demo cluster. Please reflect this in all of our tutorials. ### Reference _No response_
1.0
Document where the data is stored in the tutorials - ### Related code PR _No response_ ### Which part(s) of the docs might be affected or should be updated? And how? All of our tutorials, we do not mention where the data of the MVs is stored. For our demo cluster, the data is stored in the MinIO instance that is running in the demo cluster. Please reflect this in all of our tutorials. ### Reference _No response_
non_process
document where the data is stored in the tutorials related code pr no response which part s of the docs might be affected or should be updated and how all of our tutorials we do not mention where the data of the mvs is stored for our demo cluster the data is stored in the minio instance that is running in the demo cluster please reflect this in all of our tutorials reference no response
0
726,733
25,009,432,813
IssuesEvent
2022-11-03 14:16:52
Signbank/Global-signbank
https://api.github.com/repos/Signbank/Global-signbank
closed
Change text on login page for registration
enhancement high priority
The registration option now says "Register for free to provide feedback on the Signbank." That's long outdated, users have better reasons now to get an account: to get access to datasets. Therefore, please change to "Register in order to get access to datasets"
1.0
Change text on login page for registration - The registration option now says "Register for free to provide feedback on the Signbank." That's long outdated, users have better reasons now to get an account: to get access to datasets. Therefore, please change to "Register in order to get access to datasets"
non_process
change text on login page for registration the registration option now says register for free to provide feedback on the signbank that s long outdated users have better reasons now to get an account to get access to datasets therefore please change to register in order to get access to datasets
0
8,071
11,251,358,819
IssuesEvent
2020-01-11 00:04:59
googleapis/java-monitoring-dashboards
https://api.github.com/repos/googleapis/java-monitoring-dashboards
opened
Promote to GA
type: process
Package name: **google-cloud-monitoring-dashboard** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
Promote to GA - Package name: **google-cloud-monitoring-dashboard** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
promote to ga package name google cloud monitoring dashboard current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
17,227
22,844,205,573
IssuesEvent
2022-07-13 03:06:43
NCAR/ucomp-pipeline
https://api.github.com/repos/NCAR/ucomp-pipeline
closed
Process 20220313 and 20220316 with just RCAM and just TCAM
process
To determine if new issues with UCoMP images are due to one of the cameras, process 20220313 and 20220316 with just RCAM and just TCAM. #### RCAM - [x] 20220313 - [x] 20220316 - [x] 20220709 #### TCAM - [x] 20220313 - [x] 20220316 - [x] 20220709
1.0
Process 20220313 and 20220316 with just RCAM and just TCAM - To determine if new issues with UCoMP images are due to one of the cameras, process 20220313 and 20220316 with just RCAM and just TCAM. #### RCAM - [x] 20220313 - [x] 20220316 - [x] 20220709 #### TCAM - [x] 20220313 - [x] 20220316 - [x] 20220709
process
process and with just rcam and just tcam to determine if new issues with ucomp images are due to one of the cameras process and with just rcam and just tcam rcam tcam
1
19,155
25,235,542,029
IssuesEvent
2022-11-15 00:21:24
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
How should we pick EIP numbers?
w-stale question r-ci r-process
In light of recent events, the discussion on how to assign numbers has come up again. This issue is _solely_ for discussing assigning EIP numbers, and other topics will be deleted. I'll try to summarize the proposals here, in preparation for the next EIPIP meeting. I'm going to give each proposal a `short-name` for easy reference. | Slug | Description | | ---- | --- | | `status-quo` | No change; editors continue to assign EIP numbers arbitrarily, loosely based on the Pull Request number. | | `seq-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is exactly one more than the largest existing EIP number. | | `rand-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is equal to: `R` plus the lowest unassigned EIP number and `R` is a random number such that `0 <= R < 100`; rerolling if `N` is an existing EIP. | | `seq-rand-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is equal to: the largest existing EIP number plus a random integer `R`, where `0 < R < 11`. | | `deny-list` | Create a known list of vanity numbers that are skipped, otherwise the same as `status-quo`. |
1.0
How should we pick EIP numbers? - In light of recent events, the discussion on how to assign numbers has come up again. This issue is _solely_ for discussing assigning EIP numbers, and other topics will be deleted. I'll try to summarize the proposals here, in preparation for the next EIPIP meeting. I'm going to give each proposal a `short-name` for easy reference. | Slug | Description | | ---- | --- | | `status-quo` | No change; editors continue to assign EIP numbers arbitrarily, loosely based on the Pull Request number. | | `seq-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is exactly one more than the largest existing EIP number. | | `rand-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is equal to: `R` plus the lowest unassigned EIP number and `R` is a random number such that `0 <= R < 100`; rerolling if `N` is an existing EIP. | | `seq-rand-bot` | A bot assigns an EIP number `N` in the commit when merging a Pull Request, where `N` is equal to: the largest existing EIP number plus a random integer `R`, where `0 < R < 11`. | | `deny-list` | Create a known list of vanity numbers that are skipped, otherwise the same as `status-quo`. |
process
how should we pick eip numbers in light of recent events the discussion on how to assign numbers has come up again this issue is solely for discussing assigning eip numbers and other topics will be deleted i ll try to summarize the proposals here in preparation for the next eipip meeting i m going to give each proposal a short name for easy reference slug description status quo no change editors continue to assign eip numbers arbitrarily loosely based on the pull request number seq bot a bot assigns an eip number n in the commit when merging a pull request where n is exactly one more than the largest existing eip number rand bot a bot assigns an eip number n in the commit when merging a pull request where n is equal to r plus the lowest unassigned eip number and r is a random number such that r rerolling if n is an existing eip seq rand bot a bot assigns an eip number n in the commit when merging a pull request where n is equal to the largest existing eip number plus a random integer r where r deny list create a known list of vanity numbers that are skipped otherwise the same as status quo
1
721,097
24,817,938,008
IssuesEvent
2022-10-25 14:24:42
COS301-SE-2022/Office-Booker
https://api.github.com/repos/COS301-SE-2022/Office-Booker
closed
Cannot delete certain bookings
Priority: Low Type: Fix Status: On-hold Status: Ready
Some bookings have a bookingvotedon entry associated with them, which prevents deletion.
1.0
Cannot delete certain bookings - Some bookings have a bookingvotedon entry associated with them, which prevents deletion.
non_process
cannot delete certain bookings some bookings have a bookingvotedon entry associated with them which prevents deletion
0
15,357
19,530,500,073
IssuesEvent
2021-12-30 15:53:08
redwoodjs/redwood
https://api.github.com/repos/redwoodjs/redwood
closed
LogFormatter does not output script exec logs
triage/processing release:fix
Reported by @simoncrypta The new LogFormatter that pipes dev and api server logs to pretty print is not used when running scripts.
1.0
LogFormatter does not output script exec logs - Reported by @simoncrypta The new LogFormatter that pipes dev and api server logs to pretty print is not used when running scripts.
process
logformatter does not output script exec logs reported by simoncrypta the new logformatter that pipes dev and api server logs to pretty print is not used when running scripts
1
21,362
29,194,079,763
IssuesEvent
2023-05-20 00:31:42
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / Caieiras, São Paulo, Brazil] Fullstack Developer .NET/Angular (Híbrido) na Coodesh
SALVADOR PJ FULL-STACK MVC HTML PLENO GIT REST ANGULAR REQUISITOS ASP.NET PROCESSOS BACKEND GITHUB UMA C RAZOR AUTOMAÇÃO DE PROCESSOS HIBRIDO ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-173601126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>Fullstack developer .NET/Angular (Híbrido)</ins></strong> Nível Júnior/pleno para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - .NET - C# .NET Core - Asp.Net MVC - Angular - HTML ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - BackEnd (plataforma .net, C#); - Front (HTML, Angular, MVC, Razor / Blazor); - Banco (SqlServer); - API (API.Rest, Web Services); - Controle de versão (Git). ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer .NET/Angular (Híbrido) na Triforsys](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-173601126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
2.0
[Hibrido / Caieiras, São Paulo, Brazil] Fullstack Developer .NET/Angular (Híbrido) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-173601126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>Fullstack developer .NET/Angular (Híbrido)</ins></strong> Nível Júnior/pleno para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - .NET - C# .NET Core - Asp.Net MVC - Angular - HTML ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - BackEnd (plataforma .net, C#); - Front (HTML, Angular, MVC, Razor / Blazor); - Banco (SqlServer); - API (API.Rest, Web Services); - Controle de versão (Git). ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer .NET/Angular (Híbrido) na Triforsys](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-173601126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
process
fullstack developer net angular híbrido na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a triforsys está em busca de fullstack developer net angular híbrido nível júnior pleno para integrar seu time fábrica softwares e soluções em ti de forma personalizada atendendo a necessidade do nosso cliente com foco na automação de processos sobre a empresa somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é venha com a gente fazer seu negócio triforsys somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é habilidades net c net core asp net mvc angular html local caieiras são paulo brazil requisitos residir em caieiras ou proximidades formato híbrido backend plataforma net c front html angular mvc razor blazor banco sqlserver api api rest web services controle de versão git como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
1
13,259
15,728,470,805
IssuesEvent
2021-03-29 13:51:28
modm-io/modm
https://api.github.com/repos/modm-io/modm
closed
Create a linter tool to ensure source code formatting
feature 🚧 process 📊
While most source code roughly follows the guidelines, it has occured to me that there are many cases of inconsistent formatting across the code base. This is annoying, but good news is that it is easily automatable (so that we can focus on more important things, such as my fiber PR #439 :) Coding convention is here: https://github.com/modm-io/modm/blob/develop/docs/coding_convention.md This is an ask to either write a linter for the existing style (which we can run in CI), or to switch to a style where there already is a supported linter / formatter tool (my proposal would be Google style, but I am obviously biased). See https://google.github.io/styleguide/cppguide.html and https://github.com/cpplint/cpplint. Any opinions here?
1.0
Create a linter tool to ensure source code formatting - While most source code roughly follows the guidelines, it has occured to me that there are many cases of inconsistent formatting across the code base. This is annoying, but good news is that it is easily automatable (so that we can focus on more important things, such as my fiber PR #439 :) Coding convention is here: https://github.com/modm-io/modm/blob/develop/docs/coding_convention.md This is an ask to either write a linter for the existing style (which we can run in CI), or to switch to a style where there already is a supported linter / formatter tool (my proposal would be Google style, but I am obviously biased). See https://google.github.io/styleguide/cppguide.html and https://github.com/cpplint/cpplint. Any opinions here?
process
create a linter tool to ensure source code formatting while most source code roughly follows the guidelines it has occured to me that there are many cases of inconsistent formatting across the code base this is annoying but good news is that it is easily automatable so that we can focus on more important things such as my fiber pr coding convention is here this is an ask to either write a linter for the existing style which we can run in ci or to switch to a style where there already is a supported linter formatter tool my proposal would be google style but i am obviously biased see and any opinions here
1
297,162
9,161,301,946
IssuesEvent
2019-03-01 10:05:48
RavenProject/ravenwallet-ios
https://api.github.com/repos/RavenProject/ravenwallet-ios
closed
Qt - Wallet shows protocol raven:<address>&amount=22... etc
FIXED / DONE Low Priority enhancement
The iPhone scan should honor this protocol and fill in a send with the field info.
1.0
Qt - Wallet shows protocol raven:<address>&amount=22... etc - The iPhone scan should honor this protocol and fill in a send with the field info.
non_process
qt wallet shows protocol raven amount etc the iphone scan should honor this protocol and fill in a send with the field info
0
109,737
16,890,355,185
IssuesEvent
2021-06-23 08:30:47
epam/TimeBase
https://api.github.com/repos/epam/TimeBase
closed
CVE-2018-14042 (Medium) detected in bootstrap-3.3.7.min.js - autoclosed
security vulnerability
## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: TimeBase/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp</p> <p>Path to vulnerable library: TimeBase/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.7","packageFilePaths":["/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-14042","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-14042 (Medium) detected in bootstrap-3.3.7.min.js - autoclosed - ## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: TimeBase/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp</p> <p>Path to vulnerable library: TimeBase/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.7","packageFilePaths":["/java/timebase/web/src/main/webapp/WEB-INF/jsp/common/header.jsp"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-14042","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in bootstrap min js autoclosed cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file timebase java timebase web src main webapp web inf jsp common header jsp path to vulnerable library timebase java timebase web src main webapp web inf jsp common header jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree twitter bootstrap isminimumfixversionavailable true minimumfixversion org webjars npm bootstrap org webjars bootstrap basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the data container property of tooltip vulnerabilityurl
0
5,563
8,403,826,868
IssuesEvent
2018-10-11 10:56:58
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Travis error on realpath
type: process
We're seeing this in Travis: > toolversions.sh: line 6: realpath: command not found I'm not sure why, but we should look into it.
1.0
Travis error on realpath - We're seeing this in Travis: > toolversions.sh: line 6: realpath: command not found I'm not sure why, but we should look into it.
process
travis error on realpath we re seeing this in travis toolversions sh line realpath command not found i m not sure why but we should look into it
1
10,060
13,044,161,778
IssuesEvent
2020-07-29 03:47:26
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `AddDateIntInt` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `AddDateIntInt` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `AddDateIntInt` from TiDB - ## Description Port the scalar function `AddDateIntInt` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function adddateintint from tidb description port the scalar function adddateintint from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
537,295
15,726,757,359
IssuesEvent
2021-03-29 11:46:24
everyday-as/gmodstore-issues
https://api.github.com/repos/everyday-as/gmodstore-issues
closed
Roadmap isn't ordered properly
Low Priority 🐞 Bug
## Expected Behavior Future versions on the roadmap should be ordered by name ascending ## Actual Behavior They're currently ordered randomly, see: ![image](https://user-images.githubusercontent.com/17050455/109567235-5493bd80-7ae5-11eb-93a0-6193d9e8bf01.png) ## Specifications - GmodStore version (see footer): V8.3.49 - Example url: [https://www.gmodstore.com/about/roadmap](https://www.gmodstore.com/about/roadmap) - Browser: - Event ID:
1.0
Roadmap isn't ordered properly - ## Expected Behavior Future versions on the roadmap should be ordered by name ascending ## Actual Behavior They're currently ordered randomly, see: ![image](https://user-images.githubusercontent.com/17050455/109567235-5493bd80-7ae5-11eb-93a0-6193d9e8bf01.png) ## Specifications - GmodStore version (see footer): V8.3.49 - Example url: [https://www.gmodstore.com/about/roadmap](https://www.gmodstore.com/about/roadmap) - Browser: - Event ID:
non_process
roadmap isn t ordered properly expected behavior future versions on the roadmap should be ordered by name ascending actual behavior they re currently ordered randomly see specifications gmodstore version see footer example url browser event id
0
25,361
12,237,705,642
IssuesEvent
2020-05-04 18:29:33
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
domainJoinOptions has no Explanation
Pri2 active-directory/svc assigned-to-author domain-services/subsvc product-question triaged
In the ARM template is `domainJoinOptions` setting. I think this documentation should provide possible values or a link with more details. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2568296a-984a-5dd0-d455-96359880a5e5 * Version Independent ID: efbc5520-001c-d721-2f98-ea241d525750 * Content: [Use a template to join a Windows VM to Azure AD DS](https://docs.microsoft.com/en-us/azure/active-directory-domain-services/join-windows-vm-template) * Content Source: [articles/active-directory-domain-services/join-windows-vm-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory-domain-services/join-windows-vm-template.md) * Service: **active-directory** * Sub-service: **domain-services** * GitHub Login: @iainfoulds * Microsoft Alias: **iainfou**
1.0
domainJoinOptions has no Explanation - In the ARM template is `domainJoinOptions` setting. I think this documentation should provide possible values or a link with more details. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2568296a-984a-5dd0-d455-96359880a5e5 * Version Independent ID: efbc5520-001c-d721-2f98-ea241d525750 * Content: [Use a template to join a Windows VM to Azure AD DS](https://docs.microsoft.com/en-us/azure/active-directory-domain-services/join-windows-vm-template) * Content Source: [articles/active-directory-domain-services/join-windows-vm-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory-domain-services/join-windows-vm-template.md) * Service: **active-directory** * Sub-service: **domain-services** * GitHub Login: @iainfoulds * Microsoft Alias: **iainfou**
non_process
domainjoinoptions has no explanation in the arm template is domainjoinoptions setting i think this documentation should provide possible values or a link with more details document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service active directory sub service domain services github login iainfoulds microsoft alias iainfou
0
11,769
14,598,956,257
IssuesEvent
2020-12-21 02:40:26
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
closed
The EOS Decision and Length Extrapolation
2020 EMNLP Natural Language Processing _read_later
* https://arxiv.org/abs/2010.07174 * Blackbox NLP Workshop at EMNLP 2020 見たことのないシーケンスの長さへの外挿は、言語のニューラル生成モデルにとっての課題である。 この研究では、見落とされがちなモデル化の決定である、特別なEOS(End-of-sequence)語彙項目の使用による生成プロセスの終了予測の長さ外挿の効果を特徴づける。 我々は、EOSを予測するように訓練されたネットワーク(+EOS)と訓練されていないネットワーク(-EOS)の長さ外挿の挙動を比較するために、テスト時に正しいシーケンス長にモデルを生成するように強制するというオラクル設定を研究している。 我々は、-EOSが+EOSよりも実質的に優れていることを発見した。 例えば、ブラケットを閉じるタスクでは、トレーニング時に見られる長さの10倍の長さにうまく外挿し、困難なSCANデータセットの長さ一般化タスクでは+EOSよりも40%の改善を達成した。 EOSと+EOSモデルの隠れた状態とダイナミクスを比較することで、+EOSモデルが一般化に失敗する理由は、 (1)線形位置が連続していることによって隠れた状態を不必要に層化しているため(我々は長さマニホールドと呼ぶ構造)、または (2)EOSトークンが最高確率であるとクラスタ(我々は長さアトラクターと呼ぶ)にはまってしまうためであることがわかります。
1.0
The EOS Decision and Length Extrapolation - * https://arxiv.org/abs/2010.07174 * Blackbox NLP Workshop at EMNLP 2020 見たことのないシーケンスの長さへの外挿は、言語のニューラル生成モデルにとっての課題である。 この研究では、見落とされがちなモデル化の決定である、特別なEOS(End-of-sequence)語彙項目の使用による生成プロセスの終了予測の長さ外挿の効果を特徴づける。 我々は、EOSを予測するように訓練されたネットワーク(+EOS)と訓練されていないネットワーク(-EOS)の長さ外挿の挙動を比較するために、テスト時に正しいシーケンス長にモデルを生成するように強制するというオラクル設定を研究している。 我々は、-EOSが+EOSよりも実質的に優れていることを発見した。 例えば、ブラケットを閉じるタスクでは、トレーニング時に見られる長さの10倍の長さにうまく外挿し、困難なSCANデータセットの長さ一般化タスクでは+EOSよりも40%の改善を達成した。 EOSと+EOSモデルの隠れた状態とダイナミクスを比較することで、+EOSモデルが一般化に失敗する理由は、 (1)線形位置が連続していることによって隠れた状態を不必要に層化しているため(我々は長さマニホールドと呼ぶ構造)、または (2)EOSトークンが最高確率であるとクラスタ(我々は長さアトラクターと呼ぶ)にはまってしまうためであることがわかります。
process
the eos decision and length extrapolation blackbox nlp workshop at emnlp 見たことのないシーケンスの長さへの外挿は、言語のニューラル生成モデルにとっての課題である。 この研究では、見落とされがちなモデル化の決定である、特別なeos(end of sequence)語彙項目の使用による生成プロセスの終了予測の長さ外挿の効果を特徴づける。 我々は、eosを予測するように訓練されたネットワーク eos と訓練されていないネットワーク eos の長さ外挿の挙動を比較するために、テスト時に正しいシーケンス長にモデルを生成するように強制するというオラクル設定を研究している。 我々は、 eosが eosよりも実質的に優れていることを発見した。 例えば、ブラケットを閉じるタスクでは、 、困難なscanデータセットの長さ一般化タスクでは %の改善を達成した。 eosと eosモデルの隠れた状態とダイナミクスを比較することで、 eosモデルが一般化に失敗する理由は、 線形位置が連続していることによって隠れた状態を不必要に層化しているため(我々は長さマニホールドと呼ぶ構造)、または eosトークンが最高確率であるとクラスタ(我々は長さアトラクターと呼ぶ)にはまってしまうためであることがわかります。
1
13,951
16,725,432,124
IssuesEvent
2021-06-10 12:28:22
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
The startup query `SELECT @@socket` should be ignored when connecting to a remote database
bug/2-confirmed kind/bug process/candidate team/client topic: connections
## Bug description I have a server that is using the [TDSQL for MySQL](https://intl.cloud.tencent.com/product/dcdb) distribution database provided by Tencent Cloud. The problem is Prisma Client will get a error when startup: ``` 2021-03-23T17:11:16.362175968+08:00 prisma:info Starting a mysql pool with 91 connections. 2021-03-23T17:11:16.373958504+08:00 prisma:info Fetched a connection from the pool 2021-03-23T17:11:16.373999662+08:00 prisma:info Encountered error during initialization: 2021-03-23T17:11:16.38391327+08:00 prisma:error in 0: user_facing_errors::Error::new_non_panic_with_current_backtrace 2021-03-23T17:11:16.383934881+08:00 1: query_engine::error::<impl core::convert::From<query_engine::error::PrismaError> for user_facing_errors::Error>::from 2021-03-23T17:11:16.383941043+08:00 2: query_engine::error::PrismaError::render_as_json 2021-03-23T17:11:16.383946603+08:00 3: query_engine::main::main::{{closure}}::{{closure}} 2021-03-23T17:11:16.383954968+08:00 4: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll 2021-03-23T17:11:16.383959586+08:00 5: std::thread::local::LocalKey<T>::with 2021-03-23T17:11:16.383963003+08:00 6: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll 2021-03-23T17:11:16.383966609+08:00 7: async_io::driver::block_on 2021-03-23T17:11:16.384001264+08:00 8: async_global_executor::reactor::block_on 2021-03-23T17:11:16.384038635+08:00 9: std::thread::local::LocalKey<T>::with 2021-03-23T17:11:16.384044606+08:00 10: async_std::task::builder::Builder::blocking 2021-03-23T17:11:16.384048593+08:00 11: query_engine::main 2021-03-23T17:11:16.384078069+08:00 12: std::sys_common::backtrace::__rust_begin_short_backtrace 2021-03-23T17:11:16.384082347+08:00 13: std::rt::lang_start::{{closure}} 2021-03-23T17:11:16.384126229+08:00 14: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once 2021-03-23T17:11:16.38413206+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/core/src/ops/function.rs:259:13 2021-03-23T17:11:16.384136979+08:00 std::panicking::try::do_call 2021-03-23T17:11:16.384150434+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panicking.rs:379:40 2021-03-23T17:11:16.384155794+08:00 std::panicking::try 2021-03-23T17:11:16.384176193+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panicking.rs:343:19 2021-03-23T17:11:16.38418046+08:00 std::panic::catch_unwind 2021-03-23T17:11:16.384183135+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panic.rs:396:14 2021-03-23T17:11:16.384185549+08:00 std::rt::lang_start_internal 2021-03-23T17:11:16.384193123+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/rt.rs:51:25 2021-03-23T17:11:16.384200237+08:00 15: main 2021-03-23T17:11:16.384203643+08:00 Error querying the database: Server error: `ERROR HY000 (1193): Unknown system variable 'socket'' ``` So I had a check of the Prisma client startup queries, it's executed: ```csv Connect | root@172.17.0.1 on mydb using TCP/IP Query | SELECT @@socket Query | SELECT @@max_allowed_packet Query | SELECT @@wait_timeout ``` Prisma will always select the `@@socket` variable, it's not necessary for the remote database in my opinion. ## How to reproduce 1. Deploy a server on Tencent Cloud with a TDSQL database, with Prisma Client. ## Expected behavior 1. Prisma startup will not execute the `SELECT @@socket`, because TDSQL is not defined. 2. The server will startup correctly. ## Environment & setup - OS: CentOS 8 - Database: TDSQL with MySQL - Node.js version: 14.15.4 - Prisma version: ``` prisma : 2.19.0 @prisma/client : 2.19.0 Current platform : rhel-openssl-1.1.x Query Engine : query-engine c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x) Migration Engine : migration-engine-cli c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x) Introspection Engine : introspection-core c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x) Format Binary : prisma-fmt c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x) Studio : 0.358.0 ```
1.0
The startup query `SELECT @@socket` should be ignored when connecting to a remote database - ## Bug description I have a server that is using the [TDSQL for MySQL](https://intl.cloud.tencent.com/product/dcdb) distribution database provided by Tencent Cloud. The problem is Prisma Client will get a error when startup: ``` 2021-03-23T17:11:16.362175968+08:00 prisma:info Starting a mysql pool with 91 connections. 2021-03-23T17:11:16.373958504+08:00 prisma:info Fetched a connection from the pool 2021-03-23T17:11:16.373999662+08:00 prisma:info Encountered error during initialization: 2021-03-23T17:11:16.38391327+08:00 prisma:error in 0: user_facing_errors::Error::new_non_panic_with_current_backtrace 2021-03-23T17:11:16.383934881+08:00 1: query_engine::error::<impl core::convert::From<query_engine::error::PrismaError> for user_facing_errors::Error>::from 2021-03-23T17:11:16.383941043+08:00 2: query_engine::error::PrismaError::render_as_json 2021-03-23T17:11:16.383946603+08:00 3: query_engine::main::main::{{closure}}::{{closure}} 2021-03-23T17:11:16.383954968+08:00 4: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll 2021-03-23T17:11:16.383959586+08:00 5: std::thread::local::LocalKey<T>::with 2021-03-23T17:11:16.383963003+08:00 6: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll 2021-03-23T17:11:16.383966609+08:00 7: async_io::driver::block_on 2021-03-23T17:11:16.384001264+08:00 8: async_global_executor::reactor::block_on 2021-03-23T17:11:16.384038635+08:00 9: std::thread::local::LocalKey<T>::with 2021-03-23T17:11:16.384044606+08:00 10: async_std::task::builder::Builder::blocking 2021-03-23T17:11:16.384048593+08:00 11: query_engine::main 2021-03-23T17:11:16.384078069+08:00 12: std::sys_common::backtrace::__rust_begin_short_backtrace 2021-03-23T17:11:16.384082347+08:00 13: std::rt::lang_start::{{closure}} 2021-03-23T17:11:16.384126229+08:00 14: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once 2021-03-23T17:11:16.38413206+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/core/src/ops/function.rs:259:13 2021-03-23T17:11:16.384136979+08:00 std::panicking::try::do_call 2021-03-23T17:11:16.384150434+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panicking.rs:379:40 2021-03-23T17:11:16.384155794+08:00 std::panicking::try 2021-03-23T17:11:16.384176193+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panicking.rs:343:19 2021-03-23T17:11:16.38418046+08:00 std::panic::catch_unwind 2021-03-23T17:11:16.384183135+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/panic.rs:396:14 2021-03-23T17:11:16.384185549+08:00 std::rt::lang_start_internal 2021-03-23T17:11:16.384193123+08:00 at /rustc/cb75ad5db02783e8b0222fee363c5f63f7e2cf5b/library/std/src/rt.rs:51:25 2021-03-23T17:11:16.384200237+08:00 15: main 2021-03-23T17:11:16.384203643+08:00 Error querying the database: Server error: `ERROR HY000 (1193): Unknown system variable 'socket'' ``` So I had a check of the Prisma client startup queries, it's executed: ```csv Connect | root@172.17.0.1 on mydb using TCP/IP Query | SELECT @@socket Query | SELECT @@max_allowed_packet Query | SELECT @@wait_timeout ``` Prisma will always select the `@@socket` variable, it's not necessary for the remote database in my opinion. ## How to reproduce 1. Deploy a server on Tencent Cloud with a TDSQL database, with Prisma Client. ## Expected behavior 1. Prisma startup will not execute the `SELECT @@socket`, because TDSQL is not defined. 2. The server will startup correctly. ## Environment & setup - OS: CentOS 8 - Database: TDSQL with MySQL - Node.js version: 14.15.4 - Prisma version: ``` prisma : 2.19.0 @prisma/client : 2.19.0 Current platform : rhel-openssl-1.1.x Query Engine : query-engine c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x) Migration Engine : migration-engine-cli c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x) Introspection Engine : introspection-core c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x) Format Binary : prisma-fmt c1455d0b443d66b0d9db9bcb1bb9ee0d5bbc511d (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x) Studio : 0.358.0 ```
process
the startup query select socket should be ignored when connecting to a remote database bug description i have a server that is using the distribution database provided by tencent cloud the problem is prisma client will get a error when startup prisma info starting a mysql pool with connections prisma info fetched a connection from the pool prisma info encountered error during initialization prisma error in user facing errors error new non panic with current backtrace query engine error for user facing errors error from query engine error prismaerror render as json query engine main main closure closure as core future future future poll std thread local localkey with as core future future future poll async io driver block on async global executor reactor block on std thread local localkey with async std task builder builder blocking query engine main std sys common backtrace rust begin short backtrace std rt lang start closure core ops function impls for f call once at rustc library core src ops function rs std panicking try do call at rustc library std src panicking rs std panicking try at rustc library std src panicking rs std panic catch unwind at rustc library std src panic rs std rt lang start internal at rustc library std src rt rs main error querying the database server error error unknown system variable socket so i had a check of the prisma client startup queries it s executed csv connect root on mydb using tcp ip query select socket query select max allowed packet query select wait timeout prisma will always select the socket variable it s not necessary for the remote database in my opinion how to reproduce deploy a server on tencent cloud with a tdsql database with prisma client expected behavior prisma startup will not execute the select socket because tdsql is not defined the server will startup correctly environment setup os centos database tdsql with mysql node js version prisma version prisma prisma client current platform rhel openssl x query engine query engine at node modules prisma engines query engine rhel openssl x migration engine migration engine cli at node modules prisma engines migration engine rhel openssl x introspection engine introspection core at node modules prisma engines introspection engine rhel openssl x format binary prisma fmt at node modules prisma engines prisma fmt rhel openssl x studio
1
532,796
15,571,577,037
IssuesEvent
2021-03-17 05:17:42
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Error Occurring when Executing 3.2.0 to 4.0.0 MSSQL Database Migration Script
API-M 4.0.0 Priority/Normal Type/Bug migration-4.0.0 migration-4.0.0-docs
### Description: When executing the 3.2.0 to 4.0.0 MSSQL database migration script [1] from the terminal, the following error is encountered. ``` Msg 1779, Level 16, State 1, Server dca5a503261b, Line 28 Table 'AM_API_CLIENT_CERTIFICATE' already has a primary key defined on it. Msg 1750, Level 16, State 1, Server dca5a503261b, Line 28 Could not create constraint or index. See previous errors. ``` [1] https://github.com/wso2-extensions/apim-migration-resources/blob/master/migration-scripts/migration-3.2.0_to_4.0.0/mssql.sql ### Environment details (with versions): - OS: Ubuntu 20.04.2 LTS - Client: Ubuntu Terminal - Database docker image: microsoft/mssql-server-linux 2017-latest
1.0
Error Occurring when Executing 3.2.0 to 4.0.0 MSSQL Database Migration Script - ### Description: When executing the 3.2.0 to 4.0.0 MSSQL database migration script [1] from the terminal, the following error is encountered. ``` Msg 1779, Level 16, State 1, Server dca5a503261b, Line 28 Table 'AM_API_CLIENT_CERTIFICATE' already has a primary key defined on it. Msg 1750, Level 16, State 1, Server dca5a503261b, Line 28 Could not create constraint or index. See previous errors. ``` [1] https://github.com/wso2-extensions/apim-migration-resources/blob/master/migration-scripts/migration-3.2.0_to_4.0.0/mssql.sql ### Environment details (with versions): - OS: Ubuntu 20.04.2 LTS - Client: Ubuntu Terminal - Database docker image: microsoft/mssql-server-linux 2017-latest
non_process
error occurring when executing to mssql database migration script description when executing the to mssql database migration script from the terminal the following error is encountered msg level state server line table am api client certificate already has a primary key defined on it msg level state server line could not create constraint or index see previous errors environment details with versions os ubuntu lts client ubuntu terminal database docker image microsoft mssql server linux latest
0
97,682
4,004,553,399
IssuesEvent
2016-05-12 07:51:19
coreos/rkt
https://api.github.com/repos/coreos/rkt
opened
Function stage1/init/common#appToSystemd needs overwhaul
kind/cleanup priority/P2
The function `appToSystemd` needs an overwhaul, it is ~200 LOC, includes a lot of logic, has many parameters. I suggest to refactor it, maybe in "[functional options](http://dave.cheney.net/2014/10/17/functional-options-for-friendly-apis)" style for better future maintainability.
1.0
Function stage1/init/common#appToSystemd needs overwhaul - The function `appToSystemd` needs an overwhaul, it is ~200 LOC, includes a lot of logic, has many parameters. I suggest to refactor it, maybe in "[functional options](http://dave.cheney.net/2014/10/17/functional-options-for-friendly-apis)" style for better future maintainability.
non_process
function init common apptosystemd needs overwhaul the function apptosystemd needs an overwhaul it is loc includes a lot of logic has many parameters i suggest to refactor it maybe in style for better future maintainability
0
1,675
4,312,780,465
IssuesEvent
2016-07-22 07:41:39
matz-e/lobster
https://api.github.com/repos/matz-e/lobster
opened
Factor plotting and dashboard out into a monitor plugin like ELK
enhancement monitoring processing
We should have the "traditional" plotting have a more modular approach, like ELK currently has. Ideally, I think we should be able to specify what monitoring we want to use like this: ``` config = Config( … monitoring=[Elk(…), Plots(plotdir='…'), Dashboard()], … ) ``` Benefits: * smaller core Lobster database (a lot of the task data can be omitted) * less space consumed when running only with ELK * faster task creating and returning queries? * better logical separation of code? Thoughts?
1.0
Factor plotting and dashboard out into a monitor plugin like ELK - We should have the "traditional" plotting have a more modular approach, like ELK currently has. Ideally, I think we should be able to specify what monitoring we want to use like this: ``` config = Config( … monitoring=[Elk(…), Plots(plotdir='…'), Dashboard()], … ) ``` Benefits: * smaller core Lobster database (a lot of the task data can be omitted) * less space consumed when running only with ELK * faster task creating and returning queries? * better logical separation of code? Thoughts?
process
factor plotting and dashboard out into a monitor plugin like elk we should have the traditional plotting have a more modular approach like elk currently has ideally i think we should be able to specify what monitoring we want to use like this config config … monitoring … benefits smaller core lobster database a lot of the task data can be omitted less space consumed when running only with elk faster task creating and returning queries better logical separation of code thoughts
1
7,369
10,512,605,195
IssuesEvent
2019-09-27 18:20:48
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Open Opps - Applicant Status updates
Apply Process Approved Requirements Ready State Dept.
Who: Student intern applicants What: Status updates Why: to show when an intern has applied or been selected Acceptance Criteria: Applicant status for unpaid internships will display on the landing page as follows: - In progress - The applicant has started the application process and saved but not submitted - Applied - The applicant has submitted their student intern application - Not selected - When the non select email has been sent from ATP, pull the information and mark those applicants as non selected. - Selected - when the selected email has been sent from ATP, pull the information and mark those applicants as selected. These are applicants in the primary column from ATP. - Alternate - when someone is selected in the alternate column from ATP and has gotten the alternate email, they are marked alternate in Open Opps. Completed status will be part of a new ticket - 3526 Related Ticket: - 2915 - Close out process
1.0
Open Opps - Applicant Status updates - Who: Student intern applicants What: Status updates Why: to show when an intern has applied or been selected Acceptance Criteria: Applicant status for unpaid internships will display on the landing page as follows: - In progress - The applicant has started the application process and saved but not submitted - Applied - The applicant has submitted their student intern application - Not selected - When the non select email has been sent from ATP, pull the information and mark those applicants as non selected. - Selected - when the selected email has been sent from ATP, pull the information and mark those applicants as selected. These are applicants in the primary column from ATP. - Alternate - when someone is selected in the alternate column from ATP and has gotten the alternate email, they are marked alternate in Open Opps. Completed status will be part of a new ticket - 3526 Related Ticket: - 2915 - Close out process
process
open opps applicant status updates who student intern applicants what status updates why to show when an intern has applied or been selected acceptance criteria applicant status for unpaid internships will display on the landing page as follows in progress the applicant has started the application process and saved but not submitted applied the applicant has submitted their student intern application not selected when the non select email has been sent from atp pull the information and mark those applicants as non selected selected when the selected email has been sent from atp pull the information and mark those applicants as selected these are applicants in the primary column from atp alternate when someone is selected in the alternate column from atp and has gotten the alternate email they are marked alternate in open opps completed status will be part of a new ticket related ticket close out process
1
387,927
26,744,951,597
IssuesEvent
2023-01-30 15:24:28
h5py/h5py
https://api.github.com/repos/h5py/h5py
closed
h5py was built without ROS3 support, can't use ros3 driver
documentation
Operating System : Windows 10 Python version : 3.10 h5py version : D:\Program Files\Python310>pip uninstall h5py Found existing installation: h5py 3.7.0 发生异常: ValueError h5py was built without ROS3 support, can't use ros3 driver File "D:\hdf5Test.py", line 71, in read_hdf5 f = h5py.File(h5_path, 'r', driver='ros3', aws_region=b'xxxxx', secret_id=b'xxxx', secret_key=b'xxxx')
1.0
h5py was built without ROS3 support, can't use ros3 driver - Operating System : Windows 10 Python version : 3.10 h5py version : D:\Program Files\Python310>pip uninstall h5py Found existing installation: h5py 3.7.0 发生异常: ValueError h5py was built without ROS3 support, can't use ros3 driver File "D:\hdf5Test.py", line 71, in read_hdf5 f = h5py.File(h5_path, 'r', driver='ros3', aws_region=b'xxxxx', secret_id=b'xxxx', secret_key=b'xxxx')
non_process
was built without support can t use driver operating system windows python version version d program files pip uninstall found existing installation 发生异常 valueerror was built without support can t use driver file d py line in read f file path r driver aws region b xxxxx secret id b xxxx secret key b xxxx
0
13,633
16,243,924,285
IssuesEvent
2021-05-07 12:49:24
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Segmentation fault in dataloader after upgrading to pytorch v1.8.0
high priority module: crash module: dataloader module: multiprocessing module: regression triaged
## 🐛 Bug I get a segmentation fault in the dataloader upgrading to pytorch v1.8.0. The worker crash immediately. With pytorch v1.7.1 it works without any issues. Is there anything substantial that changed and needs to be accounted for? ## Expected behavior No segmentation fault. ## Environment ``` Collecting environment information... PyTorch version: 1.8.0 Is debug build: False CUDA used to build PyTorch: 10.2 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.5 LTS (x86_64) GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 Clang version: Could not collect CMake version: version 3.15.5 Python version: 3.9 (64-bit runtime) Is CUDA available: True CUDA runtime version: 9.1.85 GPU models and configuration: GPU 0: GeForce RTX 2080 Ti GPU 1: GeForce RTX 2080 Ti Nvidia driver version: 460.32.03 cuDNN version: /usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.7.6.4 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.5 [pip3] torch==1.8.0 [conda] blas 2.108 mkl conda-forge [conda] blas-devel 3.9.0 8_mkl conda-forge [conda] cudatoolkit 10.2.89 h8f6ccaa_8 conda-forge [conda] libblas 3.9.0 8_mkl conda-forge [conda] libcblas 3.9.0 8_mkl conda-forge [conda] liblapack 3.9.0 8_mkl conda-forge [conda] liblapacke 3.9.0 8_mkl conda-forge [conda] mkl 2020.4 h726a3e6_304 conda-forge [conda] mkl-devel 2020.4 ha770c72_305 conda-forge [conda] mkl-include 2020.4 h726a3e6_304 conda-forge [conda] numpy 1.19.5 py39hdbf815f_1 conda-forge [conda] pytorch 1.8.0 py3.9_cuda10.2_cudnn7.6.5_0 pytorch ``` ## Additional context I tried to debug with gdb: ```bash (decode_dev_cuda) lucas@pc-ries25:~/DeepSMLM$ gdb --args python -m decode.neuralfitter.train.live_engine -p config/template.yaml -i cuda:0 -w 10 --no_log --debug GNU gdb (Ubuntu 8.1.1-0ubuntu1) 8.1.1 Copyright (C) 2018 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Type "show copying" and "show warranty" for details. This GDB was configured as "x86_64-linux-gnu". Type "show configuration" for configuration details. For bug reporting instructions, please see: <http://www.gnu.org/software/gdb/bugs/>. Find the GDB manual and other documentation resources online at: <http://www.gnu.org/software/gdb/documentation/>. For help, type "help". Type "apropos word" to search for commands related to "word"... Reading symbols from python...done. (gdb) run Starting program: /home/lucas/xconda3/envs/decode_dev_cuda/bin/python -m decode.neuralfitter.train.live_engine -p config/template.yaml -i cuda:0 -w 10 --no_log --debug [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". [New Thread 0x7fff5fbd4700 (LWP 19441)] /home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py:127: RuntimeWarning: 'decode.neuralfitter.train.live_engine' found in sys.modules after import of package 'decode.neuralfitter.train', but prior to execution of 'decode.neuralfitter.train.live_engine'; this may result in unpredictable behaviour warn(RuntimeWarning(msg)) [New Thread 0x7fff5e1e9700 (LWP 19443)] [New Thread 0x7fff5d9e8700 (LWP 19444)] [New Thread 0x7fff5d1e7700 (LWP 19445)] [New Thread 0x7fff5c9e6700 (LWP 19446)] [New Thread 0x7fff57fff700 (LWP 19447)] [New Thread 0x7fff577fe700 (LWP 19448)] [New Thread 0x7fff56ffd700 (LWP 19449)] [New Thread 0x7fff567fc700 (LWP 19450)] [New Thread 0x7fff55ffb700 (LWP 19451)] [New Thread 0x7fff557fa700 (LWP 19452)] [New Thread 0x7fff54ff9700 (LWP 19453)] [New Thread 0x7fff547f8700 (LWP 19454)] [New Thread 0x7fff53ff7700 (LWP 19455)] [New Thread 0x7fff537f6700 (LWP 19456)] [New Thread 0x7fff52ff5700 (LWP 19457)] [New Thread 0x7fff527f4700 (LWP 19458)] [New Thread 0x7fff51ff3700 (LWP 19459)] [New Thread 0x7fff517f2700 (LWP 19460)] [New Thread 0x7fff50ff1700 (LWP 19461)] [New Thread 0x7fff3ffff700 (LWP 19462)] [New Thread 0x7fff3f7fe700 (LWP 19463)] [New Thread 0x7fff3effd700 (LWP 19464)] [Thread 0x7fff50ff1700 (LWP 19461) exited] [Thread 0x7fff517f2700 (LWP 19460) exited] [Thread 0x7fff51ff3700 (LWP 19459) exited] [Thread 0x7fff527f4700 (LWP 19458) exited] [Thread 0x7fff52ff5700 (LWP 19457) exited] [Thread 0x7fff537f6700 (LWP 19456) exited] [Thread 0x7fff53ff7700 (LWP 19455) exited] [Thread 0x7fff547f8700 (LWP 19454) exited] [Thread 0x7fff54ff9700 (LWP 19453) exited] [Thread 0x7fff557fa700 (LWP 19452) exited] [Thread 0x7fff55ffb700 (LWP 19451) exited] [Thread 0x7fff567fc700 (LWP 19450) exited] [Thread 0x7fff577fe700 (LWP 19448) exited] [Thread 0x7fff57fff700 (LWP 19447) exited] [Thread 0x7fff5c9e6700 (LWP 19446) exited] [Thread 0x7fff5d1e7700 (LWP 19445) exited] [Thread 0x7fff5d9e8700 (LWP 19444) exited] [Thread 0x7fff5e1e9700 (LWP 19443) exited] [Thread 0x7fff56ffd700 (LWP 19449) exited] [New Thread 0x7fff50ff1700 (LWP 19465)] [New Thread 0x7fff517f2700 (LWP 19466)] Model instantiated. Model initialised as specified in the constructor. [New Thread 0x7fff51ff3700 (LWP 19467)] [New Thread 0x7fff527f4700 (LWP 19468)] [New Thread 0x7fff5d379700 (LWP 19469)] [Thread 0x7fff3f7fe700 (LWP 19463) exited] [Thread 0x7fff3effd700 (LWP 19464) exited] [Thread 0x7fff3ffff700 (LWP 19462) exited] [New Thread 0x7fff5cb77780 (LWP 19470)] [New Thread 0x7fff54bd0800 (LWP 19471)] [New Thread 0x7fff543ce880 (LWP 19472)] Sampled dataset in 0.05s. 20828 emitters on 1025 frames. Sampled dataset in 0.01s. 2501 emitters on 129 frames. [New Thread 0x7fff3effd700 (LWP 19473)] 0%| | 0/15 [00:00<?, ?it/s]ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. [New Thread 0x7fff3f7fe700 (LWP 19493)] [Thread 0x7fff5d379700 (LWP 19469) exited] [Thread 0x7fff527f4700 (LWP 19468) exited] [Thread 0x7fff51ff3700 (LWP 19467) exited] ERROR: Unexpected segmentation fault encountered in worker. [New Thread 0x7fff5d379700 (LWP 19495)] [New Thread 0x7fff527f4700 (LWP 19496)] [New Thread 0x7fff51ff3700 (LWP 19497)] [New Thread 0x7fff3ffff700 (LWP 19498)] [New Thread 0x7fff53bcd700 (LWP 19499)] [New Thread 0x7fff533cc700 (LWP 19500)] [New Thread 0x7fff3e7fc700 (LWP 19501)] [New Thread 0x7fff3dffb700 (LWP 19502)] [New Thread 0x7fff3d7fa700 (LWP 19503)] [New Thread 0x7fff3cff9700 (LWP 19504)] 0%| | 0/15 [00:00<?, ?it/s] Traceback (most recent call last): File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 986, in _try_get_data data = self._data_queue.get(timeout=timeout) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/queue.py", line 180, in get self.not_empty.wait(remaining) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/threading.py", line 316, in wait gotit = waiter.acquire(True, timeout) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/_utils/signal_handling.py", line 66, in handler _error_if_any_worker_fails() RuntimeError: DataLoader worker (pid 19478) is killed by signal: Segmentation fault. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py", line 87, in _run_code exec(code, run_globals) File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train/live_engine.py", line 437, in <module> live_engine_setup(args.param_file, args.device, args.debug, args.no_log, args.num_worker_override, args.log_folder, File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train/live_engine.py", line 175, in live_engine_setup train_loss = decode.neuralfitter.train_val_impl.train( File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train_val_impl.py", line 21, in train for batch_num, (x, y_tar, weight) in enumerate(tqdm_enum): # model input (x), target (yt), weights (w) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/tqdm/std.py", line 1178, in __iter__ for obj in iterable: File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 517, in __next__ data = self._next_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 1182, in _next_data idx, data = self._get_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 1138, in _get_data success, data = self._try_get_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 999, in _try_get_data raise RuntimeError('DataLoader worker (pid(s) {}) exited unexpectedly'.format(pids_str)) from e RuntimeError: DataLoader worker (pid(s) 19474, 19476, 19478, 19480) exited unexpectedly [Thread 0x7fff3effd700 (LWP 19473) exited] [Thread 0x7fff3dffb700 (LWP 19502) exited] [Thread 0x7fff533cc700 (LWP 19500) exited] [Thread 0x7fff3ffff700 (LWP 19498) exited] [Thread 0x7fff527f4700 (LWP 19496) exited] [Thread 0x7fff3cff9700 (LWP 19504) exited] [Thread 0x7fff3d7fa700 (LWP 19503) exited] [Thread 0x7fff53bcd700 (LWP 19499) exited] [Thread 0x7fff51ff3700 (LWP 19497) exited] [Thread 0x7fff5d379700 (LWP 19495) exited] [Thread 0x7fff3e7fc700 (LWP 19501) exited] [Thread 0x7fff54bd0800 (LWP 19471) exited] [Thread 0x7fff543ce880 (LWP 19472) exited] [Thread 0x7fff5cb77780 (LWP 19470) exited] [Thread 0x7fff50ff1700 (LWP 19465) exited] [Thread 0x7fff517f2700 (LWP 19466) exited] [Thread 0x7fff5fbd4700 (LWP 19441) exited] [Thread 0x7ffff7fd1740 (LWP 19427) exited] [Inferior 1 (process 19427) exited with code 01] (gdb) bt No stack. (gdb) ``` cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @SsnL @VitalyFedyunin @ejguan
1.0
Segmentation fault in dataloader after upgrading to pytorch v1.8.0 - ## 🐛 Bug I get a segmentation fault in the dataloader upgrading to pytorch v1.8.0. The worker crash immediately. With pytorch v1.7.1 it works without any issues. Is there anything substantial that changed and needs to be accounted for? ## Expected behavior No segmentation fault. ## Environment ``` Collecting environment information... PyTorch version: 1.8.0 Is debug build: False CUDA used to build PyTorch: 10.2 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.5 LTS (x86_64) GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 Clang version: Could not collect CMake version: version 3.15.5 Python version: 3.9 (64-bit runtime) Is CUDA available: True CUDA runtime version: 9.1.85 GPU models and configuration: GPU 0: GeForce RTX 2080 Ti GPU 1: GeForce RTX 2080 Ti Nvidia driver version: 460.32.03 cuDNN version: /usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.7.6.4 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.5 [pip3] torch==1.8.0 [conda] blas 2.108 mkl conda-forge [conda] blas-devel 3.9.0 8_mkl conda-forge [conda] cudatoolkit 10.2.89 h8f6ccaa_8 conda-forge [conda] libblas 3.9.0 8_mkl conda-forge [conda] libcblas 3.9.0 8_mkl conda-forge [conda] liblapack 3.9.0 8_mkl conda-forge [conda] liblapacke 3.9.0 8_mkl conda-forge [conda] mkl 2020.4 h726a3e6_304 conda-forge [conda] mkl-devel 2020.4 ha770c72_305 conda-forge [conda] mkl-include 2020.4 h726a3e6_304 conda-forge [conda] numpy 1.19.5 py39hdbf815f_1 conda-forge [conda] pytorch 1.8.0 py3.9_cuda10.2_cudnn7.6.5_0 pytorch ``` ## Additional context I tried to debug with gdb: ```bash (decode_dev_cuda) lucas@pc-ries25:~/DeepSMLM$ gdb --args python -m decode.neuralfitter.train.live_engine -p config/template.yaml -i cuda:0 -w 10 --no_log --debug GNU gdb (Ubuntu 8.1.1-0ubuntu1) 8.1.1 Copyright (C) 2018 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Type "show copying" and "show warranty" for details. This GDB was configured as "x86_64-linux-gnu". Type "show configuration" for configuration details. For bug reporting instructions, please see: <http://www.gnu.org/software/gdb/bugs/>. Find the GDB manual and other documentation resources online at: <http://www.gnu.org/software/gdb/documentation/>. For help, type "help". Type "apropos word" to search for commands related to "word"... Reading symbols from python...done. (gdb) run Starting program: /home/lucas/xconda3/envs/decode_dev_cuda/bin/python -m decode.neuralfitter.train.live_engine -p config/template.yaml -i cuda:0 -w 10 --no_log --debug [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". [New Thread 0x7fff5fbd4700 (LWP 19441)] /home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py:127: RuntimeWarning: 'decode.neuralfitter.train.live_engine' found in sys.modules after import of package 'decode.neuralfitter.train', but prior to execution of 'decode.neuralfitter.train.live_engine'; this may result in unpredictable behaviour warn(RuntimeWarning(msg)) [New Thread 0x7fff5e1e9700 (LWP 19443)] [New Thread 0x7fff5d9e8700 (LWP 19444)] [New Thread 0x7fff5d1e7700 (LWP 19445)] [New Thread 0x7fff5c9e6700 (LWP 19446)] [New Thread 0x7fff57fff700 (LWP 19447)] [New Thread 0x7fff577fe700 (LWP 19448)] [New Thread 0x7fff56ffd700 (LWP 19449)] [New Thread 0x7fff567fc700 (LWP 19450)] [New Thread 0x7fff55ffb700 (LWP 19451)] [New Thread 0x7fff557fa700 (LWP 19452)] [New Thread 0x7fff54ff9700 (LWP 19453)] [New Thread 0x7fff547f8700 (LWP 19454)] [New Thread 0x7fff53ff7700 (LWP 19455)] [New Thread 0x7fff537f6700 (LWP 19456)] [New Thread 0x7fff52ff5700 (LWP 19457)] [New Thread 0x7fff527f4700 (LWP 19458)] [New Thread 0x7fff51ff3700 (LWP 19459)] [New Thread 0x7fff517f2700 (LWP 19460)] [New Thread 0x7fff50ff1700 (LWP 19461)] [New Thread 0x7fff3ffff700 (LWP 19462)] [New Thread 0x7fff3f7fe700 (LWP 19463)] [New Thread 0x7fff3effd700 (LWP 19464)] [Thread 0x7fff50ff1700 (LWP 19461) exited] [Thread 0x7fff517f2700 (LWP 19460) exited] [Thread 0x7fff51ff3700 (LWP 19459) exited] [Thread 0x7fff527f4700 (LWP 19458) exited] [Thread 0x7fff52ff5700 (LWP 19457) exited] [Thread 0x7fff537f6700 (LWP 19456) exited] [Thread 0x7fff53ff7700 (LWP 19455) exited] [Thread 0x7fff547f8700 (LWP 19454) exited] [Thread 0x7fff54ff9700 (LWP 19453) exited] [Thread 0x7fff557fa700 (LWP 19452) exited] [Thread 0x7fff55ffb700 (LWP 19451) exited] [Thread 0x7fff567fc700 (LWP 19450) exited] [Thread 0x7fff577fe700 (LWP 19448) exited] [Thread 0x7fff57fff700 (LWP 19447) exited] [Thread 0x7fff5c9e6700 (LWP 19446) exited] [Thread 0x7fff5d1e7700 (LWP 19445) exited] [Thread 0x7fff5d9e8700 (LWP 19444) exited] [Thread 0x7fff5e1e9700 (LWP 19443) exited] [Thread 0x7fff56ffd700 (LWP 19449) exited] [New Thread 0x7fff50ff1700 (LWP 19465)] [New Thread 0x7fff517f2700 (LWP 19466)] Model instantiated. Model initialised as specified in the constructor. [New Thread 0x7fff51ff3700 (LWP 19467)] [New Thread 0x7fff527f4700 (LWP 19468)] [New Thread 0x7fff5d379700 (LWP 19469)] [Thread 0x7fff3f7fe700 (LWP 19463) exited] [Thread 0x7fff3effd700 (LWP 19464) exited] [Thread 0x7fff3ffff700 (LWP 19462) exited] [New Thread 0x7fff5cb77780 (LWP 19470)] [New Thread 0x7fff54bd0800 (LWP 19471)] [New Thread 0x7fff543ce880 (LWP 19472)] Sampled dataset in 0.05s. 20828 emitters on 1025 frames. Sampled dataset in 0.01s. 2501 emitters on 129 frames. [New Thread 0x7fff3effd700 (LWP 19473)] 0%| | 0/15 [00:00<?, ?it/s]ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. ERROR: Unexpected segmentation fault encountered in worker. [New Thread 0x7fff3f7fe700 (LWP 19493)] [Thread 0x7fff5d379700 (LWP 19469) exited] [Thread 0x7fff527f4700 (LWP 19468) exited] [Thread 0x7fff51ff3700 (LWP 19467) exited] ERROR: Unexpected segmentation fault encountered in worker. [New Thread 0x7fff5d379700 (LWP 19495)] [New Thread 0x7fff527f4700 (LWP 19496)] [New Thread 0x7fff51ff3700 (LWP 19497)] [New Thread 0x7fff3ffff700 (LWP 19498)] [New Thread 0x7fff53bcd700 (LWP 19499)] [New Thread 0x7fff533cc700 (LWP 19500)] [New Thread 0x7fff3e7fc700 (LWP 19501)] [New Thread 0x7fff3dffb700 (LWP 19502)] [New Thread 0x7fff3d7fa700 (LWP 19503)] [New Thread 0x7fff3cff9700 (LWP 19504)] 0%| | 0/15 [00:00<?, ?it/s] Traceback (most recent call last): File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 986, in _try_get_data data = self._data_queue.get(timeout=timeout) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/queue.py", line 180, in get self.not_empty.wait(remaining) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/threading.py", line 316, in wait gotit = waiter.acquire(True, timeout) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/_utils/signal_handling.py", line 66, in handler _error_if_any_worker_fails() RuntimeError: DataLoader worker (pid 19478) is killed by signal: Segmentation fault. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/runpy.py", line 87, in _run_code exec(code, run_globals) File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train/live_engine.py", line 437, in <module> live_engine_setup(args.param_file, args.device, args.debug, args.no_log, args.num_worker_override, args.log_folder, File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train/live_engine.py", line 175, in live_engine_setup train_loss = decode.neuralfitter.train_val_impl.train( File "/home/lucas/RemoteDeploy/DeepSMLM/decode/neuralfitter/train_val_impl.py", line 21, in train for batch_num, (x, y_tar, weight) in enumerate(tqdm_enum): # model input (x), target (yt), weights (w) File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/tqdm/std.py", line 1178, in __iter__ for obj in iterable: File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 517, in __next__ data = self._next_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 1182, in _next_data idx, data = self._get_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 1138, in _get_data success, data = self._try_get_data() File "/home/lucas/xconda3/envs/decode_dev_cuda/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 999, in _try_get_data raise RuntimeError('DataLoader worker (pid(s) {}) exited unexpectedly'.format(pids_str)) from e RuntimeError: DataLoader worker (pid(s) 19474, 19476, 19478, 19480) exited unexpectedly [Thread 0x7fff3effd700 (LWP 19473) exited] [Thread 0x7fff3dffb700 (LWP 19502) exited] [Thread 0x7fff533cc700 (LWP 19500) exited] [Thread 0x7fff3ffff700 (LWP 19498) exited] [Thread 0x7fff527f4700 (LWP 19496) exited] [Thread 0x7fff3cff9700 (LWP 19504) exited] [Thread 0x7fff3d7fa700 (LWP 19503) exited] [Thread 0x7fff53bcd700 (LWP 19499) exited] [Thread 0x7fff51ff3700 (LWP 19497) exited] [Thread 0x7fff5d379700 (LWP 19495) exited] [Thread 0x7fff3e7fc700 (LWP 19501) exited] [Thread 0x7fff54bd0800 (LWP 19471) exited] [Thread 0x7fff543ce880 (LWP 19472) exited] [Thread 0x7fff5cb77780 (LWP 19470) exited] [Thread 0x7fff50ff1700 (LWP 19465) exited] [Thread 0x7fff517f2700 (LWP 19466) exited] [Thread 0x7fff5fbd4700 (LWP 19441) exited] [Thread 0x7ffff7fd1740 (LWP 19427) exited] [Inferior 1 (process 19427) exited with code 01] (gdb) bt No stack. (gdb) ``` cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @SsnL @VitalyFedyunin @ejguan
process
segmentation fault in dataloader after upgrading to pytorch 🐛 bug i get a segmentation fault in the dataloader upgrading to pytorch the worker crash immediately with pytorch it works without any issues is there anything substantial that changed and needs to be accounted for expected behavior no segmentation fault environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version python version bit runtime is cuda available true cuda runtime version gpu models and configuration gpu geforce rtx ti gpu geforce rtx ti nvidia driver version cudnn version usr local cuda targets linux lib libcudnn so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy torch blas mkl conda forge blas devel mkl conda forge cudatoolkit conda forge libblas mkl conda forge libcblas mkl conda forge liblapack mkl conda forge liblapacke mkl conda forge mkl conda forge mkl devel conda forge mkl include conda forge numpy conda forge pytorch pytorch additional context i tried to debug with gdb bash decode dev cuda lucas pc deepsmlm gdb args python m decode neuralfitter train live engine p config template yaml i cuda w no log debug gnu gdb ubuntu copyright c free software foundation inc license gnu gpl version or later this is free software you are free to change and redistribute it there is no warranty to the extent permitted by law type show copying and show warranty for details this gdb was configured as linux gnu type show configuration for configuration details for bug reporting instructions please see find the gdb manual and other documentation resources online at for help type help type apropos word to search for commands related to word reading symbols from python done gdb run starting program home lucas envs decode dev cuda bin python m decode neuralfitter train live engine p config template yaml i cuda w no log debug using host libthread db library lib linux gnu libthread db so home lucas envs decode dev cuda lib runpy py runtimewarning decode neuralfitter train live engine found in sys modules after import of package decode neuralfitter train but prior to execution of decode neuralfitter train live engine this may result in unpredictable behaviour warn runtimewarning msg model instantiated model initialised as specified in the constructor sampled dataset in emitters on frames sampled dataset in emitters on frames error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker error unexpected segmentation fault encountered in worker traceback most recent call last file home lucas envs decode dev cuda lib site packages torch utils data dataloader py line in try get data data self data queue get timeout timeout file home lucas envs decode dev cuda lib queue py line in get self not empty wait remaining file home lucas envs decode dev cuda lib threading py line in wait gotit waiter acquire true timeout file home lucas envs decode dev cuda lib site packages torch utils data utils signal handling py line in handler error if any worker fails runtimeerror dataloader worker pid is killed by signal segmentation fault the above exception was the direct cause of the following exception traceback most recent call last file home lucas envs decode dev cuda lib runpy py line in run module as main return run code code main globals none file home lucas envs decode dev cuda lib runpy py line in run code exec code run globals file home lucas remotedeploy deepsmlm decode neuralfitter train live engine py line in live engine setup args param file args device args debug args no log args num worker override args log folder file home lucas remotedeploy deepsmlm decode neuralfitter train live engine py line in live engine setup train loss decode neuralfitter train val impl train file home lucas remotedeploy deepsmlm decode neuralfitter train val impl py line in train for batch num x y tar weight in enumerate tqdm enum model input x target yt weights w file home lucas envs decode dev cuda lib site packages tqdm std py line in iter for obj in iterable file home lucas envs decode dev cuda lib site packages torch utils data dataloader py line in next data self next data file home lucas envs decode dev cuda lib site packages torch utils data dataloader py line in next data idx data self get data file home lucas envs decode dev cuda lib site packages torch utils data dataloader py line in get data success data self try get data file home lucas envs decode dev cuda lib site packages torch utils data dataloader py line in try get data raise runtimeerror dataloader worker pid s exited unexpectedly format pids str from e runtimeerror dataloader worker pid s exited unexpectedly gdb bt no stack gdb cc ezyang gchanan bdhirsh jbschlosser ssnl vitalyfedyunin ejguan
1
43,209
11,569,199,051
IssuesEvent
2020-02-20 17:07:23
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
"Other" heating subcategory appears (incorrectly)
Defect
Issue overview -------------- See IDF below where >12 GJ of energy appears in an "Other" heating subcategory in the End Uses By Subcategory output table. The subcategory is being automatically created. Across several hundred test IDFs, this is the _only_ file where this subcategory appears. In every other test file, heating energy correctly shows up in the "General" subcategory. (This particular test file includes a complex set of HVAC systems serving a single conditioned zone.) [in.idf.txt](https://github.com/NREL/EnergyPlus/files/3725469/in.idf.txt) ### Details Some additional details for this issue (if relevant): - Windows 10 - Version 9.2.0-921312fa1d ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
"Other" heating subcategory appears (incorrectly) - Issue overview -------------- See IDF below where >12 GJ of energy appears in an "Other" heating subcategory in the End Uses By Subcategory output table. The subcategory is being automatically created. Across several hundred test IDFs, this is the _only_ file where this subcategory appears. In every other test file, heating energy correctly shows up in the "General" subcategory. (This particular test file includes a complex set of HVAC systems serving a single conditioned zone.) [in.idf.txt](https://github.com/NREL/EnergyPlus/files/3725469/in.idf.txt) ### Details Some additional details for this issue (if relevant): - Windows 10 - Version 9.2.0-921312fa1d ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
non_process
other heating subcategory appears incorrectly issue overview see idf below where gj of energy appears in an other heating subcategory in the end uses by subcategory output table the subcategory is being automatically created across several hundred test idfs this is the only file where this subcategory appears in every other test file heating energy correctly shows up in the general subcategory this particular test file includes a complex set of hvac systems serving a single conditioned zone details some additional details for this issue if relevant windows version checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
0
4,433
7,308,529,931
IssuesEvent
2018-02-28 08:42:48
UKHomeOffice/dq-aws-transition
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
closed
Configure Maytech Mock Connectivity from NotProd Ingest Linux Server
DQ Data Ingest DQ Tranche 1 Production SSM processing
Private Key Migration - [x] Private Key for Linux Ingest NotProd (/home/SSM/.ssh/id_rsa) for Mock Maytech Server Configure NotProd ssh_remote_* parameters in sftp_oag_client_maytech.py - [x] ssh_remote_host: <mock SFTP server> - [x] ssh_remote_user: <mock oag user> - [x] ssh_remote_key: /home/SSM/.ssh/id_rsa
1.0
Configure Maytech Mock Connectivity from NotProd Ingest Linux Server - Private Key Migration - [x] Private Key for Linux Ingest NotProd (/home/SSM/.ssh/id_rsa) for Mock Maytech Server Configure NotProd ssh_remote_* parameters in sftp_oag_client_maytech.py - [x] ssh_remote_host: <mock SFTP server> - [x] ssh_remote_user: <mock oag user> - [x] ssh_remote_key: /home/SSM/.ssh/id_rsa
process
configure maytech mock connectivity from notprod ingest linux server private key migration private key for linux ingest notprod home ssm ssh id rsa for mock maytech server configure notprod ssh remote parameters in sftp oag client maytech py ssh remote host ssh remote user ssh remote key home ssm ssh id rsa
1
20,251
26,869,036,918
IssuesEvent
2023-02-04 07:54:10
threefoldtech/builders
https://api.github.com/repos/threefoldtech/builders
closed
docker install script fails on a grid full vm
type_bug process_wontfix
```sh root@VM376afe27:/code/threefoldtech/builders/scripts/installers# ./docker.sh Reading package lists... Done Building dependency tree... Done Reading state information... Done E: Unable to locate package docker-engine ++ sudo apt-get update Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease Hit:2 http://security.ubuntu.com/ubuntu jammy-security InRelease Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease Reading package lists... Done ++ sudo apt-get install ca-certificates curl gnupg lsb-release -y Reading package lists... Done Building dependency tree... Done Reading state information... Done lsb-release is already the newest version (11.1.0ubuntu4). lsb-release set to manually installed. ca-certificates is already the newest version (20211016ubuntu0.22.04.1). ca-certificates set to manually installed. curl is already the newest version (7.81.0-1ubuntu1.7). curl set to manually installed. gnupg is already the newest version (2.2.27-3ubuntu2.1). gnupg set to manually installed. 0 upgraded, 0 newly installed, 0 to remove and 19 not upgraded. ++ mkdir -p /etc/apt/keyrings ++ curl -fsSL https://download.docker.com/linux/ubuntu/gpg ++ sudo gpg --dearmor -o /etc/apt/keyrings/docker.gpg ++ sudo tee /etc/apt/sources.list.d/docker.list +++ dpkg --print-architecture +++ lsb_release -cs ++ echo 'deb [arch=amd64 signed-by=/etc/apt/keyrings/docker.gpg] https://download.docker.com/linux/ubuntu jammy stable' ++ chmod a+r /etc/apt/keyrings/docker.gpg ++ apt-get update Get:1 https://download.docker.com/linux/ubuntu jammy InRelease [48.9 kB] Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease Get:3 https://download.docker.com/linux/ubuntu jammy/stable amd64 Packages [12.7 kB] Hit:4 http://archive.ubuntu.com/ubuntu jammy-updates InRelease Get:5 http://security.ubuntu.com/ubuntu jammy-security InRelease [110 kB] Hit:6 http://archive.ubuntu.com/ubuntu jammy-backports InRelease Fetched 172 kB in 1s (127 kB/s) Reading package lists... Done ++ apt-get install docker-ce docker-ce-cli containerd.io docker-compose-plugin -y Reading package lists... Done Building dependency tree... Done Reading state information... Done The following packages were automatically installed and are no longer required: bridge-utils dns-root-data dnsmasq-base ubuntu-fan Use 'apt autoremove' to remove them. The following additional packages will be installed: docker-buildx-plugin docker-ce-rootless-extras docker-scan-plugin libltdl7 libslirp0 slirp4netns Suggested packages: aufs-tools cgroupfs-mount | cgroup-lite The following packages will be REMOVED: containerd docker.io runc The following NEW packages will be installed: containerd.io docker-buildx-plugin docker-ce docker-ce-cli docker-ce-rootless-extras docker-compose-plugin docker-scan-plugin libltdl7 libslirp0 slirp4netns 0 upgraded, 10 newly installed, 3 to remove and 19 not upgraded. Need to get 111 MB of archives. After this operation, 111 MB of additional disk space will be used. Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 libltdl7 amd64 2.4.6-15build2 [39.6 kB] Get:2 http://archive.ubuntu.com/ubuntu jammy/main amd64 libslirp0 amd64 4.6.1-1build1 [61.5 kB] Get:3 https://download.docker.com/linux/ubuntu jammy/stable amd64 containerd.io amd64 1.6.16-1 [27.7 MB] Get:4 http://archive.ubuntu.com/ubuntu jammy/universe amd64 slirp4netns amd64 1.0.1-2 [28.2 kB] Get:5 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-buildx-plugin amd64 0.10.2-1~ubuntu.22.04~jammy [25.9 MB] Get:6 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce-cli amd64 5:23.0.0-1~ubuntu.22.04~jammy [13.2 MB] Get:7 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce amd64 5:23.0.0-1~ubuntu.22.04~jammy [22.0 MB] Get:8 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce-rootless-extras amd64 5:23.0.0-1~ubuntu.22.04~jammy [8,759 kB] Get:9 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-compose-plugin amd64 2.15.1-1~ubuntu.22.04~jammy [9,570 kB] Get:10 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-scan-plugin amd64 0.23.0~ubuntu-jammy [3,623 kB] Fetched 111 MB in 3s (36.2 MB/s) (Reading database ... 79412 files and directories currently installed.) Removing docker.io (20.10.12-0ubuntu4) ... '/usr/share/docker.io/contrib/nuke-graph-directory.sh' -> '/var/lib/docker/nuke-graph-directory.sh' Warning: Stopping docker.service, but it can still be activated by: docker.socket Removing containerd (1.5.9-0ubuntu3.1) ... Removing runc (1.1.0-0ubuntu1.1) ... Selecting previously unselected package containerd.io. (Reading database ... 79150 files and directories currently installed.) Preparing to unpack .../0-containerd.io_1.6.16-1_amd64.deb ... Unpacking containerd.io (1.6.16-1) ... Selecting previously unselected package docker-buildx-plugin. Preparing to unpack .../1-docker-buildx-plugin_0.10.2-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-buildx-plugin (0.10.2-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce-cli. Preparing to unpack .../2-docker-ce-cli_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce-cli (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce. Preparing to unpack .../3-docker-ce_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce-rootless-extras. Preparing to unpack .../4-docker-ce-rootless-extras_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce-rootless-extras (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-compose-plugin. Preparing to unpack .../5-docker-compose-plugin_2.15.1-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-compose-plugin (2.15.1-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-scan-plugin. Preparing to unpack .../6-docker-scan-plugin_0.23.0~ubuntu-jammy_amd64.deb ... Unpacking docker-scan-plugin (0.23.0~ubuntu-jammy) ... Selecting previously unselected package libltdl7:amd64. Preparing to unpack .../7-libltdl7_2.4.6-15build2_amd64.deb ... Unpacking libltdl7:amd64 (2.4.6-15build2) ... Selecting previously unselected package libslirp0:amd64. Preparing to unpack .../8-libslirp0_4.6.1-1build1_amd64.deb ... Unpacking libslirp0:amd64 (4.6.1-1build1) ... Selecting previously unselected package slirp4netns. Preparing to unpack .../9-slirp4netns_1.0.1-2_amd64.deb ... Unpacking slirp4netns (1.0.1-2) ... Setting up docker-scan-plugin (0.23.0~ubuntu-jammy) ... Setting up docker-buildx-plugin (0.10.2-1~ubuntu.22.04~jammy) ... Setting up containerd.io (1.6.16-1) ... Setting up docker-compose-plugin (2.15.1-1~ubuntu.22.04~jammy) ... Setting up libltdl7:amd64 (2.4.6-15build2) ... Setting up docker-ce-cli (5:23.0.0-1~ubuntu.22.04~jammy) ... Setting up libslirp0:amd64 (4.6.1-1build1) ... Setting up docker-ce-rootless-extras (5:23.0.0-1~ubuntu.22.04~jammy) ... Setting up slirp4netns (1.0.1-2) ... Setting up docker-ce (5:23.0.0-1~ubuntu.22.04~jammy) ... Job for docker.service failed because the control process exited with error code. See "systemctl status docker.service" and "journalctl -xeu docker.service" for details. invoke-rc.d: initscript docker, action "start" failed. ● docker.service - Docker Application Container Engine Loaded: loaded (/lib/systemd/system/docker.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: exit-code) since Fri 2023-02-03 11:02:55 UTC; 17ms ago TriggeredBy: ● docker.socket Docs: https://docs.docker.com Process: 906632 ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock (code=exited, status=1/FAILURE) Main PID: 906632 (code=exited, status=1/FAILURE) CPU: 215ms dpkg: error processing package docker-ce (--configure): installed docker-ce package post-installation script subprocess returned error exit status 1 Processing triggers for man-db (2.10.2-1) ... Processing triggers for libc-bin (2.35-0ubuntu3.1) ... Errors were encountered while processing: docker-ce needrestart is being skipped since dpkg has failed E: Sub-process /usr/bin/dpkg returned an error code (1) ```
1.0
docker install script fails on a grid full vm - ```sh root@VM376afe27:/code/threefoldtech/builders/scripts/installers# ./docker.sh Reading package lists... Done Building dependency tree... Done Reading state information... Done E: Unable to locate package docker-engine ++ sudo apt-get update Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease Hit:2 http://security.ubuntu.com/ubuntu jammy-security InRelease Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease Reading package lists... Done ++ sudo apt-get install ca-certificates curl gnupg lsb-release -y Reading package lists... Done Building dependency tree... Done Reading state information... Done lsb-release is already the newest version (11.1.0ubuntu4). lsb-release set to manually installed. ca-certificates is already the newest version (20211016ubuntu0.22.04.1). ca-certificates set to manually installed. curl is already the newest version (7.81.0-1ubuntu1.7). curl set to manually installed. gnupg is already the newest version (2.2.27-3ubuntu2.1). gnupg set to manually installed. 0 upgraded, 0 newly installed, 0 to remove and 19 not upgraded. ++ mkdir -p /etc/apt/keyrings ++ curl -fsSL https://download.docker.com/linux/ubuntu/gpg ++ sudo gpg --dearmor -o /etc/apt/keyrings/docker.gpg ++ sudo tee /etc/apt/sources.list.d/docker.list +++ dpkg --print-architecture +++ lsb_release -cs ++ echo 'deb [arch=amd64 signed-by=/etc/apt/keyrings/docker.gpg] https://download.docker.com/linux/ubuntu jammy stable' ++ chmod a+r /etc/apt/keyrings/docker.gpg ++ apt-get update Get:1 https://download.docker.com/linux/ubuntu jammy InRelease [48.9 kB] Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease Get:3 https://download.docker.com/linux/ubuntu jammy/stable amd64 Packages [12.7 kB] Hit:4 http://archive.ubuntu.com/ubuntu jammy-updates InRelease Get:5 http://security.ubuntu.com/ubuntu jammy-security InRelease [110 kB] Hit:6 http://archive.ubuntu.com/ubuntu jammy-backports InRelease Fetched 172 kB in 1s (127 kB/s) Reading package lists... Done ++ apt-get install docker-ce docker-ce-cli containerd.io docker-compose-plugin -y Reading package lists... Done Building dependency tree... Done Reading state information... Done The following packages were automatically installed and are no longer required: bridge-utils dns-root-data dnsmasq-base ubuntu-fan Use 'apt autoremove' to remove them. The following additional packages will be installed: docker-buildx-plugin docker-ce-rootless-extras docker-scan-plugin libltdl7 libslirp0 slirp4netns Suggested packages: aufs-tools cgroupfs-mount | cgroup-lite The following packages will be REMOVED: containerd docker.io runc The following NEW packages will be installed: containerd.io docker-buildx-plugin docker-ce docker-ce-cli docker-ce-rootless-extras docker-compose-plugin docker-scan-plugin libltdl7 libslirp0 slirp4netns 0 upgraded, 10 newly installed, 3 to remove and 19 not upgraded. Need to get 111 MB of archives. After this operation, 111 MB of additional disk space will be used. Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 libltdl7 amd64 2.4.6-15build2 [39.6 kB] Get:2 http://archive.ubuntu.com/ubuntu jammy/main amd64 libslirp0 amd64 4.6.1-1build1 [61.5 kB] Get:3 https://download.docker.com/linux/ubuntu jammy/stable amd64 containerd.io amd64 1.6.16-1 [27.7 MB] Get:4 http://archive.ubuntu.com/ubuntu jammy/universe amd64 slirp4netns amd64 1.0.1-2 [28.2 kB] Get:5 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-buildx-plugin amd64 0.10.2-1~ubuntu.22.04~jammy [25.9 MB] Get:6 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce-cli amd64 5:23.0.0-1~ubuntu.22.04~jammy [13.2 MB] Get:7 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce amd64 5:23.0.0-1~ubuntu.22.04~jammy [22.0 MB] Get:8 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-ce-rootless-extras amd64 5:23.0.0-1~ubuntu.22.04~jammy [8,759 kB] Get:9 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-compose-plugin amd64 2.15.1-1~ubuntu.22.04~jammy [9,570 kB] Get:10 https://download.docker.com/linux/ubuntu jammy/stable amd64 docker-scan-plugin amd64 0.23.0~ubuntu-jammy [3,623 kB] Fetched 111 MB in 3s (36.2 MB/s) (Reading database ... 79412 files and directories currently installed.) Removing docker.io (20.10.12-0ubuntu4) ... '/usr/share/docker.io/contrib/nuke-graph-directory.sh' -> '/var/lib/docker/nuke-graph-directory.sh' Warning: Stopping docker.service, but it can still be activated by: docker.socket Removing containerd (1.5.9-0ubuntu3.1) ... Removing runc (1.1.0-0ubuntu1.1) ... Selecting previously unselected package containerd.io. (Reading database ... 79150 files and directories currently installed.) Preparing to unpack .../0-containerd.io_1.6.16-1_amd64.deb ... Unpacking containerd.io (1.6.16-1) ... Selecting previously unselected package docker-buildx-plugin. Preparing to unpack .../1-docker-buildx-plugin_0.10.2-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-buildx-plugin (0.10.2-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce-cli. Preparing to unpack .../2-docker-ce-cli_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce-cli (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce. Preparing to unpack .../3-docker-ce_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-ce-rootless-extras. Preparing to unpack .../4-docker-ce-rootless-extras_5%3a23.0.0-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-ce-rootless-extras (5:23.0.0-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-compose-plugin. Preparing to unpack .../5-docker-compose-plugin_2.15.1-1~ubuntu.22.04~jammy_amd64.deb ... Unpacking docker-compose-plugin (2.15.1-1~ubuntu.22.04~jammy) ... Selecting previously unselected package docker-scan-plugin. Preparing to unpack .../6-docker-scan-plugin_0.23.0~ubuntu-jammy_amd64.deb ... Unpacking docker-scan-plugin (0.23.0~ubuntu-jammy) ... Selecting previously unselected package libltdl7:amd64. Preparing to unpack .../7-libltdl7_2.4.6-15build2_amd64.deb ... Unpacking libltdl7:amd64 (2.4.6-15build2) ... Selecting previously unselected package libslirp0:amd64. Preparing to unpack .../8-libslirp0_4.6.1-1build1_amd64.deb ... Unpacking libslirp0:amd64 (4.6.1-1build1) ... Selecting previously unselected package slirp4netns. Preparing to unpack .../9-slirp4netns_1.0.1-2_amd64.deb ... Unpacking slirp4netns (1.0.1-2) ... Setting up docker-scan-plugin (0.23.0~ubuntu-jammy) ... Setting up docker-buildx-plugin (0.10.2-1~ubuntu.22.04~jammy) ... Setting up containerd.io (1.6.16-1) ... Setting up docker-compose-plugin (2.15.1-1~ubuntu.22.04~jammy) ... Setting up libltdl7:amd64 (2.4.6-15build2) ... Setting up docker-ce-cli (5:23.0.0-1~ubuntu.22.04~jammy) ... Setting up libslirp0:amd64 (4.6.1-1build1) ... Setting up docker-ce-rootless-extras (5:23.0.0-1~ubuntu.22.04~jammy) ... Setting up slirp4netns (1.0.1-2) ... Setting up docker-ce (5:23.0.0-1~ubuntu.22.04~jammy) ... Job for docker.service failed because the control process exited with error code. See "systemctl status docker.service" and "journalctl -xeu docker.service" for details. invoke-rc.d: initscript docker, action "start" failed. ● docker.service - Docker Application Container Engine Loaded: loaded (/lib/systemd/system/docker.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: exit-code) since Fri 2023-02-03 11:02:55 UTC; 17ms ago TriggeredBy: ● docker.socket Docs: https://docs.docker.com Process: 906632 ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock (code=exited, status=1/FAILURE) Main PID: 906632 (code=exited, status=1/FAILURE) CPU: 215ms dpkg: error processing package docker-ce (--configure): installed docker-ce package post-installation script subprocess returned error exit status 1 Processing triggers for man-db (2.10.2-1) ... Processing triggers for libc-bin (2.35-0ubuntu3.1) ... Errors were encountered while processing: docker-ce needrestart is being skipped since dpkg has failed E: Sub-process /usr/bin/dpkg returned an error code (1) ```
process
docker install script fails on a grid full vm sh root code threefoldtech builders scripts installers docker sh reading package lists done building dependency tree done reading state information done e unable to locate package docker engine sudo apt get update hit jammy inrelease hit jammy security inrelease hit jammy updates inrelease hit jammy backports inrelease reading package lists done sudo apt get install ca certificates curl gnupg lsb release y reading package lists done building dependency tree done reading state information done lsb release is already the newest version lsb release set to manually installed ca certificates is already the newest version ca certificates set to manually installed curl is already the newest version curl set to manually installed gnupg is already the newest version gnupg set to manually installed upgraded newly installed to remove and not upgraded mkdir p etc apt keyrings curl fssl sudo gpg dearmor o etc apt keyrings docker gpg sudo tee etc apt sources list d docker list dpkg print architecture lsb release cs echo deb jammy stable chmod a r etc apt keyrings docker gpg apt get update get jammy inrelease hit jammy inrelease get jammy stable packages hit jammy updates inrelease get jammy security inrelease hit jammy backports inrelease fetched kb in kb s reading package lists done apt get install docker ce docker ce cli containerd io docker compose plugin y reading package lists done building dependency tree done reading state information done the following packages were automatically installed and are no longer required bridge utils dns root data dnsmasq base ubuntu fan use apt autoremove to remove them the following additional packages will be installed docker buildx plugin docker ce rootless extras docker scan plugin suggested packages aufs tools cgroupfs mount cgroup lite the following packages will be removed containerd docker io runc the following new packages will be installed containerd io docker buildx plugin docker ce docker ce cli docker ce rootless extras docker compose plugin docker scan plugin upgraded newly installed to remove and not upgraded need to get mb of archives after this operation mb of additional disk space will be used get jammy main get jammy main get jammy stable containerd io get jammy universe get jammy stable docker buildx plugin ubuntu jammy get jammy stable docker ce cli ubuntu jammy get jammy stable docker ce ubuntu jammy get jammy stable docker ce rootless extras ubuntu jammy get jammy stable docker compose plugin ubuntu jammy get jammy stable docker scan plugin ubuntu jammy fetched mb in mb s reading database files and directories currently installed removing docker io usr share docker io contrib nuke graph directory sh var lib docker nuke graph directory sh warning stopping docker service but it can still be activated by docker socket removing containerd removing runc selecting previously unselected package containerd io reading database files and directories currently installed preparing to unpack containerd io deb unpacking containerd io selecting previously unselected package docker buildx plugin preparing to unpack docker buildx plugin ubuntu jammy deb unpacking docker buildx plugin ubuntu jammy selecting previously unselected package docker ce cli preparing to unpack docker ce cli ubuntu jammy deb unpacking docker ce cli ubuntu jammy selecting previously unselected package docker ce preparing to unpack docker ce ubuntu jammy deb unpacking docker ce ubuntu jammy selecting previously unselected package docker ce rootless extras preparing to unpack docker ce rootless extras ubuntu jammy deb unpacking docker ce rootless extras ubuntu jammy selecting previously unselected package docker compose plugin preparing to unpack docker compose plugin ubuntu jammy deb unpacking docker compose plugin ubuntu jammy selecting previously unselected package docker scan plugin preparing to unpack docker scan plugin ubuntu jammy deb unpacking docker scan plugin ubuntu jammy selecting previously unselected package preparing to unpack deb unpacking selecting previously unselected package preparing to unpack deb unpacking selecting previously unselected package preparing to unpack deb unpacking setting up docker scan plugin ubuntu jammy setting up docker buildx plugin ubuntu jammy setting up containerd io setting up docker compose plugin ubuntu jammy setting up setting up docker ce cli ubuntu jammy setting up setting up docker ce rootless extras ubuntu jammy setting up setting up docker ce ubuntu jammy job for docker service failed because the control process exited with error code see systemctl status docker service and journalctl xeu docker service for details invoke rc d initscript docker action start failed ● docker service docker application container engine loaded loaded lib systemd system docker service enabled vendor preset enabled active activating auto restart result exit code since fri utc ago triggeredby ● docker socket docs process execstart usr bin dockerd h fd containerd run containerd containerd sock code exited status failure main pid code exited status failure cpu dpkg error processing package docker ce configure installed docker ce package post installation script subprocess returned error exit status processing triggers for man db processing triggers for libc bin errors were encountered while processing docker ce needrestart is being skipped since dpkg has failed e sub process usr bin dpkg returned an error code
1
46,857
2,965,311,737
IssuesEvent
2015-07-10 21:58:13
riot/riot
https://api.github.com/repos/riot/riot
closed
Riot is not usable from inside web workers
low priority question to verify
I've discovered that riot can't be used from inside web workers due to init logic on the window object. It seems to be solved by changing the initializing closure call to use self instead of undefined, like this: })(typeof window != 'undefined' ? window : self); But I'm not sure if that breaks something else. The web worker's global scope is always called "self", as far as I'm aware.
1.0
Riot is not usable from inside web workers - I've discovered that riot can't be used from inside web workers due to init logic on the window object. It seems to be solved by changing the initializing closure call to use self instead of undefined, like this: })(typeof window != 'undefined' ? window : self); But I'm not sure if that breaks something else. The web worker's global scope is always called "self", as far as I'm aware.
non_process
riot is not usable from inside web workers i ve discovered that riot can t be used from inside web workers due to init logic on the window object it seems to be solved by changing the initializing closure call to use self instead of undefined like this typeof window undefined window self but i m not sure if that breaks something else the web worker s global scope is always called self as far as i m aware
0
5,972
8,792,467,319
IssuesEvent
2018-12-21 16:12:42
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
EMails cannot be imported from imap
mail processing waiting for feedback
Infos: Used Zammad version: 2.8 Installation method (source, package, ..): package Operating system: Ubuntu Server 16.04 LTS Database + version: 10.0.36-MariaDB-0ubuntu0.16.04.1 Ubuntu 16.04 Elasticsearch version: 5.6.9 Browser + version: any Error: Channel: Email::Account in Can't use Channel::Driver::Imap: #<Net::IMAP::ResponseParse Log: [2018-12-21T13:33:13.719440 #1027-70334085745740] ERROR -- : Can't use Channel::Driver::Imap: #<Net::IMAP::ResponseParseError: unexpected token LPAR (expected QUOTED or LITERAL)> E, [2018-12-21T13:33:13.719634 #1027-70334085745740] ERROR -- : ["/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3421:in `parse_error'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3273:in `match'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3205:in `string'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3186:in `nstring'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3144:in `address'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3116:in `address_list'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2344:in `envelope'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2325:in `envelope_data'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2300:in `msg_att'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2280:in `numeric_response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2222:in `response_untagged'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2198:in `response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2124:in `parse'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:1219:in `get_response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:1127:in `receive_responses'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0 /net/imap.rb:1109:in `block in initialize'", "/opt/zammad/vendor/bundle/ruby/2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context'"] I think the issue is coming up since zammad version 2.8.X The mails are related from plesk, which use < ! - as example in the mail as subject and content. As soon as I delete those e-mails and send a usual one without those symbols, emails are imported. ![mailexample](https://user-images.githubusercontent.com/15981620/50347191-0661f780-0535-11e9-814b-49326fc32bd4.jpg) To say, with 2.7.X I didn't have that troubles. Yours faithfully, Mathias
1.0
EMails cannot be imported from imap - Infos: Used Zammad version: 2.8 Installation method (source, package, ..): package Operating system: Ubuntu Server 16.04 LTS Database + version: 10.0.36-MariaDB-0ubuntu0.16.04.1 Ubuntu 16.04 Elasticsearch version: 5.6.9 Browser + version: any Error: Channel: Email::Account in Can't use Channel::Driver::Imap: #<Net::IMAP::ResponseParse Log: [2018-12-21T13:33:13.719440 #1027-70334085745740] ERROR -- : Can't use Channel::Driver::Imap: #<Net::IMAP::ResponseParseError: unexpected token LPAR (expected QUOTED or LITERAL)> E, [2018-12-21T13:33:13.719634 #1027-70334085745740] ERROR -- : ["/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3421:in `parse_error'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3273:in `match'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3205:in `string'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3186:in `nstring'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3144:in `address'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:3116:in `address_list'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2344:in `envelope'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2325:in `envelope_data'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2300:in `msg_att'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2280:in `numeric_response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2222:in `response_untagged'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2198:in `response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:2124:in `parse'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:1219:in `get_response'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0/net/imap.rb:1127:in `receive_responses'", "/opt/zammad/vendor/ruby-2.4.4/lib/ruby/2.4.0 /net/imap.rb:1109:in `block in initialize'", "/opt/zammad/vendor/bundle/ruby/2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context'"] I think the issue is coming up since zammad version 2.8.X The mails are related from plesk, which use < ! - as example in the mail as subject and content. As soon as I delete those e-mails and send a usual one without those symbols, emails are imported. ![mailexample](https://user-images.githubusercontent.com/15981620/50347191-0661f780-0535-11e9-814b-49326fc32bd4.jpg) To say, with 2.7.X I didn't have that troubles. Yours faithfully, Mathias
process
emails cannot be imported from imap infos used zammad version installation method source package package operating system ubuntu server lts database version mariadb ubuntu elasticsearch version browser version any error channel email account in can t use channel driver imap net imap responseparse log error can t use channel driver imap e error opt zammad vendor ruby lib ruby net imap rb in parse error opt zammad vendor ruby lib ruby net imap rb in match opt zammad vendor ruby lib ruby net imap rb in string opt zammad vendor ruby lib ruby net imap rb in nstring opt zammad vendor ruby lib ruby net imap rb in address opt zammad vendor ruby lib ruby net imap rb in address list opt zammad vendor ruby lib ruby net imap rb in envelope opt zammad vendor ruby lib ruby net imap rb in envelope data opt zammad vendor ruby lib ruby net imap rb in msg att opt zammad vendor ruby lib ruby net imap rb in numeric response opt zammad vendor ruby lib ruby net imap rb in response untagged opt zammad vendor ruby lib ruby net imap rb in response opt zammad vendor ruby lib ruby net imap rb in parse opt zammad vendor ruby lib ruby net imap rb in get response opt zammad vendor ruby lib ruby net imap rb in receive responses opt zammad vendor ruby lib ruby net imap rb in block in initialize opt zammad vendor bundle ruby gems logging lib logging diagnostic context rb in block in create with logging context i think the issue is coming up since zammad version x the mails are related from plesk which use as example in the mail as subject and content as soon as i delete those e mails and send a usual one without those symbols emails are imported to say with x i didn t have that troubles yours faithfully mathias
1
16,371
21,083,026,210
IssuesEvent
2022-04-03 07:21:14
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
How to see PCA components name?
enhancement preprocessing pca
I applied PCA and then feature importance but it shows important features with Component_1,2 and 3 so how can I know/understand which parameters are corresponding to these components? ![PCA3](https://user-images.githubusercontent.com/25801054/159423272-5310a224-07bf-4534-bc50-db4bd57de2f7.png)
1.0
How to see PCA components name? - I applied PCA and then feature importance but it shows important features with Component_1,2 and 3 so how can I know/understand which parameters are corresponding to these components? ![PCA3](https://user-images.githubusercontent.com/25801054/159423272-5310a224-07bf-4534-bc50-db4bd57de2f7.png)
process
how to see pca components name i applied pca and then feature importance but it shows important features with component and so how can i know understand which parameters are corresponding to these components
1
13,456
15,935,161,759
IssuesEvent
2021-04-14 09:31:39
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Better error handling for Introspection when JSON-RPC doesn't have an id
kind/improvement process/candidate team/migrations topic: error topic: introspection
See https://github.com/prisma/migrate/blob/9eec2a09d3722e81b563e0f6afccb773948e9869/src/LiftEngine.ts#L108 and https://github.com/prisma/prisma2/blob/b80e2028f0af6e7c45aa66645f2c37e8f6b1126c/cli/sdk/src/IntrospectionEngine.ts#L121 Current Example ``` if (!result.id) { console.error( `Response ${JSON.stringify( result, )} doesn't have an id and I can't handle that (yet)`, ) } ```
1.0
Better error handling for Introspection when JSON-RPC doesn't have an id - See https://github.com/prisma/migrate/blob/9eec2a09d3722e81b563e0f6afccb773948e9869/src/LiftEngine.ts#L108 and https://github.com/prisma/prisma2/blob/b80e2028f0af6e7c45aa66645f2c37e8f6b1126c/cli/sdk/src/IntrospectionEngine.ts#L121 Current Example ``` if (!result.id) { console.error( `Response ${JSON.stringify( result, )} doesn't have an id and I can't handle that (yet)`, ) } ```
process
better error handling for introspection when json rpc doesn t have an id see and current example if result id console error response json stringify result doesn t have an id and i can t handle that yet
1
35,034
7,887,543,491
IssuesEvent
2018-06-27 18:49:55
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Hackers have infaltrated com_contact through XSS and [POST:jform]
No Code Attached Yet
### Steps to reproduce the issue Upgraded to 3.8.10. As soon as I did Hackers from Germany hit com_contact with XSS attacks and [POST:jform] and send multiple emails through this component. I don't even use this component on my site. ### Expected result No hacking and XSS attack of com_contact ### Actual result emails are sent to me using this component 5 emails every 5 minutes ### System information (as much as possible) PHP 7.0.30 Joomla 3.8.10 ### Additional comments At this time I have installed Security Check Pro and it is stripping the XSS attack and stopping this. Won't upgrade any more sites until this is rectified.
1.0
Hackers have infaltrated com_contact through XSS and [POST:jform] - ### Steps to reproduce the issue Upgraded to 3.8.10. As soon as I did Hackers from Germany hit com_contact with XSS attacks and [POST:jform] and send multiple emails through this component. I don't even use this component on my site. ### Expected result No hacking and XSS attack of com_contact ### Actual result emails are sent to me using this component 5 emails every 5 minutes ### System information (as much as possible) PHP 7.0.30 Joomla 3.8.10 ### Additional comments At this time I have installed Security Check Pro and it is stripping the XSS attack and stopping this. Won't upgrade any more sites until this is rectified.
non_process
hackers have infaltrated com contact through xss and steps to reproduce the issue upgraded to as soon as i did hackers from germany hit com contact with xss attacks and and send multiple emails through this component i don t even use this component on my site expected result no hacking and xss attack of com contact actual result emails are sent to me using this component emails every minutes system information as much as possible php joomla additional comments at this time i have installed security check pro and it is stripping the xss attack and stopping this won t upgrade any more sites until this is rectified
0
34,302
9,331,423,524
IssuesEvent
2019-03-28 09:43:36
golang/go
https://api.github.com/repos/golang/go
closed
x/build/env/android: gomote debugging with Android builders is painful
Builders Documentation
@griesemer discovered that using gomote to debug things on Android is painful. We should have a wiki page with some example sessions, including how to build the exec wrapper, and how to run go tests.
1.0
x/build/env/android: gomote debugging with Android builders is painful - @griesemer discovered that using gomote to debug things on Android is painful. We should have a wiki page with some example sessions, including how to build the exec wrapper, and how to run go tests.
non_process
x build env android gomote debugging with android builders is painful griesemer discovered that using gomote to debug things on android is painful we should have a wiki page with some example sessions including how to build the exec wrapper and how to run go tests
0
2,551
5,310,583,929
IssuesEvent
2017-02-12 21:17:28
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [fr] « Il faut sortir du nucléaire » - Jean-Luc Mélenchon
Language: French Process: [6] Approved
# Video title « Il faut sortir du nucléaire » - Jean-Luc Mélenchon # URL https://www.youtube.com/watch?v=TYmeiHryrCE # Youtube subtitles language French # Duration 1:56 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&lang=fr&v=TYmeiHryrCE&tab=captions&action_mde_edit_form=1&bl=vmp&ui=hd
1.0
[subtitles] [fr] « Il faut sortir du nucléaire » - Jean-Luc Mélenchon - # Video title « Il faut sortir du nucléaire » - Jean-Luc Mélenchon # URL https://www.youtube.com/watch?v=TYmeiHryrCE # Youtube subtitles language French # Duration 1:56 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&lang=fr&v=TYmeiHryrCE&tab=captions&action_mde_edit_form=1&bl=vmp&ui=hd
process
« il faut sortir du nucléaire » jean luc mélenchon video title « il faut sortir du nucléaire » jean luc mélenchon url youtube subtitles language french duration subtitles url
1
6,501
9,574,761,668
IssuesEvent
2019-05-07 03:17:06
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
opened
Parameter is missing in the runbook code
automation/svc process-automation/subsvc
HI Team, I noticed this runbook was updated recently . However, the parameter part is missing after update . Witouout parameter part, runbook will fail . Please update doc accordingly, thanks . ===================== [OutputType("PSAzureOperationResponse")] param ( [Parameter (Mandatory=$false)] [object] $WebhookData ) $ErrorActionPreference = "stop" ================= --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 52a0d4e1-7d42-5c9f-6e73-4c15d416e5ba * Version Independent ID: 2b5ab95a-ced9-f1e5-a093-430e79b711b2 * Content: [Use an alert to trigger an Azure Automation runbook](https://docs.microsoft.com/en-us/azure/automation/automation-create-alert-triggered-runbook#feedback) * Content Source: [articles/automation/automation-create-alert-triggered-runbook.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-create-alert-triggered-runbook.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @georgewallace * Microsoft Alias: **gwallace**
1.0
Parameter is missing in the runbook code - HI Team, I noticed this runbook was updated recently . However, the parameter part is missing after update . Witouout parameter part, runbook will fail . Please update doc accordingly, thanks . ===================== [OutputType("PSAzureOperationResponse")] param ( [Parameter (Mandatory=$false)] [object] $WebhookData ) $ErrorActionPreference = "stop" ================= --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 52a0d4e1-7d42-5c9f-6e73-4c15d416e5ba * Version Independent ID: 2b5ab95a-ced9-f1e5-a093-430e79b711b2 * Content: [Use an alert to trigger an Azure Automation runbook](https://docs.microsoft.com/en-us/azure/automation/automation-create-alert-triggered-runbook#feedback) * Content Source: [articles/automation/automation-create-alert-triggered-runbook.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-create-alert-triggered-runbook.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @georgewallace * Microsoft Alias: **gwallace**
process
parameter is missing in the runbook code hi team i noticed this runbook was updated recently however the parameter part is missing after update witouout parameter part runbook will fail please update doc accordingly thanks param webhookdata erroractionpreference stop document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace
1
6,376
9,428,563,404
IssuesEvent
2019-04-12 01:46:41
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
1.14 Merge Order
area: Process
This Issue is to track which changes should try and get merged in which order early on so we can reduce churn & conflicts: Topic | Merge Status | impact to code base | PR | Owner (s) | ----- | -------------| ------------------- | ---|-----------| Move SoC code to top-level | **Merged**| impacts new SoC supports in arch/*/soc | #9776 | @nashif | DTS code generation | not ready (#8499) | | #8561| @erwango @b0661 | CMake library generation changes | not ready | | #8451 | @tejlmand | Moving to new Logging subsystem | in progress | numerous PRs | [logging PRs](https://github.com/zephyrproject-rtos/zephyr/pulls?q=is%3Apr+is%3Aopen+label%3A%22area%3A+Logging%22) | @nashif | Moving to new shell subsystem | in progress | | #9362 | @jarz-nordic @carlescufi | Multi-Repo, west tool | in progress | | #7338 | @carlescufi |
1.0
1.14 Merge Order - This Issue is to track which changes should try and get merged in which order early on so we can reduce churn & conflicts: Topic | Merge Status | impact to code base | PR | Owner (s) | ----- | -------------| ------------------- | ---|-----------| Move SoC code to top-level | **Merged**| impacts new SoC supports in arch/*/soc | #9776 | @nashif | DTS code generation | not ready (#8499) | | #8561| @erwango @b0661 | CMake library generation changes | not ready | | #8451 | @tejlmand | Moving to new Logging subsystem | in progress | numerous PRs | [logging PRs](https://github.com/zephyrproject-rtos/zephyr/pulls?q=is%3Apr+is%3Aopen+label%3A%22area%3A+Logging%22) | @nashif | Moving to new shell subsystem | in progress | | #9362 | @jarz-nordic @carlescufi | Multi-Repo, west tool | in progress | | #7338 | @carlescufi |
process
merge order this issue is to track which changes should try and get merged in which order early on so we can reduce churn conflicts topic merge status impact to code base pr owner s move soc code to top level merged impacts new soc supports in arch soc nashif dts code generation not ready erwango cmake library generation changes not ready tejlmand moving to new logging subsystem in progress numerous prs nashif moving to new shell subsystem in progress jarz nordic carlescufi multi repo west tool in progress carlescufi
1
17,607
23,427,793,778
IssuesEvent
2022-08-14 16:53:25
vortexntnu/Vortex-CV
https://api.github.com/repos/vortexntnu/Vortex-CV
closed
Hough transform manifolding as base layer (Feature Detection)
enhancement feature moderate priority Image Processing
**Time estimate:** 10 hours **Deadline:** 01.02.22 **Description of task:** The colour-shape-point-fitting based object detection that works reliably in normal environments, does not work well underwater, especially over the distances of over 3-4 meters. The cause is seemingly the blue-shift that happens over the length of the pool, and the fact that we are using HSV detection as the base for all other steps in the algorithm. Possible solution to this problem is Hough transform manifolding - searching for lines using the Hough transform that form a certain shape, e.g., gate.
1.0
Hough transform manifolding as base layer (Feature Detection) - **Time estimate:** 10 hours **Deadline:** 01.02.22 **Description of task:** The colour-shape-point-fitting based object detection that works reliably in normal environments, does not work well underwater, especially over the distances of over 3-4 meters. The cause is seemingly the blue-shift that happens over the length of the pool, and the fact that we are using HSV detection as the base for all other steps in the algorithm. Possible solution to this problem is Hough transform manifolding - searching for lines using the Hough transform that form a certain shape, e.g., gate.
process
hough transform manifolding as base layer feature detection time estimate hours deadline description of task the colour shape point fitting based object detection that works reliably in normal environments does not work well underwater especially over the distances of over meters the cause is seemingly the blue shift that happens over the length of the pool and the fact that we are using hsv detection as the base for all other steps in the algorithm possible solution to this problem is hough transform manifolding searching for lines using the hough transform that form a certain shape e g gate
1
119,016
4,759,495,532
IssuesEvent
2016-10-24 22:48:03
kdahlquist/GRNmap
https://api.github.com/repos/kdahlquist/GRNmap
opened
Data Analysis Team tasks for Week of 10/24
data analysis logistics priority 0
The tasks for this week: * @kdahlquist has "review requested" for - [ ] #230 (degree distribution charts) - [ ] #245, generating and vetting input workbooks (after requested changes are made * @bklein7 and @Nwilli31 are working on the following: - [ ] #245, generating and vetting input workbooks, finishing off requested changes - [ ] #241, update 16_tests workbooks on beta branch to be consistent with our formatting guidelines - [ ] #265 start the model runs using the v1.4.4 source * @khorstmann and @maggie-oneil are working on the following: - [ ] #259 @maggie-oneil and @khorstmann will work on getting the SBET toolbox up and running, starting with betweenness centrality. Note that nobody seems to be working on #170, so I changed the priority to 0.5 on that one.
1.0
Data Analysis Team tasks for Week of 10/24 - The tasks for this week: * @kdahlquist has "review requested" for - [ ] #230 (degree distribution charts) - [ ] #245, generating and vetting input workbooks (after requested changes are made * @bklein7 and @Nwilli31 are working on the following: - [ ] #245, generating and vetting input workbooks, finishing off requested changes - [ ] #241, update 16_tests workbooks on beta branch to be consistent with our formatting guidelines - [ ] #265 start the model runs using the v1.4.4 source * @khorstmann and @maggie-oneil are working on the following: - [ ] #259 @maggie-oneil and @khorstmann will work on getting the SBET toolbox up and running, starting with betweenness centrality. Note that nobody seems to be working on #170, so I changed the priority to 0.5 on that one.
non_process
data analysis team tasks for week of the tasks for this week kdahlquist has review requested for degree distribution charts generating and vetting input workbooks after requested changes are made and are working on the following generating and vetting input workbooks finishing off requested changes update tests workbooks on beta branch to be consistent with our formatting guidelines start the model runs using the source khorstmann and maggie oneil are working on the following maggie oneil and khorstmann will work on getting the sbet toolbox up and running starting with betweenness centrality note that nobody seems to be working on so i changed the priority to on that one
0
18,311
3,041,566,814
IssuesEvent
2015-08-07 22:15:48
francoisferland/casiousbmididriver
https://api.github.com/repos/francoisferland/casiousbmididriver
closed
Not compatible with Yosemite
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Try installing the driver on a Yosemite mac 2. Connect Casio Keyboard and open a DAW like Logic Pro X 3. What is the expected output? What do you see instead? Keyboard isn't recognized and doesn't show up in Audio MIDI setup What version of the product are you using? On what operating system? mid-2014 13 inch retina Macbook Pro running Yosemite Please provide any additional information below. ``` Original issue reported on code.google.com by `ztca...@g.uky.edu` on 17 Nov 2014 at 7:33
1.0
Not compatible with Yosemite - ``` What steps will reproduce the problem? 1. Try installing the driver on a Yosemite mac 2. Connect Casio Keyboard and open a DAW like Logic Pro X 3. What is the expected output? What do you see instead? Keyboard isn't recognized and doesn't show up in Audio MIDI setup What version of the product are you using? On what operating system? mid-2014 13 inch retina Macbook Pro running Yosemite Please provide any additional information below. ``` Original issue reported on code.google.com by `ztca...@g.uky.edu` on 17 Nov 2014 at 7:33
non_process
not compatible with yosemite what steps will reproduce the problem try installing the driver on a yosemite mac connect casio keyboard and open a daw like logic pro x what is the expected output what do you see instead keyboard isn t recognized and doesn t show up in audio midi setup what version of the product are you using on what operating system mid inch retina macbook pro running yosemite please provide any additional information below original issue reported on code google com by ztca g uky edu on nov at
0
153,655
24,168,882,775
IssuesEvent
2022-09-22 17:19:39
MetaMask/metamask-mobile
https://api.github.com/repos/MetaMask/metamask-mobile
closed
Unable to re-add account nickname after making it an empty string
type-bug needs-design Sev2-normal Priority - High stability-team
**Describe the bug** If i were to remove my account nick name i.e. leave it blank, I am unable to add a new nickname to my account. see [recording](http://recordit.co/blCne5klOv ) **To Reproduce** - launch the app - create/import your wallet - while on the wallet view - tap and hold the account nickname to highlight it - remove the default nickname - attempt to add a new account nick name. Notice you are unable to do so **Expected behavior** You should be able to update an account nick name.
1.0
Unable to re-add account nickname after making it an empty string - **Describe the bug** If i were to remove my account nick name i.e. leave it blank, I am unable to add a new nickname to my account. see [recording](http://recordit.co/blCne5klOv ) **To Reproduce** - launch the app - create/import your wallet - while on the wallet view - tap and hold the account nickname to highlight it - remove the default nickname - attempt to add a new account nick name. Notice you are unable to do so **Expected behavior** You should be able to update an account nick name.
non_process
unable to re add account nickname after making it an empty string describe the bug if i were to remove my account nick name i e leave it blank i am unable to add a new nickname to my account see to reproduce launch the app create import your wallet while on the wallet view tap and hold the account nickname to highlight it remove the default nickname attempt to add a new account nick name notice you are unable to do so expected behavior you should be able to update an account nick name
0
327,839
9,981,949,569
IssuesEvent
2019-07-10 08:45:13
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
Revert goreleaser from updating jx brew package
area/versions in progress kind/fox priority/important-soon
goreleaser is being used as a temporary measure to release new jx versions to brew, this is not being released in accordance with our CI/CD model. the homebrew-jx package will be updated by `jx step create pr brew` once merged.
1.0
Revert goreleaser from updating jx brew package - goreleaser is being used as a temporary measure to release new jx versions to brew, this is not being released in accordance with our CI/CD model. the homebrew-jx package will be updated by `jx step create pr brew` once merged.
non_process
revert goreleaser from updating jx brew package goreleaser is being used as a temporary measure to release new jx versions to brew this is not being released in accordance with our ci cd model the homebrew jx package will be updated by jx step create pr brew once merged
0
179,005
6,620,758,902
IssuesEvent
2017-09-21 16:35:55
gonetz/GLideN64
https://api.github.com/repos/gonetz/GLideN64
closed
Zelda: OoT missing fences
Priority-High Regression
You can see the original report of this issue (including mupen64plus save state) here: https://github.com/m64p/mupen64plus-GLideN64/issues/13 I verified the issue, and bisected it to this commit: 187f9ef390052906e3516e1aaa71078125d099de
1.0
Zelda: OoT missing fences - You can see the original report of this issue (including mupen64plus save state) here: https://github.com/m64p/mupen64plus-GLideN64/issues/13 I verified the issue, and bisected it to this commit: 187f9ef390052906e3516e1aaa71078125d099de
non_process
zelda oot missing fences you can see the original report of this issue including save state here i verified the issue and bisected it to this commit
0
7,980
11,170,031,210
IssuesEvent
2019-12-28 10:35:27
konlpy/konlpy
https://api.github.com/repos/konlpy/konlpy
closed
pytorch의 data loader와 konlpy 사용 (jpype error)
Keyword/multiprocess_thread Status/help wanted question
pytorch의 data loader에서 multi process를 사용하는데 다음과 같이 에러가 나오네요 ㅠ import는 다됩니다만..ㅠ (OS는 centOS입니다 ㅠ) Traceback (most recent call last): File "/usr/lib/python2.7/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/lib/python2.7/multiprocessing/process.py", line 114, in run self._target(*self._args, **self._kwargs) File "/usr/local/lib/python2.7/dist-packages/torch/utils/data/dataloader.py", line 43, in _worker_loop data_queue.put((idx, ExceptionWrapper(sys.exc_info()))) File "/usr/lib/python2.7/multiprocessing/queues.py", line 390, in put return send(obj) File "/usr/local/lib/python2.7/dist-packages/torch/multiprocessing/queue.py", line 17, in send ForkingPickler(buf, pickle.HIGHEST_PROTOCOL).dump(obj) File "/usr/lib/python2.7/pickle.py", line 224, in dump self.save(obj) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 548, in save_tuple save(element) File "/usr/lib/python2.7/pickle.py", line 331, in save self.save_reduce(obj=obj, *rv) File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce save(state) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 649, in save_dict self._batch_setitems(obj.iteritems()) File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems save(v) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 748, in save_global (obj, module, name)) PicklingError: Can't pickle <class 'jpype._jexception.java.lang.ClassFormatErrorPyRaisable'>: it's not found as jpype._jexception.java.lang.ClassFormatErrorPyRaisable <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/47763049-pytorch-data-loader-konlpy-jpype-error?utm_campaign=plugin&utm_content=tracker%2F1743549&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1743549&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
pytorch의 data loader와 konlpy 사용 (jpype error) - pytorch의 data loader에서 multi process를 사용하는데 다음과 같이 에러가 나오네요 ㅠ import는 다됩니다만..ㅠ (OS는 centOS입니다 ㅠ) Traceback (most recent call last): File "/usr/lib/python2.7/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/lib/python2.7/multiprocessing/process.py", line 114, in run self._target(*self._args, **self._kwargs) File "/usr/local/lib/python2.7/dist-packages/torch/utils/data/dataloader.py", line 43, in _worker_loop data_queue.put((idx, ExceptionWrapper(sys.exc_info()))) File "/usr/lib/python2.7/multiprocessing/queues.py", line 390, in put return send(obj) File "/usr/local/lib/python2.7/dist-packages/torch/multiprocessing/queue.py", line 17, in send ForkingPickler(buf, pickle.HIGHEST_PROTOCOL).dump(obj) File "/usr/lib/python2.7/pickle.py", line 224, in dump self.save(obj) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 548, in save_tuple save(element) File "/usr/lib/python2.7/pickle.py", line 331, in save self.save_reduce(obj=obj, *rv) File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce save(state) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 649, in save_dict self._batch_setitems(obj.iteritems()) File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems save(v) File "/usr/lib/python2.7/pickle.py", line 286, in save f(self, obj) # Call unbound method with explicit self File "/usr/lib/python2.7/pickle.py", line 748, in save_global (obj, module, name)) PicklingError: Can't pickle <class 'jpype._jexception.java.lang.ClassFormatErrorPyRaisable'>: it's not found as jpype._jexception.java.lang.ClassFormatErrorPyRaisable <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/47763049-pytorch-data-loader-konlpy-jpype-error?utm_campaign=plugin&utm_content=tracker%2F1743549&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1743549&utm_medium=issues&utm_source=github). </bountysource-plugin>
process
pytorch의 data loader와 konlpy 사용 jpype error pytorch의 data loader에서 multi process를 사용하는데 다음과 같이 에러가 나오네요 ㅠ import는 다됩니다만 ㅠ os는 centos입니다 ㅠ traceback most recent call last file usr lib multiprocessing process py line in bootstrap self run file usr lib multiprocessing process py line in run self target self args self kwargs file usr local lib dist packages torch utils data dataloader py line in worker loop data queue put idx exceptionwrapper sys exc info file usr lib multiprocessing queues py line in put return send obj file usr local lib dist packages torch multiprocessing queue py line in send forkingpickler buf pickle highest protocol dump obj file usr lib pickle py line in dump self save obj file usr lib pickle py line in save f self obj call unbound method with explicit self file usr lib pickle py line in save tuple save element file usr lib pickle py line in save self save reduce obj obj rv file usr lib pickle py line in save reduce save state file usr lib pickle py line in save f self obj call unbound method with explicit self file usr lib pickle py line in save dict self batch setitems obj iteritems file usr lib pickle py line in batch setitems save v file usr lib pickle py line in save f self obj call unbound method with explicit self file usr lib pickle py line in save global obj module name picklingerror can t pickle it s not found as jpype jexception java lang classformaterrorpyraisable want to back this issue we accept bounties via
1
69,443
30,283,614,435
IssuesEvent
2023-07-08 11:43:42
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
FR: Colorized output
Language Service Feature Request help wanted more votes needed Feature: Colorization
### Feature Request If a compilation/execution in performed in the Terminal window the output is colorized. But for the compilation the output in "Output" is not. Interactive, colorized text in "Output" will be great.
1.0
FR: Colorized output - ### Feature Request If a compilation/execution in performed in the Terminal window the output is colorized. But for the compilation the output in "Output" is not. Interactive, colorized text in "Output" will be great.
non_process
fr colorized output feature request if a compilation execution in performed in the terminal window the output is colorized but for the compilation the output in output is not interactive colorized text in output will be great
0
31,055
5,902,378,019
IssuesEvent
2017-05-19 01:05:30
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Confusion around use/guarantees of grpc.RpcError in Python
documentation python question
Modifying the Python `helloworld` example to add exception handling produces some confusing results. Changing modifying the `run()` function from https://github.com/grpc/grpc/blob/master/examples/python/helloworld/greeter_client.py to def run(): channel = grpc.insecure_channel('localhost:50051') stub = helloworld_pb2_grpc.GreeterStub(channel) try: response = stub.SayHello(helloworld_pb2.HelloRequest(name='you')) except grpc.RpcError as err: print('Type:', type(err)) print('Attributes:', dir(err)) print("Greeter client received: " + response.message) and then running the above client while the server is not running produces the following output: Type: <class 'grpc._channel._Rendezvous'> Attributes: ['__abstractmethods__', '__cause__', '__class__', '__context__', '__del__', '__delattr__', '__dict__', '__dir__', '__doc__', '__eq__', '__format__', '__ge__', '__getattribute__', '__gt__', '__hash__', '__init__', '__init_subclass__', '__iter__', '__le__', '__lt__', '__module__', '__ne__', '__new__', '__next__', '__reduce__', '__reduce_ex__', '__repr__', '__setattr__', '__setstate__', '__sizeof__', '__str__', '__subclasshook__', '__suppress_context__', '__traceback__', '__weakref__', '_abc_cache', '_abc_negative_cache', '_abc_negative_cache_version', '_abc_registry', '_call', '_deadline', '_next', '_repr', '_response_deserializer', '_state', 'add_callback', 'add_done_callback', 'args', 'cancel', 'cancelled', 'code', 'details', 'done', 'exception', 'initial_metadata', 'is_active', 'next', 'result', 'running', 'time_remaining', 'traceback', 'trailing_metadata', 'with_traceback'] Looking at the source code for `RpcError`, it's defined as: class RpcError(Exception): """Raised by the gRPC library to indicate non-OK-status RPC termination.""" So I'm wondering a few things: 1. What is this `_Rendezvous` class, and where is it defined? It doesn't appear in the package documentation (http://www.grpc.io/grpc/python/search.html?q=Rendezvous). A lot of its attributes potentially look very useful, but I'm just guessing as to what they actually do or represent. 2. Are there any guarantees about which of these attributes will always be available when RpcError is caught? E.g. if I start relying on using `err.code()` to get the error type, are there other RpcError subclasses that can be raised which don't contain this attribute? 3. Is there a list of all the possible subclasses of RpcError that can be raised? (i.e. in case I'd want different exception handling for each). I couldn't see anything in the examples or documentation, but basically, I'd like to use something like: try: # ... some gRPC call ... except grpc.RpcError as err: if err.code() == grpc.StatusCode.UNAVAILABLE: print('Service unavailable') # ... handling for different codes here ... but this looks unsafe, since `RpcError` doesn't define any attributes (beyond those inherited from `Exception`). Currently I'm using the following code, but I'm hoping there's a better option: try: # ...some gRPC call ... except grpc.RpcError as err: code = err.code() if hasattr(err, 'code') else None if code == grpc.StatusCode.UNAVAILABLE: print('Service unavailable') # ... handling for different codes here ...
1.0
Confusion around use/guarantees of grpc.RpcError in Python - Modifying the Python `helloworld` example to add exception handling produces some confusing results. Changing modifying the `run()` function from https://github.com/grpc/grpc/blob/master/examples/python/helloworld/greeter_client.py to def run(): channel = grpc.insecure_channel('localhost:50051') stub = helloworld_pb2_grpc.GreeterStub(channel) try: response = stub.SayHello(helloworld_pb2.HelloRequest(name='you')) except grpc.RpcError as err: print('Type:', type(err)) print('Attributes:', dir(err)) print("Greeter client received: " + response.message) and then running the above client while the server is not running produces the following output: Type: <class 'grpc._channel._Rendezvous'> Attributes: ['__abstractmethods__', '__cause__', '__class__', '__context__', '__del__', '__delattr__', '__dict__', '__dir__', '__doc__', '__eq__', '__format__', '__ge__', '__getattribute__', '__gt__', '__hash__', '__init__', '__init_subclass__', '__iter__', '__le__', '__lt__', '__module__', '__ne__', '__new__', '__next__', '__reduce__', '__reduce_ex__', '__repr__', '__setattr__', '__setstate__', '__sizeof__', '__str__', '__subclasshook__', '__suppress_context__', '__traceback__', '__weakref__', '_abc_cache', '_abc_negative_cache', '_abc_negative_cache_version', '_abc_registry', '_call', '_deadline', '_next', '_repr', '_response_deserializer', '_state', 'add_callback', 'add_done_callback', 'args', 'cancel', 'cancelled', 'code', 'details', 'done', 'exception', 'initial_metadata', 'is_active', 'next', 'result', 'running', 'time_remaining', 'traceback', 'trailing_metadata', 'with_traceback'] Looking at the source code for `RpcError`, it's defined as: class RpcError(Exception): """Raised by the gRPC library to indicate non-OK-status RPC termination.""" So I'm wondering a few things: 1. What is this `_Rendezvous` class, and where is it defined? It doesn't appear in the package documentation (http://www.grpc.io/grpc/python/search.html?q=Rendezvous). A lot of its attributes potentially look very useful, but I'm just guessing as to what they actually do or represent. 2. Are there any guarantees about which of these attributes will always be available when RpcError is caught? E.g. if I start relying on using `err.code()` to get the error type, are there other RpcError subclasses that can be raised which don't contain this attribute? 3. Is there a list of all the possible subclasses of RpcError that can be raised? (i.e. in case I'd want different exception handling for each). I couldn't see anything in the examples or documentation, but basically, I'd like to use something like: try: # ... some gRPC call ... except grpc.RpcError as err: if err.code() == grpc.StatusCode.UNAVAILABLE: print('Service unavailable') # ... handling for different codes here ... but this looks unsafe, since `RpcError` doesn't define any attributes (beyond those inherited from `Exception`). Currently I'm using the following code, but I'm hoping there's a better option: try: # ...some gRPC call ... except grpc.RpcError as err: code = err.code() if hasattr(err, 'code') else None if code == grpc.StatusCode.UNAVAILABLE: print('Service unavailable') # ... handling for different codes here ...
non_process
confusion around use guarantees of grpc rpcerror in python modifying the python helloworld example to add exception handling produces some confusing results changing modifying the run function from to def run channel grpc insecure channel localhost stub helloworld grpc greeterstub channel try response stub sayhello helloworld hellorequest name you except grpc rpcerror as err print type type err print attributes dir err print greeter client received response message and then running the above client while the server is not running produces the following output type attributes abstractmethods cause class context del delattr dict dir doc eq format ge getattribute gt hash init init subclass iter le lt module ne new next reduce reduce ex repr setattr setstate sizeof str subclasshook suppress context traceback weakref abc cache abc negative cache abc negative cache version abc registry call deadline next repr response deserializer state add callback add done callback args cancel cancelled code details done exception initial metadata is active next result running time remaining traceback trailing metadata with traceback looking at the source code for rpcerror it s defined as class rpcerror exception raised by the grpc library to indicate non ok status rpc termination so i m wondering a few things what is this rendezvous class and where is it defined it doesn t appear in the package documentation a lot of its attributes potentially look very useful but i m just guessing as to what they actually do or represent are there any guarantees about which of these attributes will always be available when rpcerror is caught e g if i start relying on using err code to get the error type are there other rpcerror subclasses that can be raised which don t contain this attribute is there a list of all the possible subclasses of rpcerror that can be raised i e in case i d want different exception handling for each i couldn t see anything in the examples or documentation but basically i d like to use something like try some grpc call except grpc rpcerror as err if err code grpc statuscode unavailable print service unavailable handling for different codes here but this looks unsafe since rpcerror doesn t define any attributes beyond those inherited from exception currently i m using the following code but i m hoping there s a better option try some grpc call except grpc rpcerror as err code err code if hasattr err code else none if code grpc statuscode unavailable print service unavailable handling for different codes here
0
4,434
7,308,600,203
IssuesEvent
2018-02-28 08:58:30
UKHomeOffice/dq-aws-transition
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
reopened
Add data-transfer job for OAG data to S3 archive
DQ Data Pipeline DQ Tranche 1 Production SSM processing
Add Data Transfer job for OAG data to S3 archive
1.0
Add data-transfer job for OAG data to S3 archive - Add Data Transfer job for OAG data to S3 archive
process
add data transfer job for oag data to archive add data transfer job for oag data to archive
1
383
2,823,574,458
IssuesEvent
2015-05-21 09:39:51
austundag/testing
https://api.github.com/repos/austundag/testing
closed
Add Allergies info to eHMP Toolbar
enhancement in process
1. It should be next to CARE TEAM INFORMATION and should have similarly three lines with same fonts. 2. The title should say 'ALLERGIES' and it should be red. 3. The second line should say "Local: <allergy name> <allergy name> ,,," 4. The third line should say "Remote: <allergy name> <allergy name> ,,,"
1.0
Add Allergies info to eHMP Toolbar - 1. It should be next to CARE TEAM INFORMATION and should have similarly three lines with same fonts. 2. The title should say 'ALLERGIES' and it should be red. 3. The second line should say "Local: <allergy name> <allergy name> ,,," 4. The third line should say "Remote: <allergy name> <allergy name> ,,,"
process
add allergies info to ehmp toolbar it should be next to care team information and should have similarly three lines with same fonts the title should say allergies and it should be red the second line should say local the third line should say remote
1
235,432
19,346,259,027
IssuesEvent
2021-12-15 11:07:37
mozilla-mobile/focus-ios
https://api.github.com/repos/mozilla-mobile/focus-ios
opened
[XCUITest] Create new automated tests to cover Pro Tips
eng:ui-test eng:automation
So that there is a similar automated test to the exisiting [manual test](https://testrail.stage.mozaws.net/index.php?/cases/view/1514601)
1.0
[XCUITest] Create new automated tests to cover Pro Tips - So that there is a similar automated test to the exisiting [manual test](https://testrail.stage.mozaws.net/index.php?/cases/view/1514601)
non_process
create new automated tests to cover pro tips so that there is a similar automated test to the exisiting
0
126,249
26,810,215,999
IssuesEvent
2023-02-01 21:40:47
starperov-net/universal_flash_cards_chatbot
https://api.github.com/repos/starperov-net/universal_flash_cards_chatbot
opened
unite duplicate code in /study and /selftest
Yana code_refactoring
Винести дублюючий код у файлі study.py та quick_self_test.py в окремі функції.
1.0
unite duplicate code in /study and /selftest - Винести дублюючий код у файлі study.py та quick_self_test.py в окремі функції.
non_process
unite duplicate code in study and selftest винести дублюючий код у файлі study py та quick self test py в окремі функції
0
57,015
3,081,230,526
IssuesEvent
2015-08-22 14:18:01
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Load Map option on level editor
020 enhancement imported Priority-Medium wontfix
_From [Cory.Pou...@gmail.com](https://code.google.com/u/105220164791991617712/) on September 16, 2013 17:11:56_ So i think a load map option on the level editor would be nice as you would not have to quit the level editor then open it again to get a new map. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=246_
1.0
Load Map option on level editor - _From [Cory.Pou...@gmail.com](https://code.google.com/u/105220164791991617712/) on September 16, 2013 17:11:56_ So i think a load map option on the level editor would be nice as you would not have to quit the level editor then open it again to get a new map. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=246_
non_process
load map option on level editor from on september so i think a load map option on the level editor would be nice as you would not have to quit the level editor then open it again to get a new map original issue
0
15,563
19,703,504,190
IssuesEvent
2022-01-12 19:08:03
googleapis/java-shell
https://api.github.com/repos/googleapis/java-shell
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'shell' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'shell' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname shell invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
28,417
4,104,127,006
IssuesEvent
2016-06-05 05:35:49
fossasia/engelsystem
https://api.github.com/repos/fossasia/engelsystem
opened
UI/UX Design : Register Page
design enhancement
1-Changing the layout , currently the layout is unordered and I plan to make it more ordered and user-friendly. As we can see the layout looks a bit confusing ![screenshot from 2016-06-05 11 05 18](https://cloud.githubusercontent.com/assets/14835922/15803802/695ca812-2b0d-11e6-9f0e-6fc1b94b0ff9.png) I will be dividing the layout into 2 pages. One for personal details and other for volunteer preferences as shown below.In the volunteer preferences the user can set priority also. ![screenshot from 2016-06-05 11 03 43](https://cloud.githubusercontent.com/assets/14835922/15803797/3ab912ac-2b0d-11e6-97ad-ed744777d8b6.png) ![screenshot from 2016-06-05 11 04 00](https://cloud.githubusercontent.com/assets/14835922/15803798/3aba1012-2b0d-11e6-954a-0eb64c4b47bf.png)
1.0
UI/UX Design : Register Page - 1-Changing the layout , currently the layout is unordered and I plan to make it more ordered and user-friendly. As we can see the layout looks a bit confusing ![screenshot from 2016-06-05 11 05 18](https://cloud.githubusercontent.com/assets/14835922/15803802/695ca812-2b0d-11e6-9f0e-6fc1b94b0ff9.png) I will be dividing the layout into 2 pages. One for personal details and other for volunteer preferences as shown below.In the volunteer preferences the user can set priority also. ![screenshot from 2016-06-05 11 03 43](https://cloud.githubusercontent.com/assets/14835922/15803797/3ab912ac-2b0d-11e6-97ad-ed744777d8b6.png) ![screenshot from 2016-06-05 11 04 00](https://cloud.githubusercontent.com/assets/14835922/15803798/3aba1012-2b0d-11e6-954a-0eb64c4b47bf.png)
non_process
ui ux design register page changing the layout currently the layout is unordered and i plan to make it more ordered and user friendly as we can see the layout looks a bit confusing i will be dividing the layout into pages one for personal details and other for volunteer preferences as shown below in the volunteer preferences the user can set priority also
0
12,086
14,740,055,825
IssuesEvent
2021-01-07 08:26:19
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
El Paso - SA Billing - Late Fee Account List
anc-process anp-important ant-bug has attachment
In GitLab by @kdjstudios on Oct 3, 2018, 11:02 [El_Paso.xlsx](/uploads/212b7d3ab2a040017c72b5cab156a385/El_Paso.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-33348/conversation
1.0
El Paso - SA Billing - Late Fee Account List - In GitLab by @kdjstudios on Oct 3, 2018, 11:02 [El_Paso.xlsx](/uploads/212b7d3ab2a040017c72b5cab156a385/El_Paso.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-33348/conversation
process
el paso sa billing late fee account list in gitlab by kdjstudios on oct uploads el paso xlsx hd
1
1,391
3,960,929,993
IssuesEvent
2016-05-02 09:57:46
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Make trials.recruitment_status enumerable
API Processors
http://www.who.int/ictrp/network/trds/en/ > 18. Recruitment Status > Recruitment status of this trial: > Pending: participants are not yet being recruited or enrolled at any site > Recruiting: participants are currently being recruited and enrolled > Suspended: there is a temporary halt in recruitment and enrolment > Complete: participants are no longer being recruited or enrolled > Other Will be useful (vital) for variations (discrepancy) report.
1.0
Make trials.recruitment_status enumerable - http://www.who.int/ictrp/network/trds/en/ > 18. Recruitment Status > Recruitment status of this trial: > Pending: participants are not yet being recruited or enrolled at any site > Recruiting: participants are currently being recruited and enrolled > Suspended: there is a temporary halt in recruitment and enrolment > Complete: participants are no longer being recruited or enrolled > Other Will be useful (vital) for variations (discrepancy) report.
process
make trials recruitment status enumerable recruitment status recruitment status of this trial pending participants are not yet being recruited or enrolled at any site recruiting participants are currently being recruited and enrolled suspended there is a temporary halt in recruitment and enrolment complete participants are no longer being recruited or enrolled other will be useful vital for variations discrepancy report
1
17,815
23,741,281,479
IssuesEvent
2022-08-31 12:39:45
km4ack/patmenu2
https://api.github.com/repos/km4ack/patmenu2
closed
check for internet before downloading gateway list
bug enhancement in process
If a pi boots into hotspot mode before connecting to a known good SSID, the gateway list download will fail if it is run at boot from cron. To avoid this, we need to add a check for internet when the [getardoplist-cron](https://github.com/km4ack/patmenu2/blob/master/.getardoplist-cron) script is called. This code should work at the beginning of the script: ``` #check for internet before attempting to download the list internet=1 #set the test to false while [ "$internet" = 1 ]; do echo "check for internet connection" ping -c 5 8.8.8.8 internet=$? sleep 60 #wait one minute before trying again done echo "Internet connection detected" ``` If no connection is detected, the script will retry every 60 seconds.
1.0
check for internet before downloading gateway list - If a pi boots into hotspot mode before connecting to a known good SSID, the gateway list download will fail if it is run at boot from cron. To avoid this, we need to add a check for internet when the [getardoplist-cron](https://github.com/km4ack/patmenu2/blob/master/.getardoplist-cron) script is called. This code should work at the beginning of the script: ``` #check for internet before attempting to download the list internet=1 #set the test to false while [ "$internet" = 1 ]; do echo "check for internet connection" ping -c 5 8.8.8.8 internet=$? sleep 60 #wait one minute before trying again done echo "Internet connection detected" ``` If no connection is detected, the script will retry every 60 seconds.
process
check for internet before downloading gateway list if a pi boots into hotspot mode before connecting to a known good ssid the gateway list download will fail if it is run at boot from cron to avoid this we need to add a check for internet when the script is called this code should work at the beginning of the script check for internet before attempting to download the list internet set the test to false while do echo check for internet connection ping c internet sleep wait one minute before trying again done echo internet connection detected if no connection is detected the script will retry every seconds
1
14,196
17,099,062,027
IssuesEvent
2021-07-09 08:39:03
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type
bug/1-repro-available kind/bug process/candidate team/migrations
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma migrate dev --name unique` Version: `2.26.0` Binary Version: `9b816b3aa13cc270074f172f30d6eda8a8ce867d` Report: https://prisma-errors.netlify.app/report/13408 OS: `x64 darwin 20.3.0` JS Stacktrace: ``` Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type Please create an issue with your `schema.prisma` at https://github.com/prisma/prisma/issues/new at ChildProcess.<anonymous> (/node_modules/prisma/build/index.js:55980:23) at ChildProcess.emit (events.js:315:20) at ChildProcess.EventEmitter.emit (domain.js:486:12) at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) ``` Rust Stacktrace: ``` [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type ```
1.0
Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma migrate dev --name unique` Version: `2.26.0` Binary Version: `9b816b3aa13cc270074f172f30d6eda8a8ce867d` Report: https://prisma-errors.netlify.app/report/13408 OS: `x64 darwin 20.3.0` JS Stacktrace: ``` Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type Please create an issue with your `schema.prisma` at https://github.com/prisma/prisma/issues/new at ChildProcess.<anonymous> (/node_modules/prisma/build/index.js:55980:23) at ChildProcess.emit (events.js:315:20) at ChildProcess.EventEmitter.emit (domain.js:486:12) at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) ``` Rust Stacktrace: ``` [migration-engine/connectors/sql-migration-connector/src/sql_schema_calculator/sql_schema_calculator_flavour.rs:34:9] internal error: entered unreachable code: unreachable enum_column_type ```
process
error error in migration engine reason internal error entered unreachable code unreachable enum column type command prisma migrate dev name unique version binary version report os darwin js stacktrace error error in migration engine reason internal error entered unreachable code unreachable enum column type please create an issue with your schema prisma at at childprocess node modules prisma build index js at childprocess emit events js at childprocess eventemitter emit domain js at process childprocess handle onexit internal child process js rust stacktrace internal error entered unreachable code unreachable enum column type
1
11,435
14,249,240,884
IssuesEvent
2020-11-19 14:00:57
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
cop: panic by assertion failure in BatchTableScanExecutor
severity/Moderate sig/coprocessor type/bug
## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? <!-- You can run `tikv-server --version` --> master with commit `4d8e469860292b10636f1b25bf7a19da84e0b067` ### What operating system and CPU are you using? <!-- If you're using Linux, you can run `cat /proc/cpuinfo` --> ### Steps to reproduce <!-- If possible, provide a recipe for reproducing the error. A complete runnable program is good. --> tidb-lightning integration test fails in create table tikv panic log as follows: ``` [2020-08-31T10:28:27.692Z] [2020/08/31 18:14:41.452 +08:00] [FATAL] [lib.rs:481] ["assertion failed: `(left == right)` left: `1`, right: `2`"] [backtrace="stack backtrace: 0: tikv_util::set_panic_hook::{{closure}} at components/tikv_util/src/lib.rs:480 1: std::panicking::rust_panic_with_hook at src/libstd/panicking.rs:524 2: rust_begin_unwind at src/libstd/panicking.rs:431 3: std::panicking::begin_panic_fmt at src/libstd/panicking.rs:385 4: tidb_query_datatype::codec::batch::lazy_column_vec::LazyBatchColumnVec::assert_columns_equal_length at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/macros.rs:16 5: <tidb_query_vec_executors::util::scan_executor::ScanExecutor<S,I> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/util/scan_executor.rs:172 <tidb_query_vec_executors::table_scan_executor::BatchTableScanExecutor<S> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/table_scan_executor.rs:110 6: <tidb_query_common::execute_stats::WithSummaryCollector<C,T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:108 7: <alloc::boxed::Box<T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:77 <tidb_query_vec_executors::selection_executor::BatchSelectionExecutor<Src> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/selection_executor.rs:179 <tidb_query_common::execute_stats::WithSummaryCollector<C,T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:108 8: <alloc::boxed::Box<T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:77 tidb_query_vec_executors::runner::BatchExecutorsRunner<SS>::internal_handle_request at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/runner.rs:504 9: tidb_query_vec_executors::runner::BatchExecutorsRunner<SS>::handle_request::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/runner.rs:389 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <tikv::coprocessor::dag::BatchDAGHandler as tikv::coprocessor::RequestHandler>::handle_request::__handle_request::{{closure}} at src/coprocessor/dag/mod.rs:152 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <core::pin::Pin<P> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/future.rs:119 <minitrace::future::TraceWrapped<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:107 10: <core::pin::Pin<P> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/future.rs:119 <minitrace::future::TraceWrapped<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:107 <tikv::coprocessor::interceptors::tracker::Tracker<F> as core::future::future::Future>::poll at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/interceptors/tracker.rs:45 11: <tikv::coprocessor::interceptors::concurrency_limiter::ConcurrencyLimiter<PF,F> as core::future::future::Future>::poll at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/interceptors/concurrency_limiter.rs:101 tikv::coprocessor::endpoint::Endpoint<E>::handle_unary_request_impl::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/endpoint.rs:402 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <minitrace::future::TraceSpawned<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:77 tikv::read_pool::ReadPoolHandle::spawn_handle::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:139 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 12: tikv::read_pool::ReadPoolHandle::spawn::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:115 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 13: <yatp::task::future::Runner as yatp::pool::runner::Runner>::handle at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/task/future.rs:254 14: <tikv::read_pool::ReadPoolRunner<E,R> as yatp::pool::runner::Runner>::handle at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:169 <yatp::queue::multilevel::MultilevelRunner<R> as yatp::pool::runner::Runner>::handle at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/queue/multilevel.rs:245 yatp::pool::worker::WorkerThread<T,R>::run at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/pool/worker.rs:48 yatp::pool::builder::LazyBuilder<T>::build::{{closure}} at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/pool/builder.rs:91 std::sys_common::backtrace::__rust_begin_short_backtrace at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/sys_common/backtrace.rs:130 15: std::thread::Builder::spawn_unchecked::{{closure}}::{{closure}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/thread/mod.rs:475 <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panic.rs:318 std::panicking::try::do_call at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panicking.rs:342 std::panicking::try at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panicking.rs:319 std::panic::catch_unwind at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panic.rs:394 std::thread::Builder::spawn_unchecked::{{closure}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/thread/mod.rs:474 core::ops::function::FnOnce::call_once{{vtable.shim}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/ops/function.rs:233 16: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/liballoc/boxed.rs:1078 <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/liballoc/boxed.rs:1078 std::sys::unix::thread::Thread::new::thread_start at src/libstd/sys/unix/thread.rs:87 17: start_thread 18: __clone "] [location=/rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/macros.rs:16] [thread_name=unified-read-pool-7] ``` this panic can reporduce stably. for more detail log, see: https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/lightning_ghpr_test/detail/lightning_ghpr_test/1462/pipeline/50 ### What did you expect? ### What did happened?
1.0
cop: panic by assertion failure in BatchTableScanExecutor - ## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? <!-- You can run `tikv-server --version` --> master with commit `4d8e469860292b10636f1b25bf7a19da84e0b067` ### What operating system and CPU are you using? <!-- If you're using Linux, you can run `cat /proc/cpuinfo` --> ### Steps to reproduce <!-- If possible, provide a recipe for reproducing the error. A complete runnable program is good. --> tidb-lightning integration test fails in create table tikv panic log as follows: ``` [2020-08-31T10:28:27.692Z] [2020/08/31 18:14:41.452 +08:00] [FATAL] [lib.rs:481] ["assertion failed: `(left == right)` left: `1`, right: `2`"] [backtrace="stack backtrace: 0: tikv_util::set_panic_hook::{{closure}} at components/tikv_util/src/lib.rs:480 1: std::panicking::rust_panic_with_hook at src/libstd/panicking.rs:524 2: rust_begin_unwind at src/libstd/panicking.rs:431 3: std::panicking::begin_panic_fmt at src/libstd/panicking.rs:385 4: tidb_query_datatype::codec::batch::lazy_column_vec::LazyBatchColumnVec::assert_columns_equal_length at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/macros.rs:16 5: <tidb_query_vec_executors::util::scan_executor::ScanExecutor<S,I> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/util/scan_executor.rs:172 <tidb_query_vec_executors::table_scan_executor::BatchTableScanExecutor<S> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/table_scan_executor.rs:110 6: <tidb_query_common::execute_stats::WithSummaryCollector<C,T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:108 7: <alloc::boxed::Box<T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:77 <tidb_query_vec_executors::selection_executor::BatchSelectionExecutor<Src> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/selection_executor.rs:179 <tidb_query_common::execute_stats::WithSummaryCollector<C,T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:108 8: <alloc::boxed::Box<T> as tidb_query_vec_executors::interface::BatchExecutor>::next_batch at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/interface.rs:77 tidb_query_vec_executors::runner::BatchExecutorsRunner<SS>::internal_handle_request at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/runner.rs:504 9: tidb_query_vec_executors::runner::BatchExecutorsRunner<SS>::handle_request::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/components/tidb_query_vec_executors/src/runner.rs:389 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <tikv::coprocessor::dag::BatchDAGHandler as tikv::coprocessor::RequestHandler>::handle_request::__handle_request::{{closure}} at src/coprocessor/dag/mod.rs:152 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <core::pin::Pin<P> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/future.rs:119 <minitrace::future::TraceWrapped<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:107 10: <core::pin::Pin<P> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/future.rs:119 <minitrace::future::TraceWrapped<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:107 <tikv::coprocessor::interceptors::tracker::Tracker<F> as core::future::future::Future>::poll at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/interceptors/tracker.rs:45 11: <tikv::coprocessor::interceptors::concurrency_limiter::ConcurrencyLimiter<PF,F> as core::future::future::Future>::poll at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/interceptors/concurrency_limiter.rs:101 tikv::coprocessor::endpoint::Endpoint<E>::handle_unary_request_impl::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/coprocessor/endpoint.rs:402 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 <minitrace::future::TraceSpawned<T> as core::future::future::Future>::poll at /rust/git/checkouts/minitrace-rust-4a91d55623f07cfc/de69110/src/future.rs:77 tikv::read_pool::ReadPoolHandle::spawn_handle::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:139 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 12: tikv::read_pool::ReadPoolHandle::spawn::{{closure}} at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:115 <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/future/mod.rs:74 13: <yatp::task::future::Runner as yatp::pool::runner::Runner>::handle at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/task/future.rs:254 14: <tikv::read_pool::ReadPoolRunner<E,R> as yatp::pool::runner::Runner>::handle at /home/jenkins/agent/workspace/build_tikv_multi_branch_master/tikv/src/read_pool.rs:169 <yatp::queue::multilevel::MultilevelRunner<R> as yatp::pool::runner::Runner>::handle at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/queue/multilevel.rs:245 yatp::pool::worker::WorkerThread<T,R>::run at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/pool/worker.rs:48 yatp::pool::builder::LazyBuilder<T>::build::{{closure}} at /rust/git/checkouts/yatp-e704b73c3ee279b6/3894a86/src/pool/builder.rs:91 std::sys_common::backtrace::__rust_begin_short_backtrace at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/sys_common/backtrace.rs:130 15: std::thread::Builder::spawn_unchecked::{{closure}}::{{closure}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/thread/mod.rs:475 <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panic.rs:318 std::panicking::try::do_call at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panicking.rs:342 std::panicking::try at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panicking.rs:319 std::panic::catch_unwind at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/panic.rs:394 std::thread::Builder::spawn_unchecked::{{closure}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/thread/mod.rs:474 core::ops::function::FnOnce::call_once{{vtable.shim}} at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libcore/ops/function.rs:233 16: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/liballoc/boxed.rs:1078 <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once at /rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/liballoc/boxed.rs:1078 std::sys::unix::thread::Thread::new::thread_start at src/libstd/sys/unix/thread.rs:87 17: start_thread 18: __clone "] [location=/rustc/16957bd4d3a5377263f76ed74c572aad8e4b7e59/src/libstd/macros.rs:16] [thread_name=unified-read-pool-7] ``` this panic can reporduce stably. for more detail log, see: https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/lightning_ghpr_test/detail/lightning_ghpr_test/1462/pipeline/50 ### What did you expect? ### What did happened?
process
cop panic by assertion failure in batchtablescanexecutor bug report what version of tikv are you using master with commit what operating system and cpu are you using steps to reproduce tidb lightning integration test fails in create table tikv panic log as follows assertion failed left right left right backtrace stack backtrace tikv util set panic hook closure at components tikv util src lib rs std panicking rust panic with hook at src libstd panicking rs rust begin unwind at src libstd panicking rs std panicking begin panic fmt at src libstd panicking rs tidb query datatype codec batch lazy column vec lazybatchcolumnvec assert columns equal length at rustc src libstd macros rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src util scan executor rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src table scan executor rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src interface rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src interface rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src selection executor rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src interface rs as tidb query vec executors interface batchexecutor next batch at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src interface rs tidb query vec executors runner batchexecutorsrunner internal handle request at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src runner rs tidb query vec executors runner batchexecutorsrunner handle request closure at home jenkins agent workspace build tikv multi branch master tikv components tidb query vec executors src runner rs as core future future future poll at rustc src libcore future mod rs handle request handle request closure at src coprocessor dag mod rs as core future future future poll at rustc src libcore future mod rs as core future future future poll at rustc src libcore future future rs as core future future future poll at rust git checkouts minitrace rust src future rs as core future future future poll at rustc src libcore future future rs as core future future future poll at rust git checkouts minitrace rust src future rs as core future future future poll at home jenkins agent workspace build tikv multi branch master tikv src coprocessor interceptors tracker rs as core future future future poll at home jenkins agent workspace build tikv multi branch master tikv src coprocessor interceptors concurrency limiter rs tikv coprocessor endpoint endpoint handle unary request impl closure at home jenkins agent workspace build tikv multi branch master tikv src coprocessor endpoint rs as core future future future poll at rustc src libcore future mod rs as core future future future poll at rust git checkouts minitrace rust src future rs tikv read pool readpoolhandle spawn handle closure at home jenkins agent workspace build tikv multi branch master tikv src read pool rs as core future future future poll at rustc src libcore future mod rs tikv read pool readpoolhandle spawn closure at home jenkins agent workspace build tikv multi branch master tikv src read pool rs as core future future future poll at rustc src libcore future mod rs handle at rust git checkouts yatp src task future rs as yatp pool runner runner handle at home jenkins agent workspace build tikv multi branch master tikv src read pool rs as yatp pool runner runner handle at rust git checkouts yatp src queue multilevel rs yatp pool worker workerthread run at rust git checkouts yatp src pool worker rs yatp pool builder lazybuilder build closure at rust git checkouts yatp src pool builder rs std sys common backtrace rust begin short backtrace at rustc src libstd sys common backtrace rs std thread builder spawn unchecked closure closure at rustc src libstd thread mod rs as core ops function fnonce call once at rustc src libstd panic rs std panicking try do call at rustc src libstd panicking rs std panicking try at rustc src libstd panicking rs std panic catch unwind at rustc src libstd panic rs std thread builder spawn unchecked closure at rustc src libstd thread mod rs core ops function fnonce call once vtable shim at rustc src libcore ops function rs as core ops function fnonce call once at rustc src liballoc boxed rs as core ops function fnonce call once at rustc src liballoc boxed rs std sys unix thread thread new thread start at src libstd sys unix thread rs start thread clone this panic can reporduce stably for more detail log see what did you expect what did happened
1
12,685
15,050,188,545
IssuesEvent
2021-02-03 12:32:16
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
reopened
Zonal statistics not working when raster and vector in different CRS
Bug Processing Regression
Author Name: **matteo ghetta** (@ghtmtt) Original Redmine Issue: [19027](https://issues.qgis.org/issues/19027) Affected QGIS version: 3.1(master) Redmine category:processing/qgis --- If the CRS is different the algorithm runs but the result is empty
1.0
Zonal statistics not working when raster and vector in different CRS - Author Name: **matteo ghetta** (@ghtmtt) Original Redmine Issue: [19027](https://issues.qgis.org/issues/19027) Affected QGIS version: 3.1(master) Redmine category:processing/qgis --- If the CRS is different the algorithm runs but the result is empty
process
zonal statistics not working when raster and vector in different crs author name matteo ghetta ghtmtt original redmine issue affected qgis version master redmine category processing qgis if the crs is different the algorithm runs but the result is empty
1
207,849
7,134,209,546
IssuesEvent
2018-01-22 20:02:59
particle-iot/firmware
https://api.github.com/repos/particle-iot/firmware
closed
Publishing in a multi-thread application running on UDP device during OTA results in hard fault
PR SUBMITTED bug priority track
``` Given I have a UDP device running application firmware with multithreading enabled And the application firmware publishes events to the cloud at regular intervals When I OTA flash the device with another application firmware And the device publishes an event while the OTA is in progress Then the OTA will be irreparably disrupted and the UDP device will hard fault until a manual reset ``` Application firmware running on the Electron was: ``` // This #include statement was automatically added by the Particle IDE. #include <google-maps-device-locator.h> GoogleMapsDeviceLocator locator; SerialLogHandler logHandler; SYSTEM_THREAD(ENABLED); void setup() { Serial.begin(9600); locator.withLocatePeriodic(10); } void loop() { locator.loop(); } ``` _when_ this firmware is running on the device, _and_ I OTA a new application firmware to the electron, the OTA fails mid-way through and my electron starts hard faulting. Serial logs for the device during the OTA: ``` 0000218892 [comm.protocol] INFO: rcv'd message type=13 0000219614 [comm.protocol] INFO: rcv'd message type=13 0000226230 [comm.protocol] INFO: rcv'd message type=5 0000226484 [comm.protocol] INFO: rcv'd message type=7 0000226738 [comm.protocol] INFO: rcv'd message type=7 0000226992 [comm.protocol] INFO: rcv'd message type=7 0000227246 [comm.protocol] INFO: rcv'd message type=7 0000227500 [comm.protocol] INFO: rcv'd message type=7 publishLocation 0000228099 [comm] WARN: handle received message failed - aborting transfer 0000228100 [system] INFO: Send spark/device/ota_result event 0000229100 [comm] WARN: handle received message failed - aborting transfer 0000229100 [system] INFO: Send spark/device/ota_result event 0000230101 [comm] WARN: handle received message failed - aborting transfer 0000230101 [system] INFO: Send spark/device/ota_result event 0000231102 [comm] WARN: handle received message failed - aborting transfer ``` So looks like to me the device is receiving UDP packets, but for some reason, the application code on the device is still running while receiving the OTA. The `publishLocation` is logged every time network data is sent up to the cloud. Once this happened, the following system messages suggest an issue, and the OTA is aborted. Then the device starts hard-faulting (blinking red). Note that this is only observable on an Electron (UDP) not Photon (TCP). Running the same experiment on a Photon produces the following device logs: ``` 0000055938 [comm.sparkprotocol] INFO: chunk idx=22 crc=1 fast=1 updating=1 publishLocation 0000056043 [comm.sparkprotocol] INFO: chunk 0000056043 [comm.sparkprotocol] INFO: chunk idx=23 crc=1 fast=1 updating=1 0000056146 [comm.sparkprotocol] INFO: chunk 0000056146 [comm.sparkprotocol] INFO: chunk idx=24 crc=1 fast=1 updating=1 0000056249 [comm.sparkprotocol] INFO: chunk 0000056249 [comm.sparkprotocol] INFO: chunk idx=25 crc=1 fast=1 updating=1 0000056352 [comm.sparkprotocol] INFO: chunk 0000056352 [comm.sparkprotocol] INFO: chunk idx=26 crc=1 fast=1 updating=1 0000056455 [comm.sparkprotocol] INFO: chunk ```
1.0
Publishing in a multi-thread application running on UDP device during OTA results in hard fault - ``` Given I have a UDP device running application firmware with multithreading enabled And the application firmware publishes events to the cloud at regular intervals When I OTA flash the device with another application firmware And the device publishes an event while the OTA is in progress Then the OTA will be irreparably disrupted and the UDP device will hard fault until a manual reset ``` Application firmware running on the Electron was: ``` // This #include statement was automatically added by the Particle IDE. #include <google-maps-device-locator.h> GoogleMapsDeviceLocator locator; SerialLogHandler logHandler; SYSTEM_THREAD(ENABLED); void setup() { Serial.begin(9600); locator.withLocatePeriodic(10); } void loop() { locator.loop(); } ``` _when_ this firmware is running on the device, _and_ I OTA a new application firmware to the electron, the OTA fails mid-way through and my electron starts hard faulting. Serial logs for the device during the OTA: ``` 0000218892 [comm.protocol] INFO: rcv'd message type=13 0000219614 [comm.protocol] INFO: rcv'd message type=13 0000226230 [comm.protocol] INFO: rcv'd message type=5 0000226484 [comm.protocol] INFO: rcv'd message type=7 0000226738 [comm.protocol] INFO: rcv'd message type=7 0000226992 [comm.protocol] INFO: rcv'd message type=7 0000227246 [comm.protocol] INFO: rcv'd message type=7 0000227500 [comm.protocol] INFO: rcv'd message type=7 publishLocation 0000228099 [comm] WARN: handle received message failed - aborting transfer 0000228100 [system] INFO: Send spark/device/ota_result event 0000229100 [comm] WARN: handle received message failed - aborting transfer 0000229100 [system] INFO: Send spark/device/ota_result event 0000230101 [comm] WARN: handle received message failed - aborting transfer 0000230101 [system] INFO: Send spark/device/ota_result event 0000231102 [comm] WARN: handle received message failed - aborting transfer ``` So looks like to me the device is receiving UDP packets, but for some reason, the application code on the device is still running while receiving the OTA. The `publishLocation` is logged every time network data is sent up to the cloud. Once this happened, the following system messages suggest an issue, and the OTA is aborted. Then the device starts hard-faulting (blinking red). Note that this is only observable on an Electron (UDP) not Photon (TCP). Running the same experiment on a Photon produces the following device logs: ``` 0000055938 [comm.sparkprotocol] INFO: chunk idx=22 crc=1 fast=1 updating=1 publishLocation 0000056043 [comm.sparkprotocol] INFO: chunk 0000056043 [comm.sparkprotocol] INFO: chunk idx=23 crc=1 fast=1 updating=1 0000056146 [comm.sparkprotocol] INFO: chunk 0000056146 [comm.sparkprotocol] INFO: chunk idx=24 crc=1 fast=1 updating=1 0000056249 [comm.sparkprotocol] INFO: chunk 0000056249 [comm.sparkprotocol] INFO: chunk idx=25 crc=1 fast=1 updating=1 0000056352 [comm.sparkprotocol] INFO: chunk 0000056352 [comm.sparkprotocol] INFO: chunk idx=26 crc=1 fast=1 updating=1 0000056455 [comm.sparkprotocol] INFO: chunk ```
non_process
publishing in a multi thread application running on udp device during ota results in hard fault given i have a udp device running application firmware with multithreading enabled and the application firmware publishes events to the cloud at regular intervals when i ota flash the device with another application firmware and the device publishes an event while the ota is in progress then the ota will be irreparably disrupted and the udp device will hard fault until a manual reset application firmware running on the electron was this include statement was automatically added by the particle ide include googlemapsdevicelocator locator serialloghandler loghandler system thread enabled void setup serial begin locator withlocateperiodic void loop locator loop when this firmware is running on the device and i ota a new application firmware to the electron the ota fails mid way through and my electron starts hard faulting serial logs for the device during the ota info rcv d message type info rcv d message type info rcv d message type info rcv d message type info rcv d message type info rcv d message type info rcv d message type info rcv d message type publishlocation warn handle received message failed aborting transfer info send spark device ota result event warn handle received message failed aborting transfer info send spark device ota result event warn handle received message failed aborting transfer info send spark device ota result event warn handle received message failed aborting transfer so looks like to me the device is receiving udp packets but for some reason the application code on the device is still running while receiving the ota the publishlocation is logged every time network data is sent up to the cloud once this happened the following system messages suggest an issue and the ota is aborted then the device starts hard faulting blinking red note that this is only observable on an electron udp not photon tcp running the same experiment on a photon produces the following device logs info chunk idx crc fast updating publishlocation info chunk info chunk idx crc fast updating info chunk info chunk idx crc fast updating info chunk info chunk idx crc fast updating info chunk info chunk idx crc fast updating info chunk
0
259,821
27,728,652,336
IssuesEvent
2023-03-15 05:47:09
vasind/try-ember
https://api.github.com/repos/vasind/try-ember
closed
CVE-2022-21670 (Medium) detected in markdown-it-8.4.2.tgz
security vulnerability no-issue-activity
## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-8.4.2.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - ember-cli-3.7.1.tgz (Root Library) - :x: **markdown-it-8.4.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: markdown-it - 12.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21670 (Medium) detected in markdown-it-8.4.2.tgz - ## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-8.4.2.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - ember-cli-3.7.1.tgz (Root Library) - :x: **markdown-it-8.4.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: markdown-it - 12.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in markdown it tgz cve medium severity vulnerability vulnerable library markdown it tgz markdown it modern pluggable markdown parser library home page a href path to dependency file package json path to vulnerable library node modules markdown it package json dependency hierarchy ember cli tgz root library x markdown it tgz vulnerable library vulnerability details markdown it is a markdown parser prior to version special patterns with length greater than thousand characterss could slow down the parser significantly users should upgrade to version to receive a patch there are no known workarounds aside from upgrading publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution markdown it step up your open source security game with mend
0
385,338
11,418,829,902
IssuesEvent
2020-02-03 06:07:59
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Error occurred when manually sourcing the Oracle sql script
Priority/Highest Severity/Blocker Type/Bug
Enviornment DB - Oracle 12. wso2is-5.9.0-alpha pack Priority: High Severity: Medium Steps: Took the latest oracle script Source that oracle (/wso2is-5.9.0-alpha/dbscripts) **Error** Error starting at line : 783 in command - CREATE INDEX SYSTEM_ROLE_IND_BY_RN_TI ON UM_SYSTEM_ROLE(UM_ROLE_NAME, UM_TENANT_ID) Error report - ORA-01408: such column list already indexed 01408. 00000 - "such column list already indexed" *Cause: *Action: Error starting at line : 783 in command - CREATE INDEX REG_PATH_IND_BY_PATH_VALUE ON REG_PATH(REG_PATH_VALUE, REG_TENANT_ID) Error report - ORA-01408: such column list already indexed 01408. 00000 - "such column list already indexed" *Cause: *Action:
1.0
Error occurred when manually sourcing the Oracle sql script - Enviornment DB - Oracle 12. wso2is-5.9.0-alpha pack Priority: High Severity: Medium Steps: Took the latest oracle script Source that oracle (/wso2is-5.9.0-alpha/dbscripts) **Error** Error starting at line : 783 in command - CREATE INDEX SYSTEM_ROLE_IND_BY_RN_TI ON UM_SYSTEM_ROLE(UM_ROLE_NAME, UM_TENANT_ID) Error report - ORA-01408: such column list already indexed 01408. 00000 - "such column list already indexed" *Cause: *Action: Error starting at line : 783 in command - CREATE INDEX REG_PATH_IND_BY_PATH_VALUE ON REG_PATH(REG_PATH_VALUE, REG_TENANT_ID) Error report - ORA-01408: such column list already indexed 01408. 00000 - "such column list already indexed" *Cause: *Action:
non_process
error occurred when manually sourcing the oracle sql script enviornment db oracle alpha pack priority high severity medium steps took the latest oracle script source that oracle alpha dbscripts error error starting at line in command create index system role ind by rn ti on um system role um role name um tenant id error report ora such column list already indexed such column list already indexed cause action error starting at line in command create index reg path ind by path value on reg path reg path value reg tenant id error report ora such column list already indexed such column list already indexed cause action
0
285,049
24,639,576,199
IssuesEvent
2022-10-17 10:32:08
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
scaletesting: Update Sourcegraph to 4.0.1
devx/q3b1/scaletesting/devops
``` ✅ Fetched deployed version ✅ Done updating list of commits 💡 Live on "scaletesting": e18d1344e0ee (built on 2022-09-23) ``` Absolute latest version is v4.0.1
1.0
scaletesting: Update Sourcegraph to 4.0.1 - ``` ✅ Fetched deployed version ✅ Done updating list of commits 💡 Live on "scaletesting": e18d1344e0ee (built on 2022-09-23) ``` Absolute latest version is v4.0.1
non_process
scaletesting update sourcegraph to ✅ fetched deployed version ✅ done updating list of commits 💡 live on scaletesting built on absolute latest version is
0
133,142
18,279,651,171
IssuesEvent
2021-10-05 00:24:28
ghc-dev/Thomas-Stone
https://api.github.com/repos/ghc-dev/Thomas-Stone
opened
CVE-2020-1738 (Low) detected in ansible-2.9.9.tar.gz
security vulnerability
## CVE-2020-1738 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Thomas-Stone/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Thomas-Stone/commit/05a2c8bc9363030191fdc3188014e1d8aa8a4ec0">05a2c8bc9363030191fdc3188014e1d8aa8a4ec0</a></p> <p>Found in base branch: <b>feature_branch</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Ansible Engine when the module package or service is used and the parameter 'use' is not specified. If a previous task is executed with a malicious user, the module sent can be selected by the attacker using the ansible facts file. All versions in 2.7.x, 2.8.x and 2.9.x branches are believed to be vulnerable. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1738>CVE-2020-1738</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-1738">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-1738</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: ansible-engine 2.9.7</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansible-engine 2.9.7"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-1738","vulnerabilityDetails":"A flaw was found in Ansible Engine when the module package or service is used and the parameter \u0027use\u0027 is not specified. If a previous task is executed with a malicious user, the module sent can be selected by the attacker using the ansible facts file. All versions in 2.7.x, 2.8.x and 2.9.x branches are believed to be vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1738","cvss3Severity":"low","cvss3Score":"3.9","cvss3Metrics":{"A":"Low","AC":"High","PR":"Low","S":"Changed","C":"None","UI":"Required","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-1738 (Low) detected in ansible-2.9.9.tar.gz - ## CVE-2020-1738 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Thomas-Stone/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Thomas-Stone/commit/05a2c8bc9363030191fdc3188014e1d8aa8a4ec0">05a2c8bc9363030191fdc3188014e1d8aa8a4ec0</a></p> <p>Found in base branch: <b>feature_branch</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Ansible Engine when the module package or service is used and the parameter 'use' is not specified. If a previous task is executed with a malicious user, the module sent can be selected by the attacker using the ansible facts file. All versions in 2.7.x, 2.8.x and 2.9.x branches are believed to be vulnerable. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1738>CVE-2020-1738</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-1738">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-1738</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: ansible-engine 2.9.7</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansible-engine 2.9.7"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-1738","vulnerabilityDetails":"A flaw was found in Ansible Engine when the module package or service is used and the parameter \u0027use\u0027 is not specified. If a previous task is executed with a malicious user, the module sent can be selected by the attacker using the ansible facts file. All versions in 2.7.x, 2.8.x and 2.9.x branches are believed to be vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1738","cvss3Severity":"low","cvss3Score":"3.9","cvss3Metrics":{"A":"Low","AC":"High","PR":"Low","S":"Changed","C":"None","UI":"Required","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve low detected in ansible tar gz cve low severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file thomas stone requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library found in head commit a href found in base branch feature branch vulnerability details a flaw was found in ansible engine when the module package or service is used and the parameter use is not specified if a previous task is executed with a malicious user the module sent can be selected by the attacker using the ansible facts file all versions in x x and x branches are believed to be vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction required scope changed impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansible engine rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ansible isminimumfixversionavailable true minimumfixversion ansible engine basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was found in ansible engine when the module package or service is used and the parameter is not specified if a previous task is executed with a malicious user the module sent can be selected by the attacker using the ansible facts file all versions in x x and x branches are believed to be vulnerable vulnerabilityurl
0
16,868
22,148,945,668
IssuesEvent
2022-06-03 14:48:51
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Bump main to 0.2.0-SNAPSHOT
enhancement P1 process
### Problem 0.1.0-rc1 has been branched off of main ### Solution Bump main to 0.2.0-SNAPSHOT ### Alternatives _No response_
1.0
Bump main to 0.2.0-SNAPSHOT - ### Problem 0.1.0-rc1 has been branched off of main ### Solution Bump main to 0.2.0-SNAPSHOT ### Alternatives _No response_
process
bump main to snapshot problem has been branched off of main solution bump main to snapshot alternatives no response
1
186,144
6,733,792,695
IssuesEvent
2017-10-18 15:50:49
EFForg/privacybadger
https://api.github.com/repos/EFForg/privacybadger
closed
Option to cookieblock trackers by default
enhancement heuristic low priority privacy yellowlist
It might be a nice feature to be able to have the default setting for detected trackers to cookie block them instead of allowing them until we see them three times. This should be off by default but have a setting in the options menu.
1.0
Option to cookieblock trackers by default - It might be a nice feature to be able to have the default setting for detected trackers to cookie block them instead of allowing them until we see them three times. This should be off by default but have a setting in the options menu.
non_process
option to cookieblock trackers by default it might be a nice feature to be able to have the default setting for detected trackers to cookie block them instead of allowing them until we see them three times this should be off by default but have a setting in the options menu
0
386,787
26,701,442,039
IssuesEvent
2023-01-27 14:42:22
insightsengineering/tern
https://api.github.com/repos/insightsengineering/tern
closed
Incorrect reference list in the tern pkgdown
bug documentation sme
https://bookish-invention-97673167.pages.github.io/reference/index.html#statistics-functions The one family functions (prefix with h_, a_, s_) are duplicated across different subsections, the structure is highly misleading. This comes from roxygen2 `describeIn` functionality as influence the pkgdown.yaml. It looks like there is no control for that from the perspective of pkgdown. https://github.com/r-lib/pkgdown/issues/986 One of the solutions is to rewrite the pkgdown yaml file to assume all siblings are in the one place. The other solution is to rewrite the roxygen2 which does not looks to be a right one. Or `man-roxygen` ... Or simplify the reference and leave only the analyze functions (end product) Ps. if updating pkgdown yaml please use `- starts_with("a_")` syntax to make code more elegant and easier to maintain.
1.0
Incorrect reference list in the tern pkgdown - https://bookish-invention-97673167.pages.github.io/reference/index.html#statistics-functions The one family functions (prefix with h_, a_, s_) are duplicated across different subsections, the structure is highly misleading. This comes from roxygen2 `describeIn` functionality as influence the pkgdown.yaml. It looks like there is no control for that from the perspective of pkgdown. https://github.com/r-lib/pkgdown/issues/986 One of the solutions is to rewrite the pkgdown yaml file to assume all siblings are in the one place. The other solution is to rewrite the roxygen2 which does not looks to be a right one. Or `man-roxygen` ... Or simplify the reference and leave only the analyze functions (end product) Ps. if updating pkgdown yaml please use `- starts_with("a_")` syntax to make code more elegant and easier to maintain.
non_process
incorrect reference list in the tern pkgdown the one family functions prefix with h a s are duplicated across different subsections the structure is highly misleading this comes from describein functionality as influence the pkgdown yaml it looks like there is no control for that from the perspective of pkgdown one of the solutions is to rewrite the pkgdown yaml file to assume all siblings are in the one place the other solution is to rewrite the which does not looks to be a right one or man roxygen or simplify the reference and leave only the analyze functions end product ps if updating pkgdown yaml please use starts with a syntax to make code more elegant and easier to maintain
0
6,412
9,498,290,756
IssuesEvent
2019-04-24 01:09:58
googleapis/nodejs-firestore-session
https://api.github.com/repos/googleapis/nodejs-firestore-session
closed
Add a `.repo-metadata.json`
release blocking type: process
Let's add that file (see https://github.com/googleapis/nodejs-storage/blob/master/.repo-metadata.json for an example) and use that to generate the README.
1.0
Add a `.repo-metadata.json` - Let's add that file (see https://github.com/googleapis/nodejs-storage/blob/master/.repo-metadata.json for an example) and use that to generate the README.
process
add a repo metadata json let s add that file see for an example and use that to generate the readme
1
433,622
12,507,880,512
IssuesEvent
2020-06-02 14:45:35
GeyserMC/Geyser
https://api.github.com/repos/GeyserMC/Geyser
closed
Geyser Packet Errors
Confirmed Bug Priority: Low
**Describe the bug** Geyser spamming error when join the server. Note : This issue only occurs when geyser is run alone. If run in bungeecord or another server, not giving this error. **Screenshots / Videos** ![image](https://user-images.githubusercontent.com/47111280/81970042-940d1e00-9627-11ea-913c-889284a34e4e.png) ![image](https://user-images.githubusercontent.com/47111280/81970026-8e173d00-9627-11ea-9b61-d7606e6ab472.png) ![image](https://user-images.githubusercontent.com/47111280/81970072-9ec7b300-9627-11ea-89e9-f5597f1b2388.png) ![image](https://user-images.githubusercontent.com/47111280/81970175-c1f26280-9627-11ea-9525-6677d3e75029.png) **Server Version** ![image](https://user-images.githubusercontent.com/47111280/81970257-e2bab800-9627-11ea-835d-ac58f8c450a8.png) Geyser Version Dev build #148 Minecraft: Bedrock Edition Version 1.14.60 Server ip: Oyna.SonCesurlar.Com (its java version ip)
1.0
Geyser Packet Errors - **Describe the bug** Geyser spamming error when join the server. Note : This issue only occurs when geyser is run alone. If run in bungeecord or another server, not giving this error. **Screenshots / Videos** ![image](https://user-images.githubusercontent.com/47111280/81970042-940d1e00-9627-11ea-913c-889284a34e4e.png) ![image](https://user-images.githubusercontent.com/47111280/81970026-8e173d00-9627-11ea-9b61-d7606e6ab472.png) ![image](https://user-images.githubusercontent.com/47111280/81970072-9ec7b300-9627-11ea-89e9-f5597f1b2388.png) ![image](https://user-images.githubusercontent.com/47111280/81970175-c1f26280-9627-11ea-9525-6677d3e75029.png) **Server Version** ![image](https://user-images.githubusercontent.com/47111280/81970257-e2bab800-9627-11ea-835d-ac58f8c450a8.png) Geyser Version Dev build #148 Minecraft: Bedrock Edition Version 1.14.60 Server ip: Oyna.SonCesurlar.Com (its java version ip)
non_process
geyser packet errors describe the bug geyser spamming error when join the server note this issue only occurs when geyser is run alone if run in bungeecord or another server not giving this error screenshots videos server version geyser version dev build minecraft bedrock edition version server ip oyna soncesurlar com its java version ip
0
234
2,663,098,054
IssuesEvent
2015-03-20 00:56:11
niranjv/ucscamsms2015
https://api.github.com/repos/niranjv/ucscamsms2015
opened
Use knitr chunk references for code snippets to generate plots & tables
doc: 2-Single Processor doc: 3-Multi Processor doc: 6-Appendix Project Setup
* Various plots are used in Chapters 2 and 3 * Code for these plots must be shown in Appendix B * Keep the code in a separate R file (`bin/plot.r`) in labelled chunks * Run this code via knitr in the main text * Display the same code in the Appendix * Use knitr chunk references to display the necessary code * This will ensure that the code in the appendix is valid and was used to generate the plots in the main text
2.0
Use knitr chunk references for code snippets to generate plots & tables - * Various plots are used in Chapters 2 and 3 * Code for these plots must be shown in Appendix B * Keep the code in a separate R file (`bin/plot.r`) in labelled chunks * Run this code via knitr in the main text * Display the same code in the Appendix * Use knitr chunk references to display the necessary code * This will ensure that the code in the appendix is valid and was used to generate the plots in the main text
process
use knitr chunk references for code snippets to generate plots tables various plots are used in chapters and code for these plots must be shown in appendix b keep the code in a separate r file bin plot r in labelled chunks run this code via knitr in the main text display the same code in the appendix use knitr chunk references to display the necessary code this will ensure that the code in the appendix is valid and was used to generate the plots in the main text
1
79,357
28,125,554,011
IssuesEvent
2023-03-31 17:23:21
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
ZFS "drowns" Linux kernel audit
Type: Defect
### System information Gentoo Linux AMD64 using Linux kernel 6.2.8 and ZFS 2.1.9 (GCC 12.2.1) CPU is an AMD 7950X (32 logical cores) ### Describe the problem you're observing When "emerging" (i.e. compiling) _some_ packages like net-fs/samba (4.18.0 but any 4.x version should trigger the issue) __**AND**__ auditd is running, osq_lock() goes into the ceiling as shown by a **perf top** and gcc becomes as slow as a snail: ``` 76.76% [kernel] [k] osq_lock 2.10% [kernel] [k] mutex_spin_on_owner 1.41% [kernel] [k] native_sched_clock 1.30% [kernel] [k] find_busiest_group 0.74% [kernel] [k] psi_group_change 0.54% [kernel] [k] _raw_spin_lock_irqsave 0.52% [kernel] [k] __schedule 0.51% [kernel] [k] update_rq_clock.part.0 0.48% [kernel] [k] __update_load_avg_cfs_rq 0.43% [kernel] [k] update_load_avg ``` With this issue happening, recompiling Samba 4 takes more than 90 minutes (normal build time is around 10 minutes). If I flush the auditd ruleset or if I stop auditd this behavior does not happen, the same if I put my stage 4 Gentoo environment on a non-ZFS stuff (e.g. a XFS filesystem). Some preliminary random thoughs: ``` # auditctl -l (by default as provided by sys-process/audit) -a never,exit -F arch=b32 -S read,write,open,close,brk,fcntl,dup2,mmap,munmap,stat,fstat,nanosleep,rt_sigaction -a never,exit -F arch=b64 -S read,write,open,close,stat,fstat,mmap,munmap,brk,rt_sigaction,dup2,nanosleep,fcntl # bpftrace -e 'kprobe:osq_lock { @[arg0] = count(); }' Attaching 1 probe... ^C @[-113557655720192]: 1 @[-113461265342996]: 1 @[-113473551484556]: 1 @[-113557664708716]: 1 (...) @[-113557216933532]: 6423 @[-113557216934140]: 8925 @[-113557216933836]: 11901 @[-113557216935052]: 14951 @[-113457054216836]: 90877916 <=== OUTCH! # bpftrace -e 'kprobe:osq_lock /arg0 == -113457054216836/ { @[kstack(10)] = count() }' Attaching 1 probe... ^C (...) @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_dirent_lock+189 zfs_get_xattrdir+222 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 ]: 847095 @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_dirent_unlock+32 zfs_get_xattrdir+291 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 ]: 982700 @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_getattr_fast+104 zio_do_crypt_abd+10778 vfs_statx+160 vfs_fstatat+81 __do_sys_newfstatat+19 do_syscall_64+58 entry_SYSCALL_64_after_hwframe+99 ]: 1005841 @[ osq_lock+1 __mutex_lock.constprop.0+369 zpl_prune_sb+153 zpl_prune_sb+3243 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 audit_copy_inode+123 filename_lookup+339 vfs_statx+124 ]: 1047785 @[ osq_lock+1 __mutex_lock.constprop.0+369 __cv_broadcast+231 zfs_dirent_lock+348 zfs_get_xattrdir+222 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 ]: 26117572 ``` The common point in those most frequent kernel call stack frames: some ZFS system calls. For those who are unfamiliar, the last one is the one that happens the most (over 2.6 million times in a couple of seconds). ### Describe how to reproduce the problem 1. Put a Gentoo stage 3 (or 4) on a ZFS data set (e.g zpool1/mystage4) 2. Mount everything required (mount -o bind /proc /zpool1/mystage4/proc, mount -o bind /sys /zpool1/mystage4/sys and so on) 3. chroot there (chroot /zpool1/mystage4 /bin/bash) 4. env-update 5. emerge net-fs/samba Did not try (yet) but installing a common Linux distribution on a ZFS pool and recompiling Samba 4.x should trigger the issue.
1.0
ZFS "drowns" Linux kernel audit - ### System information Gentoo Linux AMD64 using Linux kernel 6.2.8 and ZFS 2.1.9 (GCC 12.2.1) CPU is an AMD 7950X (32 logical cores) ### Describe the problem you're observing When "emerging" (i.e. compiling) _some_ packages like net-fs/samba (4.18.0 but any 4.x version should trigger the issue) __**AND**__ auditd is running, osq_lock() goes into the ceiling as shown by a **perf top** and gcc becomes as slow as a snail: ``` 76.76% [kernel] [k] osq_lock 2.10% [kernel] [k] mutex_spin_on_owner 1.41% [kernel] [k] native_sched_clock 1.30% [kernel] [k] find_busiest_group 0.74% [kernel] [k] psi_group_change 0.54% [kernel] [k] _raw_spin_lock_irqsave 0.52% [kernel] [k] __schedule 0.51% [kernel] [k] update_rq_clock.part.0 0.48% [kernel] [k] __update_load_avg_cfs_rq 0.43% [kernel] [k] update_load_avg ``` With this issue happening, recompiling Samba 4 takes more than 90 minutes (normal build time is around 10 minutes). If I flush the auditd ruleset or if I stop auditd this behavior does not happen, the same if I put my stage 4 Gentoo environment on a non-ZFS stuff (e.g. a XFS filesystem). Some preliminary random thoughs: ``` # auditctl -l (by default as provided by sys-process/audit) -a never,exit -F arch=b32 -S read,write,open,close,brk,fcntl,dup2,mmap,munmap,stat,fstat,nanosleep,rt_sigaction -a never,exit -F arch=b64 -S read,write,open,close,stat,fstat,mmap,munmap,brk,rt_sigaction,dup2,nanosleep,fcntl # bpftrace -e 'kprobe:osq_lock { @[arg0] = count(); }' Attaching 1 probe... ^C @[-113557655720192]: 1 @[-113461265342996]: 1 @[-113473551484556]: 1 @[-113557664708716]: 1 (...) @[-113557216933532]: 6423 @[-113557216934140]: 8925 @[-113557216933836]: 11901 @[-113557216935052]: 14951 @[-113457054216836]: 90877916 <=== OUTCH! # bpftrace -e 'kprobe:osq_lock /arg0 == -113457054216836/ { @[kstack(10)] = count() }' Attaching 1 probe... ^C (...) @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_dirent_lock+189 zfs_get_xattrdir+222 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 ]: 847095 @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_dirent_unlock+32 zfs_get_xattrdir+291 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 ]: 982700 @[ osq_lock+1 __mutex_lock.constprop.0+369 zfs_getattr_fast+104 zio_do_crypt_abd+10778 vfs_statx+160 vfs_fstatat+81 __do_sys_newfstatat+19 do_syscall_64+58 entry_SYSCALL_64_after_hwframe+99 ]: 1005841 @[ osq_lock+1 __mutex_lock.constprop.0+369 zpl_prune_sb+153 zpl_prune_sb+3243 zpl_prune_sb+3518 __vfs_getxattr+126 get_vfs_caps_from_disk+96 audit_copy_inode+123 filename_lookup+339 vfs_statx+124 ]: 1047785 @[ osq_lock+1 __mutex_lock.constprop.0+369 __cv_broadcast+231 zfs_dirent_lock+348 zfs_get_xattrdir+222 zfs_lookup+226 zpl_prune_sb+475 zpl_prune_sb+3105 zpl_prune_sb+3518 __vfs_getxattr+126 ]: 26117572 ``` The common point in those most frequent kernel call stack frames: some ZFS system calls. For those who are unfamiliar, the last one is the one that happens the most (over 2.6 million times in a couple of seconds). ### Describe how to reproduce the problem 1. Put a Gentoo stage 3 (or 4) on a ZFS data set (e.g zpool1/mystage4) 2. Mount everything required (mount -o bind /proc /zpool1/mystage4/proc, mount -o bind /sys /zpool1/mystage4/sys and so on) 3. chroot there (chroot /zpool1/mystage4 /bin/bash) 4. env-update 5. emerge net-fs/samba Did not try (yet) but installing a common Linux distribution on a ZFS pool and recompiling Samba 4.x should trigger the issue.
non_process
zfs drowns linux kernel audit system information gentoo linux using linux kernel and zfs gcc cpu is an amd logical cores describe the problem you re observing when emerging i e compiling some packages like net fs samba but any x version should trigger the issue and auditd is running osq lock goes into the ceiling as shown by a perf top and gcc becomes as slow as a snail osq lock mutex spin on owner native sched clock find busiest group psi group change raw spin lock irqsave schedule update rq clock part update load avg cfs rq update load avg with this issue happening recompiling samba takes more than minutes normal build time is around minutes if i flush the auditd ruleset or if i stop auditd this behavior does not happen the same if i put my stage gentoo environment on a non zfs stuff e g a xfs filesystem some preliminary random thoughs auditctl l by default as provided by sys process audit a never exit f arch s read write open close brk fcntl mmap munmap stat fstat nanosleep rt sigaction a never exit f arch s read write open close stat fstat mmap munmap brk rt sigaction nanosleep fcntl bpftrace e kprobe osq lock count attaching probe c outch bpftrace e kprobe osq lock count attaching probe c osq lock mutex lock constprop zfs dirent lock zfs get xattrdir zfs lookup zpl prune sb zpl prune sb zpl prune sb vfs getxattr get vfs caps from disk osq lock mutex lock constprop zfs dirent unlock zfs get xattrdir zfs lookup zpl prune sb zpl prune sb zpl prune sb vfs getxattr get vfs caps from disk osq lock mutex lock constprop zfs getattr fast zio do crypt abd vfs statx vfs fstatat do sys newfstatat do syscall entry syscall after hwframe osq lock mutex lock constprop zpl prune sb zpl prune sb zpl prune sb vfs getxattr get vfs caps from disk audit copy inode filename lookup vfs statx osq lock mutex lock constprop cv broadcast zfs dirent lock zfs get xattrdir zfs lookup zpl prune sb zpl prune sb zpl prune sb vfs getxattr the common point in those most frequent kernel call stack frames some zfs system calls for those who are unfamiliar the last one is the one that happens the most over million times in a couple of seconds describe how to reproduce the problem put a gentoo stage or on a zfs data set e g mount everything required mount o bind proc proc mount o bind sys sys and so on chroot there chroot bin bash env update emerge net fs samba did not try yet but installing a common linux distribution on a zfs pool and recompiling samba x should trigger the issue
0
79,578
28,433,087,831
IssuesEvent
2023-04-15 01:59:33
zealdocs/zeal
https://api.github.com/repos/zealdocs/zeal
closed
I used the doc2dash tool to generate the Polars document, but its content was not displayed properly
type/defect resolution/awaiting-response scope/ui/webview
Problem description All the content is compressed into a small scroll bar area, as shown below: ![image](https://user-images.githubusercontent.com/29591339/231759433-3a6eec99-86b8-4127-a970-1743f468aaa2.png) expect work: https://pola-rs.github.io/polars/py-polars/html/reference/ requirement: numpy pandas pyarrow hypothesis==6.71.0 autodocsumm==0.2.9 commonmark==0.9.1 numpydoc==1.5.0 pydata-sphinx-theme==0.13.0 sphinx-autosummary-accessors==2022.4.0 sphinx-copybutton==0.5.1 sphinx-design==0.3.0 sphinx==5.3.0 sphinxcontrib-applehelp==1.0.2 sphinxcontrib-devhelp==1.0.2 sphinxcontrib-htmlhelp==2.0.0 sphinxcontrib-jsmath==1.0.1 sphinxcontrib-napoleon==0.7 sphinxcontrib-qthelp==1.0.3 sphinxcontrib-serializinghtml==1.1.5
1.0
I used the doc2dash tool to generate the Polars document, but its content was not displayed properly - Problem description All the content is compressed into a small scroll bar area, as shown below: ![image](https://user-images.githubusercontent.com/29591339/231759433-3a6eec99-86b8-4127-a970-1743f468aaa2.png) expect work: https://pola-rs.github.io/polars/py-polars/html/reference/ requirement: numpy pandas pyarrow hypothesis==6.71.0 autodocsumm==0.2.9 commonmark==0.9.1 numpydoc==1.5.0 pydata-sphinx-theme==0.13.0 sphinx-autosummary-accessors==2022.4.0 sphinx-copybutton==0.5.1 sphinx-design==0.3.0 sphinx==5.3.0 sphinxcontrib-applehelp==1.0.2 sphinxcontrib-devhelp==1.0.2 sphinxcontrib-htmlhelp==2.0.0 sphinxcontrib-jsmath==1.0.1 sphinxcontrib-napoleon==0.7 sphinxcontrib-qthelp==1.0.3 sphinxcontrib-serializinghtml==1.1.5
non_process
i used the tool to generate the polars document but its content was not displayed properly problem description all the content is compressed into a small scroll bar area as shown below expect work requirement numpy pandas pyarrow hypothesis autodocsumm commonmark numpydoc pydata sphinx theme sphinx autosummary accessors sphinx copybutton sphinx design sphinx sphinxcontrib applehelp sphinxcontrib devhelp sphinxcontrib htmlhelp sphinxcontrib jsmath sphinxcontrib napoleon sphinxcontrib qthelp sphinxcontrib serializinghtml
0
16,341
21,001,258,463
IssuesEvent
2022-03-29 17:41:07
googleapis/nodejs-logging
https://api.github.com/repos/googleapis/nodejs-logging
closed
`@google-cloud/logging-min` NPM package README
type: process api: logging lang: nodejs
The NPM page for [@google-cloud/logging-min](https://www.npmjs.com/package/@google-cloud/logging-min) is the same as [@google-cloud/logging](https://www.npmjs.com/package/@google-cloud/logging). This can cause some confusion on what the difference are and how to install the correct package: `npm install @google-cloud/logging` vs `npm install @google-cloud/logging-min`.
1.0
`@google-cloud/logging-min` NPM package README - The NPM page for [@google-cloud/logging-min](https://www.npmjs.com/package/@google-cloud/logging-min) is the same as [@google-cloud/logging](https://www.npmjs.com/package/@google-cloud/logging). This can cause some confusion on what the difference are and how to install the correct package: `npm install @google-cloud/logging` vs `npm install @google-cloud/logging-min`.
process
google cloud logging min npm package readme the npm page for is the same as this can cause some confusion on what the difference are and how to install the correct package npm install google cloud logging vs npm install google cloud logging min
1
12,105
14,740,387,835
IssuesEvent
2021-01-07 09:00:33
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
File Upload for Outbound Site Billing
anc-process anp-1.5 ant-enhancement grt-ui processes
In GitLab by @kdjstudios on Nov 2, 2018, 14:38 **Submitted by:** <cori.bartlett@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-02-38333/conversation **Server:** Internal **Client/Site:** Multiple **Account:** NA **Issue:** What would it take to have the Outbound Sites setup to be able to upload file that would capture data to generate charges?
2.0
File Upload for Outbound Site Billing - In GitLab by @kdjstudios on Nov 2, 2018, 14:38 **Submitted by:** <cori.bartlett@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-02-38333/conversation **Server:** Internal **Client/Site:** Multiple **Account:** NA **Issue:** What would it take to have the Outbound Sites setup to be able to upload file that would capture data to generate charges?
process
file upload for outbound site billing in gitlab by kdjstudios on nov submitted by helpdesk server internal client site multiple account na issue what would it take to have the outbound sites setup to be able to upload file that would capture data to generate charges
1
137,213
18,752,659,456
IssuesEvent
2021-11-05 05:45:45
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2018-1065 (Medium) detected in multiple libraries
security vulnerability
## CVE-2018-1065 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stablev4.17.12</b>, <b>linux-stablev4.17.12</b>, <b>linux-stablev4.17.12</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The netfilter subsystem in the Linux kernel through 4.15.7 mishandles the case of a rule blob that contains a jump but lacks a user-defined chain, which allows local users to cause a denial of service (NULL pointer dereference) by leveraging the CAP_NET_RAW or CAP_NET_ADMIN capability, related to arpt_do_table in net/ipv4/netfilter/arp_tables.c, ipt_do_table in net/ipv4/netfilter/ip_tables.c, and ip6t_do_table in net/ipv6/netfilter/ip6_tables.c. <p>Publish Date: 2018-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1065>CVE-2018-1065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1065">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1065</a></p> <p>Release Date: 2018-03-02</p> <p>Fix Resolution: v4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1065 (Medium) detected in multiple libraries - ## CVE-2018-1065 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stablev4.17.12</b>, <b>linux-stablev4.17.12</b>, <b>linux-stablev4.17.12</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The netfilter subsystem in the Linux kernel through 4.15.7 mishandles the case of a rule blob that contains a jump but lacks a user-defined chain, which allows local users to cause a denial of service (NULL pointer dereference) by leveraging the CAP_NET_RAW or CAP_NET_ADMIN capability, related to arpt_do_table in net/ipv4/netfilter/arp_tables.c, ipt_do_table in net/ipv4/netfilter/ip_tables.c, and ip6t_do_table in net/ipv6/netfilter/ip6_tables.c. <p>Publish Date: 2018-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1065>CVE-2018-1065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1065">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1065</a></p> <p>Release Date: 2018-03-02</p> <p>Fix Resolution: v4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries linux linux linux vulnerability details the netfilter subsystem in the linux kernel through mishandles the case of a rule blob that contains a jump but lacks a user defined chain which allows local users to cause a denial of service null pointer dereference by leveraging the cap net raw or cap net admin capability related to arpt do table in net netfilter arp tables c ipt do table in net netfilter ip tables c and do table in net netfilter tables c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
5,164
7,936,782,273
IssuesEvent
2018-07-09 10:33:17
cypress-io/cypress-example-kitchensink
https://api.github.com/repos/cypress-io/cypress-example-kitchensink
closed
Fix Codeship Basic NPM install
process: ci
Failing on run, because Cypress binary is not installed https://app.codeship.com/projects/134609/builds/35549801?pipeline=07b7a962-0632-408f-aaea-627753de4fc3 ``` > cypress-example-kitchensink@1.0.2 e2e:record /home/rof/src/github.com/cypress-io/cypress-example-kitchensink > cypress run --record cypress:cli cli starts with arguments ["/home/rof/.nvm/versions/node/v6.5.0/bin/node","/home/rof/src/github.com/cypress-io/cypress-example-kitchensink/node_modules/.bin/cypress","run","--record"] +0ms cypress:cli NODE_OPTIONS is not set +0ms cypress:cli program parsing arguments +3ms cypress:cli running Cypress +2ms cypress:cli parsed cli options { record: true } +85ms cypress:cli verifying Cypress app +0ms cypress:cli checking environment variables +1ms cypress:cli checking if executable exists /home/rof/.cache/Cypress/3.0.2/Cypress/Cypress +4ms The cypress npm package is installed, but the Cypress binary is missing. We expected the binary to be installed here: /home/rof/.cache/Cypress/3.0.2/Cypress/Cypress Reasons it may be missing: - You're caching 'node_modules' but are not caching this path: /home/rof/.cache/Cypress - You ran 'npm install' at an earlier build step but did not persist: /home/rof/.cache/Cypress Properly caching the binary will fix this error and avoid downloading and unzipping Cypress. Alternatively, you can run 'cypress install' to download the binary again. https://on.cypress.io/not-installed-ci-error ---------- Platform: linux (Ubuntu Linux - 14.04) Cypress Version: 3.0.2 ```
1.0
Fix Codeship Basic NPM install - Failing on run, because Cypress binary is not installed https://app.codeship.com/projects/134609/builds/35549801?pipeline=07b7a962-0632-408f-aaea-627753de4fc3 ``` > cypress-example-kitchensink@1.0.2 e2e:record /home/rof/src/github.com/cypress-io/cypress-example-kitchensink > cypress run --record cypress:cli cli starts with arguments ["/home/rof/.nvm/versions/node/v6.5.0/bin/node","/home/rof/src/github.com/cypress-io/cypress-example-kitchensink/node_modules/.bin/cypress","run","--record"] +0ms cypress:cli NODE_OPTIONS is not set +0ms cypress:cli program parsing arguments +3ms cypress:cli running Cypress +2ms cypress:cli parsed cli options { record: true } +85ms cypress:cli verifying Cypress app +0ms cypress:cli checking environment variables +1ms cypress:cli checking if executable exists /home/rof/.cache/Cypress/3.0.2/Cypress/Cypress +4ms The cypress npm package is installed, but the Cypress binary is missing. We expected the binary to be installed here: /home/rof/.cache/Cypress/3.0.2/Cypress/Cypress Reasons it may be missing: - You're caching 'node_modules' but are not caching this path: /home/rof/.cache/Cypress - You ran 'npm install' at an earlier build step but did not persist: /home/rof/.cache/Cypress Properly caching the binary will fix this error and avoid downloading and unzipping Cypress. Alternatively, you can run 'cypress install' to download the binary again. https://on.cypress.io/not-installed-ci-error ---------- Platform: linux (Ubuntu Linux - 14.04) Cypress Version: 3.0.2 ```
process
fix codeship basic npm install failing on run because cypress binary is not installed cypress example kitchensink record home rof src github com cypress io cypress example kitchensink cypress run record cypress cli cli starts with arguments cypress cli node options is not set cypress cli program parsing arguments cypress cli running cypress cypress cli parsed cli options record true cypress cli verifying cypress app cypress cli checking environment variables cypress cli checking if executable exists home rof cache cypress cypress cypress the cypress npm package is installed but the cypress binary is missing we expected the binary to be installed here home rof cache cypress cypress cypress reasons it may be missing you re caching node modules but are not caching this path home rof cache cypress you ran npm install at an earlier build step but did not persist home rof cache cypress properly caching the binary will fix this error and avoid downloading and unzipping cypress alternatively you can run cypress install to download the binary again platform linux ubuntu linux cypress version
1
12,295
14,852,796,022
IssuesEvent
2021-01-18 09:05:17
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
R session buffer hangs while waiting for help to show up
process:eval
I thought this was the same issue as https://github.com/emacs-ess/ESS/issues/1062, but perhaps not. When I start a new R session, and I type e.g. `?data.frame`, I get an `Retrieving RDS aliases...done` message in the minibuffer, but other than that it keeps hanging! When I C-g a couple of times, the help buffer shows up but the `? data.frame` line is still running, and keeps running even after I <kbd>q</kbd> on the help buffer. I then have to C-c on the R terminal and I'm back to normal. Perhaps the last bit is irrelevant because q doesn't appear to kill the buffer, so it may make sense to have the help command in the R session still run. If I manually kill the help buffer, it also doesn't unfreeze the interactive session. This time, when I tried to run the previous command again (M-p) History item 1, it shows up but now the whole emacs session seems t hang until I repeatedly hit C-g again. Then tried again and it worked normally this time. Running another command, e.g. `? tibble::tibble` repeats the process.
1.0
R session buffer hangs while waiting for help to show up - I thought this was the same issue as https://github.com/emacs-ess/ESS/issues/1062, but perhaps not. When I start a new R session, and I type e.g. `?data.frame`, I get an `Retrieving RDS aliases...done` message in the minibuffer, but other than that it keeps hanging! When I C-g a couple of times, the help buffer shows up but the `? data.frame` line is still running, and keeps running even after I <kbd>q</kbd> on the help buffer. I then have to C-c on the R terminal and I'm back to normal. Perhaps the last bit is irrelevant because q doesn't appear to kill the buffer, so it may make sense to have the help command in the R session still run. If I manually kill the help buffer, it also doesn't unfreeze the interactive session. This time, when I tried to run the previous command again (M-p) History item 1, it shows up but now the whole emacs session seems t hang until I repeatedly hit C-g again. Then tried again and it worked normally this time. Running another command, e.g. `? tibble::tibble` repeats the process.
process
r session buffer hangs while waiting for help to show up i thought this was the same issue as but perhaps not when i start a new r session and i type e g data frame i get an retrieving rds aliases done message in the minibuffer but other than that it keeps hanging when i c g a couple of times the help buffer shows up but the data frame line is still running and keeps running even after i q on the help buffer i then have to c c on the r terminal and i m back to normal perhaps the last bit is irrelevant because q doesn t appear to kill the buffer so it may make sense to have the help command in the r session still run if i manually kill the help buffer it also doesn t unfreeze the interactive session this time when i tried to run the previous command again m p history item it shows up but now the whole emacs session seems t hang until i repeatedly hit c g again then tried again and it worked normally this time running another command e g tibble tibble repeats the process
1
84,799
7,934,320,283
IssuesEvent
2018-07-08 17:48:17
SunwellTracker/issues
https://api.github.com/repos/SunwellTracker/issues
closed
Quest: Buzzbox 411 isn't available.
In progress Works locally | Requires testing enhancement
Decription: The former quest which leads up to this one is "Buzzbox 827" which was available to accept and complete, when turned in the mechanical box should give you the quest: Buzzbox 411 and yet it doesn't. How it works: It doesn't offer the quest Buzzbox 411 after completing Buzzbox 827. That's it. How it should work: The quest should be offered to the recipient per usual and yet in my case it didn't. Not much else I can provide other than that it's a quest given by a mechanical box sitting in the grass next to a wooden platform in Auberdine. Source (you should point out proofs of your report, please give us some source): Imgur [Mirror](https://imgur.com/a/HLZB27Y)
1.0
Quest: Buzzbox 411 isn't available. - Decription: The former quest which leads up to this one is "Buzzbox 827" which was available to accept and complete, when turned in the mechanical box should give you the quest: Buzzbox 411 and yet it doesn't. How it works: It doesn't offer the quest Buzzbox 411 after completing Buzzbox 827. That's it. How it should work: The quest should be offered to the recipient per usual and yet in my case it didn't. Not much else I can provide other than that it's a quest given by a mechanical box sitting in the grass next to a wooden platform in Auberdine. Source (you should point out proofs of your report, please give us some source): Imgur [Mirror](https://imgur.com/a/HLZB27Y)
non_process
quest buzzbox isn t available decription the former quest which leads up to this one is buzzbox which was available to accept and complete when turned in the mechanical box should give you the quest buzzbox and yet it doesn t how it works it doesn t offer the quest buzzbox after completing buzzbox that s it how it should work the quest should be offered to the recipient per usual and yet in my case it didn t not much else i can provide other than that it s a quest given by a mechanical box sitting in the grass next to a wooden platform in auberdine source you should point out proofs of your report please give us some source imgur
0
24,103
10,981,333,634
IssuesEvent
2019-11-30 20:58:41
Agoric/evaluator-shim
https://api.github.com/repos/Agoric/evaluator-shim
opened
Clarify the behavior of defineProperty
TC39 question security shim
Referencing https://github.com/tc39/proposal-realms/issues/141 by @erights: Because of the weird semantics of the browser's Window vs WindowProxy (not a proxy btw), the spec of `Object.defineProperty` was changed to allow it to indicate failure by returning false rather than throwing. The `defineProperty` exported by commons.js should wrap this to turn this false result back into a thrown error. See also https://github.com/tc39/ecma262/issues/672 Needs a safety label
True
Clarify the behavior of defineProperty - Referencing https://github.com/tc39/proposal-realms/issues/141 by @erights: Because of the weird semantics of the browser's Window vs WindowProxy (not a proxy btw), the spec of `Object.defineProperty` was changed to allow it to indicate failure by returning false rather than throwing. The `defineProperty` exported by commons.js should wrap this to turn this false result back into a thrown error. See also https://github.com/tc39/ecma262/issues/672 Needs a safety label
non_process
clarify the behavior of defineproperty referencing by erights because of the weird semantics of the browser s window vs windowproxy not a proxy btw the spec of object defineproperty was changed to allow it to indicate failure by returning false rather than throwing the defineproperty exported by commons js should wrap this to turn this false result back into a thrown error see also needs a safety label
0
19,670
26,029,992,785
IssuesEvent
2022-12-21 20:07:59
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
Unwanted gapic regeneration PR opened by yoshi-automation
type: process
The PR chore(all): auto-regenerate gapics #6320 appears to be the result of an error in internal/gapicgen, possibly a flaw in interpreting the git status. Recurrent generation of this PR should be prevented.
1.0
Unwanted gapic regeneration PR opened by yoshi-automation - The PR chore(all): auto-regenerate gapics #6320 appears to be the result of an error in internal/gapicgen, possibly a flaw in interpreting the git status. Recurrent generation of this PR should be prevented.
process
unwanted gapic regeneration pr opened by yoshi automation the pr chore all auto regenerate gapics appears to be the result of an error in internal gapicgen possibly a flaw in interpreting the git status recurrent generation of this pr should be prevented
1
134,013
29,802,379,656
IssuesEvent
2023-06-16 09:04:56
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: REST API: application crashes when typing in dynamic binding in REST API query URL input box
Bug Frontend High Community Reported Needs Triaging BE Coders Pod REST API plugin Integrations Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description - When user types in a URL with dynamic binding in REST API query URL input box then application crashes. - This issue was reported as part of #24218 - [Ref](https://theappsmith.slack.com/archives/C040LHZN03V/p1686577919515999?thread_ts=1686300736.679729&cid=C040LHZN03V) https://github.com/appsmithorg/appsmith/assets/1757421/214f87e6-7d8b-45dd-a8e4-a0309cc02326 ### Steps To Reproduce 1. Create a JS Object and store a url in appsmith store say key is api_url. 2. Create new API query. 3. Edit the API query URL input box and write slowly : {{appsmith.store.api_url}} ### Public Sample App _No response_ ### Environment Production ### Issue video log _No response_ ### Version cloud, self hosted
1.0
[Bug]: REST API: application crashes when typing in dynamic binding in REST API query URL input box - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description - When user types in a URL with dynamic binding in REST API query URL input box then application crashes. - This issue was reported as part of #24218 - [Ref](https://theappsmith.slack.com/archives/C040LHZN03V/p1686577919515999?thread_ts=1686300736.679729&cid=C040LHZN03V) https://github.com/appsmithorg/appsmith/assets/1757421/214f87e6-7d8b-45dd-a8e4-a0309cc02326 ### Steps To Reproduce 1. Create a JS Object and store a url in appsmith store say key is api_url. 2. Create new API query. 3. Edit the API query URL input box and write slowly : {{appsmith.store.api_url}} ### Public Sample App _No response_ ### Environment Production ### Issue video log _No response_ ### Version cloud, self hosted
non_process
rest api application crashes when typing in dynamic binding in rest api query url input box is there an existing issue for this i have searched the existing issues description when user types in a url with dynamic binding in rest api query url input box then application crashes this issue was reported as part of steps to reproduce create a js object and store a url in appsmith store say key is api url create new api query edit the api query url input box and write slowly appsmith store api url public sample app no response environment production issue video log no response version cloud self hosted
0
692,439
23,734,871,087
IssuesEvent
2022-08-31 07:09:38
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.youtube.com - see bug description
browser-firefox priority-critical engine-gecko
<!-- @browser: Firefox 104.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:104.0) Gecko/20100101 Firefox/104.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/110014 --> **URL**: http://www.youtube.com **Browser / Version**: Firefox 104.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Opera **Problem type**: Something else **Description**: Every time I go to youtube in firefox now, the page just keep refreshing to the point I cant do anything. A recent update has not fixed this issue. I dont understand why when going on youtube the site just refreshes and refreshes automatically. Please fix this! **Steps to Reproduce**: Already descirbed the issue. Every time I go to youtube now the home page just constantly refreshes for now reason. I cant even watch videos anymore. Ive tries using youtube on another browser and it is fine. I would like to use firefox though. This is a recent issue and a update has not helped fix it. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.youtube.com - see bug description - <!-- @browser: Firefox 104.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:104.0) Gecko/20100101 Firefox/104.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/110014 --> **URL**: http://www.youtube.com **Browser / Version**: Firefox 104.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Opera **Problem type**: Something else **Description**: Every time I go to youtube in firefox now, the page just keep refreshing to the point I cant do anything. A recent update has not fixed this issue. I dont understand why when going on youtube the site just refreshes and refreshes automatically. Please fix this! **Steps to Reproduce**: Already descirbed the issue. Every time I go to youtube now the home page just constantly refreshes for now reason. I cant even watch videos anymore. Ive tries using youtube on another browser and it is fine. I would like to use firefox though. This is a recent issue and a update has not helped fix it. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox operating system windows tested another browser yes opera problem type something else description every time i go to youtube in firefox now the page just keep refreshing to the point i cant do anything a recent update has not fixed this issue i dont understand why when going on youtube the site just refreshes and refreshes automatically please fix this steps to reproduce already descirbed the issue every time i go to youtube now the home page just constantly refreshes for now reason i cant even watch videos anymore ive tries using youtube on another browser and it is fine i would like to use firefox though this is a recent issue and a update has not helped fix it browser configuration none from with ❤️
0
321,407
9,798,157,220
IssuesEvent
2019-06-11 11:43:37
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
mega.nz - design is broken
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://mega.nz/fm **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Design is broken **Description**: it wont let me download files from mega or goggle **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
mega.nz - design is broken - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://mega.nz/fm **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Design is broken **Description**: it wont let me download files from mega or goggle **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
mega nz design is broken url browser version firefox mobile operating system android tested another browser no problem type design is broken description it wont let me download files from mega or goggle steps to reproduce browser configuration none from with ❤️
0
19,049
25,050,379,185
IssuesEvent
2022-11-05 20:14:29
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
opened
Migrate to Rescue Prime Optimized hash function
processor air v0.4
Currently, the native hash function of the VM is a non-standard variation of Rescue Prime. Besides being non-standard, it has several disadvantages which we can remedy by switching to Rescue Prime Optimized (RPO), which was designed by the original authors of Rescue/Rescue Prime. We already have a Rust implementation of RPO [here](https://github.com/0xPolygonMiden/crypto/blob/main/src/hash/rpo/mod.rs). The main difference between Rescue Prime (in the variant that we are using) and Rescue Prime Optimized are as follows: * **Overwrite mode**: during the absorption step, instead of adding new elements to the elements in rate portion of the state, we can just overwrite them. * **Better padding rule**: the padding rule used in the variant of Rescue Prime is one of the things that make it non-standard. This padding rule has questionable security. The padding rule described in RPO specifications is much better. * **No inverse MDS matrix**: because of a slightly different arrangement of operations within a round, multiplication by an inverse MDS matrix is no longer needed for evaluating AIR constraints. To replace Rescue Prime with RPO, we'll need to: * Update how the processor works. Specifically, the hasher chiplet but probably a few other things too (e.g., chiplet bus, decoder). * Update AIR constraints for the affected components. One nice thing is that these updates should actually reduce the overall complexity.
1.0
Migrate to Rescue Prime Optimized hash function - Currently, the native hash function of the VM is a non-standard variation of Rescue Prime. Besides being non-standard, it has several disadvantages which we can remedy by switching to Rescue Prime Optimized (RPO), which was designed by the original authors of Rescue/Rescue Prime. We already have a Rust implementation of RPO [here](https://github.com/0xPolygonMiden/crypto/blob/main/src/hash/rpo/mod.rs). The main difference between Rescue Prime (in the variant that we are using) and Rescue Prime Optimized are as follows: * **Overwrite mode**: during the absorption step, instead of adding new elements to the elements in rate portion of the state, we can just overwrite them. * **Better padding rule**: the padding rule used in the variant of Rescue Prime is one of the things that make it non-standard. This padding rule has questionable security. The padding rule described in RPO specifications is much better. * **No inverse MDS matrix**: because of a slightly different arrangement of operations within a round, multiplication by an inverse MDS matrix is no longer needed for evaluating AIR constraints. To replace Rescue Prime with RPO, we'll need to: * Update how the processor works. Specifically, the hasher chiplet but probably a few other things too (e.g., chiplet bus, decoder). * Update AIR constraints for the affected components. One nice thing is that these updates should actually reduce the overall complexity.
process
migrate to rescue prime optimized hash function currently the native hash function of the vm is a non standard variation of rescue prime besides being non standard it has several disadvantages which we can remedy by switching to rescue prime optimized rpo which was designed by the original authors of rescue rescue prime we already have a rust implementation of rpo the main difference between rescue prime in the variant that we are using and rescue prime optimized are as follows overwrite mode during the absorption step instead of adding new elements to the elements in rate portion of the state we can just overwrite them better padding rule the padding rule used in the variant of rescue prime is one of the things that make it non standard this padding rule has questionable security the padding rule described in rpo specifications is much better no inverse mds matrix because of a slightly different arrangement of operations within a round multiplication by an inverse mds matrix is no longer needed for evaluating air constraints to replace rescue prime with rpo we ll need to update how the processor works specifically the hasher chiplet but probably a few other things too e g chiplet bus decoder update air constraints for the affected components one nice thing is that these updates should actually reduce the overall complexity
1
21,960
30,454,236,458
IssuesEvent
2023-07-16 17:28:06
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add The Goof Who Sat By the Door: The Thomas Washington Story from "Atlanta" (Screenshots and Title Screen added)
suggested title in process
Please add as much of the following info as you can: Title: The Goof Who Sat By the Door: The Thomas Washington Story Type (film/tv show): film - documentary Film or show in which it appears: Atlanta Synopsis: A detailed account of the rise and fall of Thomas Washington, an unassuming Black animator from Atlanta, Georgia, who, in 1992, was accidentially appointed as CEO of The Walt Disney Company, and his quest to use the opportunity to create "the Blackest movie of all time:" _A Goofy Movie_.* Is the parent film/show streaming anywhere? yes - Hulu About when in the parent film/show does it appear? Ep. 4x08 - "The Goof Who Sat By the Door" It consists of the entire episode. Actual footage of the film/show can be seen (yes/no)? yes Director: Devonte Jackson Production studio: B.A.N. (Black American Network) Documentary focus: Thomas Washington Featuring: Evelyn Jackson (Thomas' mother), Phillip Washington (Thomas' cousin), Mark Dillard (S. C. A. D. professor), Jenna Wortham (journalist), Leonard Trussell (former Disney executive), Joel Zempel (former Disney Company board member), Gary Welch (former Disney animator), Frank Rolls (director), Anna Speilman (Thomas' wife), Maxwell Washington (Thomas' son), Chris Spinx (former Disney animator), Brian McKnight (singer), & Sinbad (comedian) Fun Fact: The title of the documentary is based on _The Spook Who Sat By the Door_, the novel (and later film adaptation) by Sam Greenlee about the first Black CIA officer. *_At the request of B.A.N.'s lawyers, we asknowledge that though the story you're about to see has been fact-checked to the best of our ability, certain details persented may not be endorsed by The Walt Disney Company._ ![Goof_Title Screen](https://user-images.githubusercontent.com/88982629/201405218-5e384c4d-f27f-4b01-b0da-0fbd93ddfd64.jpg) ![Goof_1](https://user-images.githubusercontent.com/88982629/201381484-0a1c19ea-dc79-4e50-ae1a-b1a5f4ab7495.jpg) ![Goof_2](https://user-images.githubusercontent.com/88982629/201381486-5438410f-30c3-4391-8206-24c8696468f7.jpg) ![Goof_3](https://user-images.githubusercontent.com/88982629/201381487-c44a757f-28fe-4cf7-91a6-c274da1b8262.jpg) ![Goof_4](https://user-images.githubusercontent.com/88982629/201381490-37c2ddf8-7323-439d-812e-9262cf561f88.jpg) ![Goof_5](https://user-images.githubusercontent.com/88982629/201381493-4ae18cb4-2393-4634-a094-d2340cac8128.jpg) ![Goof_6](https://user-images.githubusercontent.com/88982629/201381495-f2faedbf-b39e-4610-a24e-33ac74f7b339.jpg) ![Goof_7](https://user-images.githubusercontent.com/88982629/201381498-5988e694-1f21-41fe-a68a-4d3f8470c4b8.jpg) ![Goof_9](https://user-images.githubusercontent.com/88982629/201381506-e7d2fb30-0757-4f84-99da-62c96fc032c8.jpg) ![Goof_10](https://user-images.githubusercontent.com/88982629/201381507-f6bfdea0-bc46-42b9-8418-e25c7521e341.jpg) ![Goof_12](https://user-images.githubusercontent.com/88982629/201381849-63dc6d31-5be1-4af4-9e72-73cc8938eea1.jpg)
1.0
Add The Goof Who Sat By the Door: The Thomas Washington Story from "Atlanta" (Screenshots and Title Screen added) - Please add as much of the following info as you can: Title: The Goof Who Sat By the Door: The Thomas Washington Story Type (film/tv show): film - documentary Film or show in which it appears: Atlanta Synopsis: A detailed account of the rise and fall of Thomas Washington, an unassuming Black animator from Atlanta, Georgia, who, in 1992, was accidentially appointed as CEO of The Walt Disney Company, and his quest to use the opportunity to create "the Blackest movie of all time:" _A Goofy Movie_.* Is the parent film/show streaming anywhere? yes - Hulu About when in the parent film/show does it appear? Ep. 4x08 - "The Goof Who Sat By the Door" It consists of the entire episode. Actual footage of the film/show can be seen (yes/no)? yes Director: Devonte Jackson Production studio: B.A.N. (Black American Network) Documentary focus: Thomas Washington Featuring: Evelyn Jackson (Thomas' mother), Phillip Washington (Thomas' cousin), Mark Dillard (S. C. A. D. professor), Jenna Wortham (journalist), Leonard Trussell (former Disney executive), Joel Zempel (former Disney Company board member), Gary Welch (former Disney animator), Frank Rolls (director), Anna Speilman (Thomas' wife), Maxwell Washington (Thomas' son), Chris Spinx (former Disney animator), Brian McKnight (singer), & Sinbad (comedian) Fun Fact: The title of the documentary is based on _The Spook Who Sat By the Door_, the novel (and later film adaptation) by Sam Greenlee about the first Black CIA officer. *_At the request of B.A.N.'s lawyers, we asknowledge that though the story you're about to see has been fact-checked to the best of our ability, certain details persented may not be endorsed by The Walt Disney Company._ ![Goof_Title Screen](https://user-images.githubusercontent.com/88982629/201405218-5e384c4d-f27f-4b01-b0da-0fbd93ddfd64.jpg) ![Goof_1](https://user-images.githubusercontent.com/88982629/201381484-0a1c19ea-dc79-4e50-ae1a-b1a5f4ab7495.jpg) ![Goof_2](https://user-images.githubusercontent.com/88982629/201381486-5438410f-30c3-4391-8206-24c8696468f7.jpg) ![Goof_3](https://user-images.githubusercontent.com/88982629/201381487-c44a757f-28fe-4cf7-91a6-c274da1b8262.jpg) ![Goof_4](https://user-images.githubusercontent.com/88982629/201381490-37c2ddf8-7323-439d-812e-9262cf561f88.jpg) ![Goof_5](https://user-images.githubusercontent.com/88982629/201381493-4ae18cb4-2393-4634-a094-d2340cac8128.jpg) ![Goof_6](https://user-images.githubusercontent.com/88982629/201381495-f2faedbf-b39e-4610-a24e-33ac74f7b339.jpg) ![Goof_7](https://user-images.githubusercontent.com/88982629/201381498-5988e694-1f21-41fe-a68a-4d3f8470c4b8.jpg) ![Goof_9](https://user-images.githubusercontent.com/88982629/201381506-e7d2fb30-0757-4f84-99da-62c96fc032c8.jpg) ![Goof_10](https://user-images.githubusercontent.com/88982629/201381507-f6bfdea0-bc46-42b9-8418-e25c7521e341.jpg) ![Goof_12](https://user-images.githubusercontent.com/88982629/201381849-63dc6d31-5be1-4af4-9e72-73cc8938eea1.jpg)
process
add the goof who sat by the door the thomas washington story from atlanta screenshots and title screen added please add as much of the following info as you can title the goof who sat by the door the thomas washington story type film tv show film documentary film or show in which it appears atlanta synopsis a detailed account of the rise and fall of thomas washington an unassuming black animator from atlanta georgia who in was accidentially appointed as ceo of the walt disney company and his quest to use the opportunity to create the blackest movie of all time a goofy movie is the parent film show streaming anywhere yes hulu about when in the parent film show does it appear ep the goof who sat by the door it consists of the entire episode actual footage of the film show can be seen yes no yes director devonte jackson production studio b a n black american network documentary focus thomas washington featuring evelyn jackson thomas mother phillip washington thomas cousin mark dillard s c a d professor jenna wortham journalist leonard trussell former disney executive joel zempel former disney company board member gary welch former disney animator frank rolls director anna speilman thomas wife maxwell washington thomas son chris spinx former disney animator brian mcknight singer sinbad comedian fun fact the title of the documentary is based on the spook who sat by the door the novel and later film adaptation by sam greenlee about the first black cia officer at the request of b a n s lawyers we asknowledge that though the story you re about to see has been fact checked to the best of our ability certain details persented may not be endorsed by the walt disney company
1
503,074
14,579,064,545
IssuesEvent
2020-12-18 06:29:41
kubeflow/website
https://api.github.com/repos/kubeflow/website
closed
How to enable caching is missing
area/docs area/pipelines kind/feature lifecycle/stale priority/p2
In the section : https://www.kubeflow.org/docs/pipelines/caching/#enabling-caching. It is assumed that we check caching is enable but no way technique is given what to do if caching is not enabled!! Bummer right
1.0
How to enable caching is missing - In the section : https://www.kubeflow.org/docs/pipelines/caching/#enabling-caching. It is assumed that we check caching is enable but no way technique is given what to do if caching is not enabled!! Bummer right
non_process
how to enable caching is missing in the section it is assumed that we check caching is enable but no way technique is given what to do if caching is not enabled bummer right
0
16,433
21,316,756,549
IssuesEvent
2022-04-16 12:16:13
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Chunk to-content does not preserve (HTML) or correctly reflect (PDF) topicheads within the chunk
bug preprocess/chunking enhancement stale
## Expected Behavior Topicheads are reflected in the resulting chunk as though they were title-only topics. This both the only reasonable result (omitting them would be data loss, which is unacceptable) and is consistent with the "to-content generates title-only topics" rule for topicheads. ## Actual Behavior HTML: The topicheads are reflected in the ToC but not in the resulting chunk. PDF: The topicheads are reflected in the PDF after all the non-topichead descendants of the chunk root. ## Possible Solution ## Steps to Reproduce <!-- Test case, Gist, set of files or steps required to reproduce the issue. --> 1. Process the root map https://github.com/dita-community/dita-test-cases/blob/master/topichead-chunking/topichead-chunking-test-03.ditamap to both XHTML and PDF 2. In the HTML observe that the topicheads are not reflected in the resulting HTML file for the chunk. 3. In the PDF observe that the topicheads occur after their subordinate topics in the flow and ToC. ## Copy of the error message, log file or stack trace No unexpected messages from the chunk stage. ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: 2.5.3 * Operating system and version: macOS * How did you run DITA-OT? oXygen * Transformation type: xhtml, PDF2 <!-- Before submitting, check the Preview tab above to verify the XML markup appears correctly and remember you can edit the description later to add information. -->
1.0
Chunk to-content does not preserve (HTML) or correctly reflect (PDF) topicheads within the chunk - ## Expected Behavior Topicheads are reflected in the resulting chunk as though they were title-only topics. This both the only reasonable result (omitting them would be data loss, which is unacceptable) and is consistent with the "to-content generates title-only topics" rule for topicheads. ## Actual Behavior HTML: The topicheads are reflected in the ToC but not in the resulting chunk. PDF: The topicheads are reflected in the PDF after all the non-topichead descendants of the chunk root. ## Possible Solution ## Steps to Reproduce <!-- Test case, Gist, set of files or steps required to reproduce the issue. --> 1. Process the root map https://github.com/dita-community/dita-test-cases/blob/master/topichead-chunking/topichead-chunking-test-03.ditamap to both XHTML and PDF 2. In the HTML observe that the topicheads are not reflected in the resulting HTML file for the chunk. 3. In the PDF observe that the topicheads occur after their subordinate topics in the flow and ToC. ## Copy of the error message, log file or stack trace No unexpected messages from the chunk stage. ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: 2.5.3 * Operating system and version: macOS * How did you run DITA-OT? oXygen * Transformation type: xhtml, PDF2 <!-- Before submitting, check the Preview tab above to verify the XML markup appears correctly and remember you can edit the description later to add information. -->
process
chunk to content does not preserve html or correctly reflect pdf topicheads within the chunk expected behavior topicheads are reflected in the resulting chunk as though they were title only topics this both the only reasonable result omitting them would be data loss which is unacceptable and is consistent with the to content generates title only topics rule for topicheads actual behavior html the topicheads are reflected in the toc but not in the resulting chunk pdf the topicheads are reflected in the pdf after all the non topichead descendants of the chunk root possible solution steps to reproduce process the root map to both xhtml and pdf in the html observe that the topicheads are not reflected in the resulting html file for the chunk in the pdf observe that the topicheads occur after their subordinate topics in the flow and toc copy of the error message log file or stack trace no unexpected messages from the chunk stage environment dita ot version operating system and version macos how did you run dita ot oxygen transformation type xhtml before submitting check the preview tab above to verify the xml markup appears correctly and remember you can edit the description later to add information
1
66,291
16,584,043,266
IssuesEvent
2021-05-31 15:44:35
VROOM-Project/vroom
https://api.github.com/repos/VROOM-Project/vroom
closed
Add libvroom example to CI builds
CI builds enhancement
Every now and then, we do some refactoring of the C++ API or internal data structures that break the build for `libvroom_examples/libvroom.cpp`. This usually go unnoticed until someone reports it like in #510 so it would be nice to add a simple build check in the CI setup. Note: this is not about treating failed builds as an error as we don't provide any guarantee of stability of the C++ API, but rather a way to spot changes so that the example stays in sync with core updates.
1.0
Add libvroom example to CI builds - Every now and then, we do some refactoring of the C++ API or internal data structures that break the build for `libvroom_examples/libvroom.cpp`. This usually go unnoticed until someone reports it like in #510 so it would be nice to add a simple build check in the CI setup. Note: this is not about treating failed builds as an error as we don't provide any guarantee of stability of the C++ API, but rather a way to spot changes so that the example stays in sync with core updates.
non_process
add libvroom example to ci builds every now and then we do some refactoring of the c api or internal data structures that break the build for libvroom examples libvroom cpp this usually go unnoticed until someone reports it like in so it would be nice to add a simple build check in the ci setup note this is not about treating failed builds as an error as we don t provide any guarantee of stability of the c api but rather a way to spot changes so that the example stays in sync with core updates
0