Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
257,968
8,149,305,678
IssuesEvent
2018-08-22 09:12:58
Xceptance/neodymium-library
https://api.github.com/repos/Xceptance/neodymium-library
closed
Attachments in Allure report missing
Medium Priority bug doneInDevelop
Selenide attachments created while testing are currently not attached to Allure report.
1.0
Attachments in Allure report missing - Selenide attachments created while testing are currently not attached to Allure report.
priority
attachments in allure report missing selenide attachments created while testing are currently not attached to allure report
1
660,060
21,950,461,107
IssuesEvent
2022-05-24 07:26:12
dodona-edu/dodona
https://api.github.com/repos/dodona-edu/dodona
closed
Show transient messages to users
feature medium priority
We want a way to easily show messages to all users (or all course admins/staff/...). Such message would be displayed until it is dismissed or until we disable it. This could be used to announce new releases, scheduled downtime, etc. We want to be able to enable/disable this using an interface on Dodona and not doing a deploy.
1.0
Show transient messages to users - We want a way to easily show messages to all users (or all course admins/staff/...). Such message would be displayed until it is dismissed or until we disable it. This could be used to announce new releases, scheduled downtime, etc. We want to be able to enable/disable this using an interface on Dodona and not doing a deploy.
priority
show transient messages to users we want a way to easily show messages to all users or all course admins staff such message would be displayed until it is dismissed or until we disable it this could be used to announce new releases scheduled downtime etc we want to be able to enable disable this using an interface on dodona and not doing a deploy
1
103,346
4,167,530,056
IssuesEvent
2016-06-20 09:52:19
GreatEmerald/RGIC1601
https://api.github.com/repos/GreatEmerald/RGIC1601
closed
Implement a metadata chain
Priority 4: Medium
Starting from the ClassifyToZones module, operations that result in new rasters should contain metadata about the parameters used in creating the output. The way to do it, for raster objects, when `input` is an input raster, `output` is the output raster, and you want to add `newvar` set to `newvalue`, is: ```r metadata(output) = append(metadata(input), list(newvar="newvalue", newvar2="newvalue2")) ``` This metadata should be written to files in the end (put into the Spatial*DataFrame before sending to the ExportToFile module).
1.0
Implement a metadata chain - Starting from the ClassifyToZones module, operations that result in new rasters should contain metadata about the parameters used in creating the output. The way to do it, for raster objects, when `input` is an input raster, `output` is the output raster, and you want to add `newvar` set to `newvalue`, is: ```r metadata(output) = append(metadata(input), list(newvar="newvalue", newvar2="newvalue2")) ``` This metadata should be written to files in the end (put into the Spatial*DataFrame before sending to the ExportToFile module).
priority
implement a metadata chain starting from the classifytozones module operations that result in new rasters should contain metadata about the parameters used in creating the output the way to do it for raster objects when input is an input raster output is the output raster and you want to add newvar set to newvalue is r metadata output append metadata input list newvar newvalue this metadata should be written to files in the end put into the spatial dataframe before sending to the exporttofile module
1
260,749
8,214,348,435
IssuesEvent
2018-09-04 22:54:20
wevote/WebApp
https://api.github.com/repos/wevote/WebApp
closed
Force /src/index.html to reload every 12 hours
Difficulty: Medium Priority: 1
### Please describe the issue (What happens? What do you expect?) Add code that sets an expiration to the /src/index.html so that the page has to reload every 12 hours. With a new index.html page, the name of the bundle will change if there has been an update, which will force the download of the latest Javascript bundle.
1.0
Force /src/index.html to reload every 12 hours - ### Please describe the issue (What happens? What do you expect?) Add code that sets an expiration to the /src/index.html so that the page has to reload every 12 hours. With a new index.html page, the name of the bundle will change if there has been an update, which will force the download of the latest Javascript bundle.
priority
force src index html to reload every hours please describe the issue what happens what do you expect add code that sets an expiration to the src index html so that the page has to reload every hours with a new index html page the name of the bundle will change if there has been an update which will force the download of the latest javascript bundle
1
116,114
4,696,970,118
IssuesEvent
2016-10-12 07:35:54
PowerlineApp/powerline-mobile
https://api.github.com/repos/PowerlineApp/powerline-mobile
closed
Notification Action Button: Fundraiser/Crowdfunder "Donate" should load Item Detail Screen
bug P2 - Medium Priority
Currently shows newsfeed. Should show item detail screen for the fundraiser.
1.0
Notification Action Button: Fundraiser/Crowdfunder "Donate" should load Item Detail Screen - Currently shows newsfeed. Should show item detail screen for the fundraiser.
priority
notification action button fundraiser crowdfunder donate should load item detail screen currently shows newsfeed should show item detail screen for the fundraiser
1
20,342
2,622,843,561
IssuesEvent
2015-03-04 08:01:25
max99x/pagemon-chrome-ext
https://api.github.com/repos/max99x/pagemon-chrome-ext
closed
Optionally allow hiding of the removals ("red parts") in a diff
auto-migrated Priority-Medium Type-Enhancement
``` When viewing a diff, text that has been removed is marked in red and new text is marked in green. This is nice, but distorts the web page structure and makes it more difficult to read the diff (we are interested in "what is new?" more than "what has changed?"). As an option (and maybe by default) diffs should be shown without the text that has been removed (the parts currently in red). ``` Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Mar 2010 at 7:30
1.0
Optionally allow hiding of the removals ("red parts") in a diff - ``` When viewing a diff, text that has been removed is marked in red and new text is marked in green. This is nice, but distorts the web page structure and makes it more difficult to read the diff (we are interested in "what is new?" more than "what has changed?"). As an option (and maybe by default) diffs should be shown without the text that has been removed (the parts currently in red). ``` Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Mar 2010 at 7:30
priority
optionally allow hiding of the removals red parts in a diff when viewing a diff text that has been removed is marked in red and new text is marked in green this is nice but distorts the web page structure and makes it more difficult to read the diff we are interested in what is new more than what has changed as an option and maybe by default diffs should be shown without the text that has been removed the parts currently in red original issue reported on code google com by gkaemp gmail com on mar at
1
499,462
14,447,684,186
IssuesEvent
2020-12-08 04:26:13
JuezUN/INGInious
https://api.github.com/repos/JuezUN/INGInious
closed
Changes on subproblems tab
Change request Frontend Medium Priority Plugins Task
- [x] Allow languages in multilang depending on grading environment, that is, allowing VHDL and Verilog for VHDL environment, python for Data Science and the others for multilang. - [x] Restrict the creation of subproblems to maximum 1. - [x] Show multiple languages in alphabetic order.
1.0
Changes on subproblems tab - - [x] Allow languages in multilang depending on grading environment, that is, allowing VHDL and Verilog for VHDL environment, python for Data Science and the others for multilang. - [x] Restrict the creation of subproblems to maximum 1. - [x] Show multiple languages in alphabetic order.
priority
changes on subproblems tab allow languages in multilang depending on grading environment that is allowing vhdl and verilog for vhdl environment python for data science and the others for multilang restrict the creation of subproblems to maximum show multiple languages in alphabetic order
1
29,497
2,716,242,341
IssuesEvent
2015-04-10 17:53:12
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
The event BeforeRowEdit does not allow access to the widget responsible for row editing
bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.2.0
_From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 17:12:55_ The event BeforeRowEdit does not allow access to the widget responsible for row editing _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=419_
1.0
The event BeforeRowEdit does not allow access to the widget responsible for row editing - _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 17:12:55_ The event BeforeRowEdit does not allow access to the widget responsible for row editing _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=419_
priority
the event beforerowedit does not allow access to the widget responsible for row editing from on july the event beforerowedit does not allow access to the widget responsible for row editing original issue
1
70,203
3,321,165,151
IssuesEvent
2015-11-09 06:43:55
ekux44/LampShade
https://api.github.com/repos/ekux44/LampShade
closed
java.lang.NullPointerException in com.kuxhausen.huemore.MainFragment.onStateChanged
crash medium priority
LampShade version: 79, 82 Android version: 5.0, 5.1, 6.0 Devices: OnePlus One (A0001), Nexus 6 (shamu), Nexus 5 (hammerhead), Nexus 4 (mako), Galaxy S6 Edge (zeroltebmc) java.lang.NullPointerException: Attempt to invoke virtual method 'com.kuxhausen.huemore.net.DeviceManager com.kuxhausen.huemore.net.ConnectivityService.getDeviceManager()' on a null object reference at com.kuxhausen.huemore.MainFragment.onStateChanged(MainFragment.java:263) at com.kuxhausen.huemore.net.DeviceManager.onStateChanged(DeviceManager.java:186) at com.kuxhausen.huemore.net.hue.HubConnection.reportStateChangeSucess(HubConnection.java:288) at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:22) at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:7) at com.kuxhausen.huemore.net.hue.api.GsonRequest.deliverResponse(GsonRequest.java:83) at com.android.volley.ExecutorDelivery$ResponseDeliveryRunnable.run(ExecutorDelivery.java:99) at android.os.Handler.handleCallback(Handler.java:739) at android.os.Handler.dispatchMessage(Handler.java:95) at android.os.Looper.loop(Looper.java:148) at android.app.ActivityThread.main(ActivityThread.java:5417) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616)
1.0
java.lang.NullPointerException in com.kuxhausen.huemore.MainFragment.onStateChanged - LampShade version: 79, 82 Android version: 5.0, 5.1, 6.0 Devices: OnePlus One (A0001), Nexus 6 (shamu), Nexus 5 (hammerhead), Nexus 4 (mako), Galaxy S6 Edge (zeroltebmc) java.lang.NullPointerException: Attempt to invoke virtual method 'com.kuxhausen.huemore.net.DeviceManager com.kuxhausen.huemore.net.ConnectivityService.getDeviceManager()' on a null object reference at com.kuxhausen.huemore.MainFragment.onStateChanged(MainFragment.java:263) at com.kuxhausen.huemore.net.DeviceManager.onStateChanged(DeviceManager.java:186) at com.kuxhausen.huemore.net.hue.HubConnection.reportStateChangeSucess(HubConnection.java:288) at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:22) at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:7) at com.kuxhausen.huemore.net.hue.api.GsonRequest.deliverResponse(GsonRequest.java:83) at com.android.volley.ExecutorDelivery$ResponseDeliveryRunnable.run(ExecutorDelivery.java:99) at android.os.Handler.handleCallback(Handler.java:739) at android.os.Handler.dispatchMessage(Handler.java:95) at android.os.Looper.loop(Looper.java:148) at android.app.ActivityThread.main(ActivityThread.java:5417) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616)
priority
java lang nullpointerexception in com kuxhausen huemore mainfragment onstatechanged lampshade version android version devices oneplus one nexus shamu nexus hammerhead nexus mako galaxy edge zeroltebmc java lang nullpointerexception attempt to invoke virtual method com kuxhausen huemore net devicemanager com kuxhausen huemore net connectivityservice getdevicemanager on a null object reference at com kuxhausen huemore mainfragment onstatechanged mainfragment java at com kuxhausen huemore net devicemanager onstatechanged devicemanager java at com kuxhausen huemore net hue hubconnection reportstatechangesucess hubconnection java at com kuxhausen huemore net hue api statesuccesslistener onresponse statesuccesslistener java at com kuxhausen huemore net hue api statesuccesslistener onresponse statesuccesslistener java at com kuxhausen huemore net hue api gsonrequest deliverresponse gsonrequest java at com android volley executordelivery responsedeliveryrunnable run executordelivery java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java
1
268,788
8,414,496,734
IssuesEvent
2018-10-13 03:10:43
compodoc/compodoc
https://api.github.com/repos/compodoc/compodoc
closed
[BUG] Route docs are not generated
1. Type: Bug Context : routing Priority: Medium Status: Accepted Time: ~1 hour
<!-- > Please follow the issue template below for bug reports and queries. > For issue, start the label of the title with [BUG] > For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant. --> ##### **Overview of the issue** Routes information is not populated in generated docs. Used [angular-started](https://github.com/gdi2290/angular-starter) as example for testing. [Generated docs screenshot](http://prntscr.com/k64f0e) ##### **Operating System, Node.js, npm, compodoc version(s)** * **Node**: `v9.11.1` * **OS**: macOS 10.13, Windows 10 * **Compodoc**: `1.1.3` ##### **Angular configuration, a `package.json` file in the root folder** Use [angular-starter](https://github.com/gdi2290/angular-starter) sources. ##### **Compodoc installed globally or locally ?** Both ##### **If possible sourcecode of the file where it breaks** Use [angular-starter](https://github.com/gdi2290/angular-starter) sources. ##### **Reproduce the error** 1. Clone [angular-starter](https://github.com/gdi2290/angular-starter). 2. Run `npm install` or `yarn install` 3. Generate docs using compodoc: `npm run docs:compodoc` 4. View docs: `npm run docs:compodoc:generate` No route section generated ##### **Related issues** https://github.com/compodoc/compodoc/issues/257
1.0
[BUG] Route docs are not generated - <!-- > Please follow the issue template below for bug reports and queries. > For issue, start the label of the title with [BUG] > For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant. --> ##### **Overview of the issue** Routes information is not populated in generated docs. Used [angular-started](https://github.com/gdi2290/angular-starter) as example for testing. [Generated docs screenshot](http://prntscr.com/k64f0e) ##### **Operating System, Node.js, npm, compodoc version(s)** * **Node**: `v9.11.1` * **OS**: macOS 10.13, Windows 10 * **Compodoc**: `1.1.3` ##### **Angular configuration, a `package.json` file in the root folder** Use [angular-starter](https://github.com/gdi2290/angular-starter) sources. ##### **Compodoc installed globally or locally ?** Both ##### **If possible sourcecode of the file where it breaks** Use [angular-starter](https://github.com/gdi2290/angular-starter) sources. ##### **Reproduce the error** 1. Clone [angular-starter](https://github.com/gdi2290/angular-starter). 2. Run `npm install` or `yarn install` 3. Generate docs using compodoc: `npm run docs:compodoc` 4. View docs: `npm run docs:compodoc:generate` No route section generated ##### **Related issues** https://github.com/compodoc/compodoc/issues/257
priority
route docs are not generated please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue routes information is not populated in generated docs used as example for testing operating system node js npm compodoc version s node os macos windows compodoc angular configuration a package json file in the root folder use sources compodoc installed globally or locally both if possible sourcecode of the file where it breaks use sources reproduce the error clone run npm install or yarn install generate docs using compodoc npm run docs compodoc view docs npm run docs compodoc generate no route section generated related issues
1
296,737
9,125,653,180
IssuesEvent
2019-02-24 15:31:39
pixijs/pixi.js
https://api.github.com/repos/pixijs/pixi.js
closed
Prefiltered line technique to dramatically improve the rendering of lines an shapes
Difficulty: Medium Domain: API Plugin: Graphics Priority: Low Resolution: Won't Fix Status: Accepting PRs Type: Feature Request Version: v5.x
My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA) I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI. http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
1.0
Prefiltered line technique to dramatically improve the rendering of lines an shapes - My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA) I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI. http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
priority
prefiltered line technique to dramatically improve the rendering of lines an shapes my understanding is that currently pixi uses the stencil buffer to draw graphics objects like lines and shapes since the stencil buffer does not support alpha all lines are very jagged aliased this is fine with aa or fxaa on but both of these come with other sets of issues that may or may not be acceptable for a given project gpu browser support performance for aa and text rendering issues for fxaa i ve stumbled across a different technique which appears to complete solve this issue and if my understanding is right it should be faster than the current method i thought i would share the link in case it is something you guys might want to look into for a future version of pixi
1
231,650
7,641,700,393
IssuesEvent
2018-05-08 06:30:12
vmware/harbor
https://api.github.com/repos/vmware/harbor
closed
Sometimes label filter does not close
area/ui priority/medium
Harbor version: v1.5 rc3 Go into project repo, open search filter, select a label, then click refresh button on the right, click search filter to close search filter, the label filter does not close ![labeeel](https://user-images.githubusercontent.com/23101283/39354426-25cf3c9e-4a3d-11e8-9f17-fcef4e920c18.PNG)
1.0
Sometimes label filter does not close - Harbor version: v1.5 rc3 Go into project repo, open search filter, select a label, then click refresh button on the right, click search filter to close search filter, the label filter does not close ![labeeel](https://user-images.githubusercontent.com/23101283/39354426-25cf3c9e-4a3d-11e8-9f17-fcef4e920c18.PNG)
priority
sometimes label filter does not close harbor version go into project repo open search filter select a label then click refresh button on the right click search filter to close search filter the label filter does not close
1
704,162
24,187,698,110
IssuesEvent
2022-09-23 14:37:46
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
cross-compile toolchain variant doesn't working properly with multilib toolchain
bug priority: medium area: Toolchains
**Describe the bug** In case of `ZEPHYR_TOOLCHAIN_VARIANT=cross-compile` we don't set `SYSROOT_DIR` cmake variable, so we don't execute block of code which sets `LIBC_LIBRARY_DIR` variable based on the toolchain C flags (including mcpu flag): https://github.com/zephyrproject-rtos/zephyr/blob/88aa873fbc544b99c1e48e37d5c74662f90b3de0/cmake/compiler/gcc/target.cmake#L74-L88 So, in case of multilib cross-compile toolchain we always use default libraries set no mater which mcpu value we've provided. Obviously it's incorrect. **Impact** Implicit & unexpected usage of libraries which are build for different mcpu than was specified. If we are lucky we may get compile error - for example if we got something not linkable, like 64bit libraries from toolchain and 32 bit object files from rest of zephyr. However we may get successful linkage and use some not optimal libraries (i.e built for processor which use smaller instruction set) or get runtime issues. **possible solutions** * add additional environment variable for cross-compile toolchain variant, which will be pointing to correct SYSROOT_DIR path. * try to extract SYSROOT_DIR path from the `CROSS_COMPILE` variable. I've checked the value of `CROSS_COMPILE` I've previously used - it doesn't seems to be feasible for these toolchains I.e: ``` CROSS_COMPILE=/home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/bin/arc64-snps-elf- SYSROOT_DIR should be /home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/arc64-unknown-elf/ CROSS_COMPILE=/global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/bin/arc-elf32- SYSROOT_DIR should be /global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/arc-snps-elf/ ```
1.0
cross-compile toolchain variant doesn't working properly with multilib toolchain - **Describe the bug** In case of `ZEPHYR_TOOLCHAIN_VARIANT=cross-compile` we don't set `SYSROOT_DIR` cmake variable, so we don't execute block of code which sets `LIBC_LIBRARY_DIR` variable based on the toolchain C flags (including mcpu flag): https://github.com/zephyrproject-rtos/zephyr/blob/88aa873fbc544b99c1e48e37d5c74662f90b3de0/cmake/compiler/gcc/target.cmake#L74-L88 So, in case of multilib cross-compile toolchain we always use default libraries set no mater which mcpu value we've provided. Obviously it's incorrect. **Impact** Implicit & unexpected usage of libraries which are build for different mcpu than was specified. If we are lucky we may get compile error - for example if we got something not linkable, like 64bit libraries from toolchain and 32 bit object files from rest of zephyr. However we may get successful linkage and use some not optimal libraries (i.e built for processor which use smaller instruction set) or get runtime issues. **possible solutions** * add additional environment variable for cross-compile toolchain variant, which will be pointing to correct SYSROOT_DIR path. * try to extract SYSROOT_DIR path from the `CROSS_COMPILE` variable. I've checked the value of `CROSS_COMPILE` I've previously used - it doesn't seems to be feasible for these toolchains I.e: ``` CROSS_COMPILE=/home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/bin/arc64-snps-elf- SYSROOT_DIR should be /home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/arc64-unknown-elf/ CROSS_COMPILE=/global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/bin/arc-elf32- SYSROOT_DIR should be /global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/arc-snps-elf/ ```
priority
cross compile toolchain variant doesn t working properly with multilib toolchain describe the bug in case of zephyr toolchain variant cross compile we don t set sysroot dir cmake variable so we don t execute block of code which sets libc library dir variable based on the toolchain c flags including mcpu flag so in case of multilib cross compile toolchain we always use default libraries set no mater which mcpu value we ve provided obviously it s incorrect impact implicit unexpected usage of libraries which are build for different mcpu than was specified if we are lucky we may get compile error for example if we got something not linkable like libraries from toolchain and bit object files from rest of zephyr however we may get successful linkage and use some not optimal libraries i e built for processor which use smaller instruction set or get runtime issues possible solutions add additional environment variable for cross compile toolchain variant which will be pointing to correct sysroot dir path try to extract sysroot dir path from the cross compile variable i ve checked the value of cross compile i ve previously used it doesn t seems to be feasible for these toolchains i e cross compile home user dev unknown elf unknown elf bin snps elf sysroot dir should be home user dev unknown elf unknown elf unknown elf cross compile global tools linux arc gnu prebuilt le bin arc sysroot dir should be global tools linux arc gnu prebuilt le arc snps elf
1
293,471
8,996,207,878
IssuesEvent
2019-02-02 00:04:08
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
USB Power Event Panic
area: USB bug priority: medium
**Describe the bug** A panic appears to occur on a USB power event. **To Reproduce** As best as I can tell, I've had a tough time reproducing this consistently, but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj.conf seems to cause the problem sometimes. My prj.conf for the app that causes the fault. ``` CONFIG_FLOAT=y CONFIG_COMPILER_OPT="-ffast-math" # GPIO Enable CONFIG_GPIO=y # Logging Config CONFIG_LOG=y CONFIG_LOG_PRINTK=y CONFIG_LOG_BACKEND_RTT=y CONFIG_ISR_STACK_SIZE=4096 # Workqueue stack needs to be larger for some tasks CONFIG_SYSTEM_WORKQUEUE_STACK_SIZE=8192 ## USB CONFIG_USB=y CONFIG_USB_DEVICE_STACK=y CONFIG_USB_DEVICE_PRODUCT="The Boss" CONFIG_USB_UART_CONSOLE=y CONFIG_UART_INTERRUPT_DRIVEN=y CONFIG_UART_LINE_CTRL=y ## RTT and SystemView CONFIG_USE_SEGGER_RTT=n CONFIG_RTT_CONSOLE=y CONFIG_SEGGER_SYSTEMVIEW=n # Debug help CONFIG_DEBUG=n CONFIG_ASSERT=n CONFIG_OUTPUT_DISASSEMBLY=y CONFIG_STACK_CANARIES=n CONFIG_STACK_SENTINEL=n CONFIG_DEBUG_INFO=y CONFIG_INIT_STACKS=n CONFIG_STACK_USAGE=n # Enable mcumgr. CONFIG_MCUMGR=y CONFIG_MCUMGR_SMP_BT=y #CONFIG_MCUMGR_SMP_SHELL=n #CONFIG_MCUMGR_SMP_UART=y CONFIG_MCUMGR_CMD_FS_MGMT=y CONFIG_MCUMGR_CMD_IMG_MGMT=y CONFIG_MCUMGR_CMD_OS_MGMT=y CONFIG_MCUMGR_CMD_STAT_MGMT=y # Ensure an MCUboot-compatible binary is generated. CONFIG_BOOTLOADER_MCUBOOT=n # Setup watchdog support CONFIG_WATCHDOG=n # Enable ADC for battery level monitoring CONFIG_ADC=y CONFIG_ADC_0=y CONFIG_ADC_NRFX_SAADC=y #CONFIG_ADC_NRFX_ADC_CHANNEL_COUNT=1 # Setup flash access and file system settings CONFIG_FLASH=y CONFIG_FLASH_PAGE_LAYOUT=y CONFIG_SOC_FLASH_NRF_RADIO_SYNC=y CONFIG_MPU_ALLOW_FLASH_WRITE=y CONFIG_FILE_SYSTEM=y CONFIG_FILE_SYSTEM_NFFS=y CONFIG_FS_NFFS_FLASH_DEV_NAME="NRF_FLASH_DRV_NAME" CONFIG_FS_NFFS_NUM_FILES=4 CONFIG_FS_NFFS_NUM_DIRS=4 CONFIG_FS_NFFS_NUM_INODES=1024 CONFIG_FS_NFFS_NUM_BLOCKS=1024 CONFIG_FS_NFFS_NUM_CACHE_INODES=1 CONFIG_FS_NFFS_NUM_CACHE_BLOCKS=1 CONFIG_FILE_SYSTEM_NFFS=y CONFIG_NFFS_FILESYSTEM_MAX_AREAS=12 # Settings CONFIG_SETTINGS=y CONFIG_SETTINGS_FS_DIR="/nffs/settings" CONFIG_SETTINGS_FS_FILE="/nffs/settings/run" # Bluetooth Config CONFIG_BT=y #CONFIG_BT_PERIPHERAL=y #CONFIG_BT_DEBUG_LOG=n #CONFIG_BT_DEBUG_CONN=y #CONFIG_BT_DEBUG_SETTINGS=y #CONFIG_BT_DEBUG_GATT=y #CONFIG_BT_DEBUG_ATT=y #CONFIG_BT_DEBUG_L2CAP=y #CONFIG_BT_DEBUG_HCI_CORE=y #CONFIG_BT_DEBUG_HCI_DRIVER=y #CONFIG_BT_DEBUG_MONITOR=n # dumps protocol on UART #CONFIG_BT_CTLR_PROFILE_ISR=y CONFIG_BT_DEVICE_NAME="My Device" #CONFIG_BT_DEVICE_NAME_DYNAMIC=y #CONFIG_BT_DEVICE_NAME_GATT_WRITABLE=y #CONFIG_BT_DEVICE_NAME_MAX=64 CONFIG_BT_RX_BUF_COUNT=4 CONFIG_BT_RX_BUF_LEN=251 CONFIG_BT_RX_STACK_SIZE=2048 #CONFIG_BT_SETTINGS=n CONFIG_BT_L2CAP_TX_BUF_COUNT=10 #CONFIG_BT_L2CAP_TX_FRAG_COUNT=4 CONFIG_BT_L2CAP_TX_MTU=251 #CONFIG_BT_CONN_TX_MAX=10 #CONFIG_BT_ATT_PREPARE_COUNT=10 CONFIG_BT_ATT_TX_MAX=10 #CONFIG_BT_SMP=n #CONFIG_BT_PRIVACY=n #CONFIG_BT_SIGNING=n #CONFIG_BT_CTLR=y CONFIG_BT_CTLR_TX_BUFFERS=10 CONFIG_BT_CTLR_RX_BUFFERS=4 CONFIG_BT_CTLR_TX_BUFFER_SIZE=251 CONFIG_BT_CTLR_DATA_LENGTH_MAX=251 CONFIG_BT_CTLR_ADVANCED_FEATURES=y CONFIG_BT_CTLR_WORKER_PRIO=0 CONFIG_BT_CTLR_JOB_PRIO=0 CONFIG_BT_CTLR_PHY_2M=y #toggle suppoort of 2M phy CONFIG_BT_CTLR_CONN_RSSI=y CONFIG_BT_CTLR_SCAN_REQ_RSSI=y # Bluetooth support of mcumgr requires a net_buf user_data size >= 7. CONFIG_NET_BUF_USER_DATA_SIZE=16 # Required by the `taskstat` command. CONFIG_THREAD_MONITOR=y # Enable statistics and statistic names. CONFIG_STATS=y CONFIG_STATS_NAMES=y CONFIG_I2C=y CONFIG_I2C_NRFX=y CONFIG_I2C_0=y CONFIG_I2C_0_NRF_TWIM=y CONFIG_SPI=y CONFIG_SPI_NRFX=y CONFIG_SPI_1=y CONFIG_SPI_1_NRF_SPIM=y CONFIG_SENSOR=y # Magnetometer Sensor CONFIG_LIS3MDL=y CONFIG_LIS3MDL_TRIGGER=y CONFIG_LIS3MDL_THREAD_PRIORITY=0 CONFIG_LIS3MDL_TRIGGER_OWN_THREAD=y CONFIG_LIS3MDL_TRIGGER_IRQ=n CONFIG_LIS3MDL_TRIGGER_GLOBAL_THREAD=n CONFIG_LIS3MDL_DATA_RAW=y CONFIG_LIS3MDL_DATA_CONVERTED=n CONFIG_LIS3MDL_ODR="1000" CONFIG_LIS3MDL_FS=4 # LP5521 LED controller CONFIG_LED=y CONFIG_LP5521=y ``` **Expected behavior** No panics! **Impact** This prevents my application from working with mcuboot, which prevents OTA updates. **Screenshots or console output** GDB output as I was debugging this problem ``` (gdb) break usb_dc_nrfx_power_event_callback Breakpoint 1 at 0x14a60: file /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c, line 477. (gdb) break usb_enable Breakpoint 2 at 0x126a4: file /home/tburdick/src/zephyr/subsys/usb/usb_device.c, line 999. (gdb) monitor reset 2 mResets core & peripherals using RESET pin. (gdb) monitor reset Resetting target (gdb) c Continuing. Breakpoint 1, usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:477 477 switch (event) { (gdb) print usbd_ctx $1 = {status_cb = 0x0, attached = false, ready = false, usb_work = {_reserved = 0x0, handler = 0x0, flags = {0}}, work_queue = {_queue = {data_q = {head = 0x0, tail = 0x0}, {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, poll_events = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}}}, drv_lock = {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, owner = 0x0, lock_count = 0, owner_orig_prio = 0}, ep_ctx = {{cfg = {cb = 0x0, max_sz = 0, en = false, addr = 0 '\000', type = USB_DC_EP_CONTROL}, buf = {len = 0, block = {data = 0x0, id = {pool = 0, level = 0, block = 0}}, data = 0x0, curr = 0x0}, read_complete = false, read_pending = false, write_in_progress = false, write_fragmented = false} <repeats 18 times>}, ctrl_read_len = 0} (gdb) break _Sys _SysFatalErrorHandler _SysNmiOnReset (gdb) break _SysFatalErrorHandler Breakpoint 3 at 0x156ac: file /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c, line 52. (gdb) c Continuing. Breakpoint 3, _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52 52 if (reason == _NANO_ERR_KERNEL_PANIC) { (gdb) bt #0 _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52 #1 0x00015636 in __usage_fault () at /home/tburdick/src/zephyr/arch/arm/core/fault_s.S:143 #2 <signal handler called> #3 signal_poll_event (event=0x0, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:342 #4 0x000309a4 in _handle_obj_poll_events (events=events@entry=0x20027164 <usbd_ctx+28>, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:371 #5 0x0002f790 in handle_poll_events (state=4, queue=0x2002715c <usbd_ctx+20>) at /home/tburdick/src/zephyr/kernel/queue.c:119 #6 queue_insert (queue=queue@entry=0x2002715c <usbd_ctx+20>, prev=0x0, data=<optimized out>, alloc=alloc@entry=false) at /home/tburdick/src/zephyr/kernel/queue.c:179 #7 0x0002f7ee in k_queue_append (queue=queue@entry=0x2002715c <usbd_ctx+20>, data=<optimized out>) at /home/tburdick/src/zephyr/kernel/queue.c:193 #8 0x000141ac in usbd_evt_put (ev=<optimized out>) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:395 #9 0x00014a88 in usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:501 #10 0x00013a56 in power_event_cb (event=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:275 #11 _power_clock_isr (arg=0x0) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:415 #12 0x0001577a in _isr_wrapper () at /home/tburdick/src/zephyr/arch/arm/core/isr_wrapper.S:120 #13 <signal handler called> #14 0x17fb7d56 in ?? () #15 0x28311824 in ?? () Backtrace stopped: previous frame identical to this frame (corrupt stack?) (gdb) ``` **Environment (please complete the following information):** - gcc 6.2.1 **Additional context** It appears that sometimes a usb power event occurs before the usbd_ctx has been initialized from usb_dc_attach. I think that is the cause of the problem, because then the usb_work struct is likely to be uninitialized. I'd guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an IRQ, this IRQ must be occuring before the subsequent usb_enable is called. Why that happens beyond my current understanding of zephyr and the usb device driver.
1.0
USB Power Event Panic - **Describe the bug** A panic appears to occur on a USB power event. **To Reproduce** As best as I can tell, I've had a tough time reproducing this consistently, but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj.conf seems to cause the problem sometimes. My prj.conf for the app that causes the fault. ``` CONFIG_FLOAT=y CONFIG_COMPILER_OPT="-ffast-math" # GPIO Enable CONFIG_GPIO=y # Logging Config CONFIG_LOG=y CONFIG_LOG_PRINTK=y CONFIG_LOG_BACKEND_RTT=y CONFIG_ISR_STACK_SIZE=4096 # Workqueue stack needs to be larger for some tasks CONFIG_SYSTEM_WORKQUEUE_STACK_SIZE=8192 ## USB CONFIG_USB=y CONFIG_USB_DEVICE_STACK=y CONFIG_USB_DEVICE_PRODUCT="The Boss" CONFIG_USB_UART_CONSOLE=y CONFIG_UART_INTERRUPT_DRIVEN=y CONFIG_UART_LINE_CTRL=y ## RTT and SystemView CONFIG_USE_SEGGER_RTT=n CONFIG_RTT_CONSOLE=y CONFIG_SEGGER_SYSTEMVIEW=n # Debug help CONFIG_DEBUG=n CONFIG_ASSERT=n CONFIG_OUTPUT_DISASSEMBLY=y CONFIG_STACK_CANARIES=n CONFIG_STACK_SENTINEL=n CONFIG_DEBUG_INFO=y CONFIG_INIT_STACKS=n CONFIG_STACK_USAGE=n # Enable mcumgr. CONFIG_MCUMGR=y CONFIG_MCUMGR_SMP_BT=y #CONFIG_MCUMGR_SMP_SHELL=n #CONFIG_MCUMGR_SMP_UART=y CONFIG_MCUMGR_CMD_FS_MGMT=y CONFIG_MCUMGR_CMD_IMG_MGMT=y CONFIG_MCUMGR_CMD_OS_MGMT=y CONFIG_MCUMGR_CMD_STAT_MGMT=y # Ensure an MCUboot-compatible binary is generated. CONFIG_BOOTLOADER_MCUBOOT=n # Setup watchdog support CONFIG_WATCHDOG=n # Enable ADC for battery level monitoring CONFIG_ADC=y CONFIG_ADC_0=y CONFIG_ADC_NRFX_SAADC=y #CONFIG_ADC_NRFX_ADC_CHANNEL_COUNT=1 # Setup flash access and file system settings CONFIG_FLASH=y CONFIG_FLASH_PAGE_LAYOUT=y CONFIG_SOC_FLASH_NRF_RADIO_SYNC=y CONFIG_MPU_ALLOW_FLASH_WRITE=y CONFIG_FILE_SYSTEM=y CONFIG_FILE_SYSTEM_NFFS=y CONFIG_FS_NFFS_FLASH_DEV_NAME="NRF_FLASH_DRV_NAME" CONFIG_FS_NFFS_NUM_FILES=4 CONFIG_FS_NFFS_NUM_DIRS=4 CONFIG_FS_NFFS_NUM_INODES=1024 CONFIG_FS_NFFS_NUM_BLOCKS=1024 CONFIG_FS_NFFS_NUM_CACHE_INODES=1 CONFIG_FS_NFFS_NUM_CACHE_BLOCKS=1 CONFIG_FILE_SYSTEM_NFFS=y CONFIG_NFFS_FILESYSTEM_MAX_AREAS=12 # Settings CONFIG_SETTINGS=y CONFIG_SETTINGS_FS_DIR="/nffs/settings" CONFIG_SETTINGS_FS_FILE="/nffs/settings/run" # Bluetooth Config CONFIG_BT=y #CONFIG_BT_PERIPHERAL=y #CONFIG_BT_DEBUG_LOG=n #CONFIG_BT_DEBUG_CONN=y #CONFIG_BT_DEBUG_SETTINGS=y #CONFIG_BT_DEBUG_GATT=y #CONFIG_BT_DEBUG_ATT=y #CONFIG_BT_DEBUG_L2CAP=y #CONFIG_BT_DEBUG_HCI_CORE=y #CONFIG_BT_DEBUG_HCI_DRIVER=y #CONFIG_BT_DEBUG_MONITOR=n # dumps protocol on UART #CONFIG_BT_CTLR_PROFILE_ISR=y CONFIG_BT_DEVICE_NAME="My Device" #CONFIG_BT_DEVICE_NAME_DYNAMIC=y #CONFIG_BT_DEVICE_NAME_GATT_WRITABLE=y #CONFIG_BT_DEVICE_NAME_MAX=64 CONFIG_BT_RX_BUF_COUNT=4 CONFIG_BT_RX_BUF_LEN=251 CONFIG_BT_RX_STACK_SIZE=2048 #CONFIG_BT_SETTINGS=n CONFIG_BT_L2CAP_TX_BUF_COUNT=10 #CONFIG_BT_L2CAP_TX_FRAG_COUNT=4 CONFIG_BT_L2CAP_TX_MTU=251 #CONFIG_BT_CONN_TX_MAX=10 #CONFIG_BT_ATT_PREPARE_COUNT=10 CONFIG_BT_ATT_TX_MAX=10 #CONFIG_BT_SMP=n #CONFIG_BT_PRIVACY=n #CONFIG_BT_SIGNING=n #CONFIG_BT_CTLR=y CONFIG_BT_CTLR_TX_BUFFERS=10 CONFIG_BT_CTLR_RX_BUFFERS=4 CONFIG_BT_CTLR_TX_BUFFER_SIZE=251 CONFIG_BT_CTLR_DATA_LENGTH_MAX=251 CONFIG_BT_CTLR_ADVANCED_FEATURES=y CONFIG_BT_CTLR_WORKER_PRIO=0 CONFIG_BT_CTLR_JOB_PRIO=0 CONFIG_BT_CTLR_PHY_2M=y #toggle suppoort of 2M phy CONFIG_BT_CTLR_CONN_RSSI=y CONFIG_BT_CTLR_SCAN_REQ_RSSI=y # Bluetooth support of mcumgr requires a net_buf user_data size >= 7. CONFIG_NET_BUF_USER_DATA_SIZE=16 # Required by the `taskstat` command. CONFIG_THREAD_MONITOR=y # Enable statistics and statistic names. CONFIG_STATS=y CONFIG_STATS_NAMES=y CONFIG_I2C=y CONFIG_I2C_NRFX=y CONFIG_I2C_0=y CONFIG_I2C_0_NRF_TWIM=y CONFIG_SPI=y CONFIG_SPI_NRFX=y CONFIG_SPI_1=y CONFIG_SPI_1_NRF_SPIM=y CONFIG_SENSOR=y # Magnetometer Sensor CONFIG_LIS3MDL=y CONFIG_LIS3MDL_TRIGGER=y CONFIG_LIS3MDL_THREAD_PRIORITY=0 CONFIG_LIS3MDL_TRIGGER_OWN_THREAD=y CONFIG_LIS3MDL_TRIGGER_IRQ=n CONFIG_LIS3MDL_TRIGGER_GLOBAL_THREAD=n CONFIG_LIS3MDL_DATA_RAW=y CONFIG_LIS3MDL_DATA_CONVERTED=n CONFIG_LIS3MDL_ODR="1000" CONFIG_LIS3MDL_FS=4 # LP5521 LED controller CONFIG_LED=y CONFIG_LP5521=y ``` **Expected behavior** No panics! **Impact** This prevents my application from working with mcuboot, which prevents OTA updates. **Screenshots or console output** GDB output as I was debugging this problem ``` (gdb) break usb_dc_nrfx_power_event_callback Breakpoint 1 at 0x14a60: file /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c, line 477. (gdb) break usb_enable Breakpoint 2 at 0x126a4: file /home/tburdick/src/zephyr/subsys/usb/usb_device.c, line 999. (gdb) monitor reset 2 mResets core & peripherals using RESET pin. (gdb) monitor reset Resetting target (gdb) c Continuing. Breakpoint 1, usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:477 477 switch (event) { (gdb) print usbd_ctx $1 = {status_cb = 0x0, attached = false, ready = false, usb_work = {_reserved = 0x0, handler = 0x0, flags = {0}}, work_queue = {_queue = {data_q = {head = 0x0, tail = 0x0}, {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, poll_events = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}}}, drv_lock = {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, owner = 0x0, lock_count = 0, owner_orig_prio = 0}, ep_ctx = {{cfg = {cb = 0x0, max_sz = 0, en = false, addr = 0 '\000', type = USB_DC_EP_CONTROL}, buf = {len = 0, block = {data = 0x0, id = {pool = 0, level = 0, block = 0}}, data = 0x0, curr = 0x0}, read_complete = false, read_pending = false, write_in_progress = false, write_fragmented = false} <repeats 18 times>}, ctrl_read_len = 0} (gdb) break _Sys _SysFatalErrorHandler _SysNmiOnReset (gdb) break _SysFatalErrorHandler Breakpoint 3 at 0x156ac: file /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c, line 52. (gdb) c Continuing. Breakpoint 3, _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52 52 if (reason == _NANO_ERR_KERNEL_PANIC) { (gdb) bt #0 _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52 #1 0x00015636 in __usage_fault () at /home/tburdick/src/zephyr/arch/arm/core/fault_s.S:143 #2 <signal handler called> #3 signal_poll_event (event=0x0, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:342 #4 0x000309a4 in _handle_obj_poll_events (events=events@entry=0x20027164 <usbd_ctx+28>, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:371 #5 0x0002f790 in handle_poll_events (state=4, queue=0x2002715c <usbd_ctx+20>) at /home/tburdick/src/zephyr/kernel/queue.c:119 #6 queue_insert (queue=queue@entry=0x2002715c <usbd_ctx+20>, prev=0x0, data=<optimized out>, alloc=alloc@entry=false) at /home/tburdick/src/zephyr/kernel/queue.c:179 #7 0x0002f7ee in k_queue_append (queue=queue@entry=0x2002715c <usbd_ctx+20>, data=<optimized out>) at /home/tburdick/src/zephyr/kernel/queue.c:193 #8 0x000141ac in usbd_evt_put (ev=<optimized out>) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:395 #9 0x00014a88 in usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:501 #10 0x00013a56 in power_event_cb (event=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:275 #11 _power_clock_isr (arg=0x0) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:415 #12 0x0001577a in _isr_wrapper () at /home/tburdick/src/zephyr/arch/arm/core/isr_wrapper.S:120 #13 <signal handler called> #14 0x17fb7d56 in ?? () #15 0x28311824 in ?? () Backtrace stopped: previous frame identical to this frame (corrupt stack?) (gdb) ``` **Environment (please complete the following information):** - gcc 6.2.1 **Additional context** It appears that sometimes a usb power event occurs before the usbd_ctx has been initialized from usb_dc_attach. I think that is the cause of the problem, because then the usb_work struct is likely to be uninitialized. I'd guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an IRQ, this IRQ must be occuring before the subsequent usb_enable is called. Why that happens beyond my current understanding of zephyr and the usb device driver.
priority
usb power event panic describe the bug a panic appears to occur on a usb power event to reproduce as best as i can tell i ve had a tough time reproducing this consistently but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj conf seems to cause the problem sometimes my prj conf for the app that causes the fault config float y config compiler opt ffast math gpio enable config gpio y logging config config log y config log printk y config log backend rtt y config isr stack size workqueue stack needs to be larger for some tasks config system workqueue stack size usb config usb y config usb device stack y config usb device product the boss config usb uart console y config uart interrupt driven y config uart line ctrl y rtt and systemview config use segger rtt n config rtt console y config segger systemview n debug help config debug n config assert n config output disassembly y config stack canaries n config stack sentinel n config debug info y config init stacks n config stack usage n enable mcumgr config mcumgr y config mcumgr smp bt y config mcumgr smp shell n config mcumgr smp uart y config mcumgr cmd fs mgmt y config mcumgr cmd img mgmt y config mcumgr cmd os mgmt y config mcumgr cmd stat mgmt y ensure an mcuboot compatible binary is generated config bootloader mcuboot n setup watchdog support config watchdog n enable adc for battery level monitoring config adc y config adc y config adc nrfx saadc y config adc nrfx adc channel count setup flash access and file system settings config flash y config flash page layout y config soc flash nrf radio sync y config mpu allow flash write y config file system y config file system nffs y config fs nffs flash dev name nrf flash drv name config fs nffs num files config fs nffs num dirs config fs nffs num inodes config fs nffs num blocks config fs nffs num cache inodes config fs nffs num cache blocks config file system nffs y config nffs filesystem max areas settings config settings y config settings fs dir nffs settings config settings fs file nffs settings run bluetooth config config bt y config bt peripheral y config bt debug log n config bt debug conn y config bt debug settings y config bt debug gatt y config bt debug att y config bt debug y config bt debug hci core y config bt debug hci driver y config bt debug monitor n dumps protocol on uart config bt ctlr profile isr y config bt device name my device config bt device name dynamic y config bt device name gatt writable y config bt device name max config bt rx buf count config bt rx buf len config bt rx stack size config bt settings n config bt tx buf count config bt tx frag count config bt tx mtu config bt conn tx max config bt att prepare count config bt att tx max config bt smp n config bt privacy n config bt signing n config bt ctlr y config bt ctlr tx buffers config bt ctlr rx buffers config bt ctlr tx buffer size config bt ctlr data length max config bt ctlr advanced features y config bt ctlr worker prio config bt ctlr job prio config bt ctlr phy y toggle suppoort of phy config bt ctlr conn rssi y config bt ctlr scan req rssi y bluetooth support of mcumgr requires a net buf user data size config net buf user data size required by the taskstat command config thread monitor y enable statistics and statistic names config stats y config stats names y config y config nrfx y config y config nrf twim y config spi y config spi nrfx y config spi y config spi nrf spim y config sensor y magnetometer sensor config y config trigger y config thread priority config trigger own thread y config trigger irq n config trigger global thread n config data raw y config data converted n config odr config fs led controller config led y config y expected behavior no panics impact this prevents my application from working with mcuboot which prevents ota updates screenshots or console output gdb output as i was debugging this problem gdb break usb dc nrfx power event callback breakpoint at file home tburdick src zephyr drivers usb device usb dc nrfx c line gdb break usb enable breakpoint at file home tburdick src zephyr subsys usb usb device c line gdb monitor reset mresets core peripherals using reset pin gdb monitor reset resetting target gdb c continuing breakpoint usb dc nrfx power event callback event event entry nrf power event usbdetected at home tburdick src zephyr drivers usb device usb dc nrfx c switch event gdb print usbd ctx status cb attached false ready false usb work reserved handler flags work queue queue data q head tail wait q waitq head next tail prev poll events head next tail prev drv lock wait q waitq head next tail prev owner lock count owner orig prio ep ctx cfg cb max sz en false addr type usb dc ep control buf len block data id pool level block data curr read complete false read pending false write in progress false write fragmented false ctrl read len gdb break sys sysfatalerrorhandler sysnmionreset gdb break sysfatalerrorhandler breakpoint at file home tburdick src zephyr arch arm core sys fatal error handler c line gdb c continuing breakpoint sysfatalerrorhandler reason pesf at home tburdick src zephyr arch arm core sys fatal error handler c if reason nano err kernel panic gdb bt sysfatalerrorhandler reason pesf at home tburdick src zephyr arch arm core sys fatal error handler c in usage fault at home tburdick src zephyr arch arm core fault s s signal poll event event state state entry at home tburdick src zephyr kernel poll c in handle obj poll events events events entry state state entry at home tburdick src zephyr kernel poll c in handle poll events state queue at home tburdick src zephyr kernel queue c queue insert queue queue entry prev data alloc alloc entry false at home tburdick src zephyr kernel queue c in k queue append queue queue entry data at home tburdick src zephyr kernel queue c in usbd evt put ev at home tburdick src zephyr drivers usb device usb dc nrfx c in usb dc nrfx power event callback event event entry nrf power event usbdetected at home tburdick src zephyr drivers usb device usb dc nrfx c in power event cb event nrf power event usbdetected at home tburdick src zephyr drivers clock control nrf power clock c power clock isr arg at home tburdick src zephyr drivers clock control nrf power clock c in isr wrapper at home tburdick src zephyr arch arm core isr wrapper s in in backtrace stopped previous frame identical to this frame corrupt stack gdb environment please complete the following information gcc additional context it appears that sometimes a usb power event occurs before the usbd ctx has been initialized from usb dc attach i think that is the cause of the problem because then the usb work struct is likely to be uninitialized i d guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an irq this irq must be occuring before the subsequent usb enable is called why that happens beyond my current understanding of zephyr and the usb device driver
1
298,769
9,201,112,713
IssuesEvent
2019-03-07 18:46:54
trimstray/htrace.sh
https://api.github.com/repos/trimstray/htrace.sh
closed
Wiki: updated 'Examples' chapter
Priority: Medium Status: Completed Type: Enhancement
Wiki [Examples](https://github.com/trimstray/htrace.sh/wiki/Examples) chapter. - [x] - Check redirects - [x] - POST method
1.0
Wiki: updated 'Examples' chapter - Wiki [Examples](https://github.com/trimstray/htrace.sh/wiki/Examples) chapter. - [x] - Check redirects - [x] - POST method
priority
wiki updated examples chapter wiki chapter check redirects post method
1
261,381
8,230,606,048
IssuesEvent
2018-09-07 13:29:51
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
Work items created in Analytics Insights can't be opened in Chromium and Firefox.
SEV3-medium area/browser-compatibility priority/P2 status/in-progress team/analytics type/bug
The work item created in the Analytics Insights menu can't be opened, see blank screen and Chrome log: "Redirect ... was blocked by CORS policy...". And I can't see it in the Planner backlog. Scenario: 1) Open 'View complete stack report' from 'Stack Report Recommendations' 2) Select 'Insights' and tab 'Companion Dependency Details' 3) Select a dependency and click 'Create work item' 4) Click 'View here' after the WI has been created 5) I got forwarded to a black page, see picture below 6) In addition the WI is not present in the Planner backlog Tested on Chrome 56.0.2924.87 Fedora Project (64-bit) and Firefox. ![insights_wi_created](https://user-images.githubusercontent.com/26871874/39049800-3bdc15d8-44a3-11e8-8164-bf283491e1ec.png) ![insights_wi_view](https://user-images.githubusercontent.com/26871874/39049824-4f944cd0-44a3-11e8-8a5a-892417a028f5.png)
1.0
Work items created in Analytics Insights can't be opened in Chromium and Firefox. - The work item created in the Analytics Insights menu can't be opened, see blank screen and Chrome log: "Redirect ... was blocked by CORS policy...". And I can't see it in the Planner backlog. Scenario: 1) Open 'View complete stack report' from 'Stack Report Recommendations' 2) Select 'Insights' and tab 'Companion Dependency Details' 3) Select a dependency and click 'Create work item' 4) Click 'View here' after the WI has been created 5) I got forwarded to a black page, see picture below 6) In addition the WI is not present in the Planner backlog Tested on Chrome 56.0.2924.87 Fedora Project (64-bit) and Firefox. ![insights_wi_created](https://user-images.githubusercontent.com/26871874/39049800-3bdc15d8-44a3-11e8-8164-bf283491e1ec.png) ![insights_wi_view](https://user-images.githubusercontent.com/26871874/39049824-4f944cd0-44a3-11e8-8a5a-892417a028f5.png)
priority
work items created in analytics insights can t be opened in chromium and firefox the work item created in the analytics insights menu can t be opened see blank screen and chrome log redirect was blocked by cors policy and i can t see it in the planner backlog scenario open view complete stack report from stack report recommendations select insights and tab companion dependency details select a dependency and click create work item click view here after the wi has been created i got forwarded to a black page see picture below in addition the wi is not present in the planner backlog tested on chrome fedora project bit and firefox
1
753,930
26,367,686,186
IssuesEvent
2023-01-11 17:52:05
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
yb-admin list_snapshots payload size limited
kind/bug area/docdb priority/medium
Jira Link: [DB-2362](https://yugabyte.atlassian.net/browse/DB-2362) There is a hardcoded limit that effectively gates rpc_max_message_size at the protobuf bytes limit. `yb-admin -rpc_max_message_size 1500000000 list_snapshots` `tcmalloc: large alloc 1256480768 bytes == 0x2bac000 @ 0x7f8621f5c6d3 0x7f8621f7d09b 0x7f8629ae5e58 0x7f8629b63d28 0x7f8629aea522 0x7f8629b54cdc 0x7f8629b58a56 0x7f8629b595b6 0x7f8624d80b7b 0x7f8624d84c8a 0x7f8629b1587c 0x7f862639c3df 0x7f8620dcb694 [libprotobuf ERROR google/protobuf/io/coded_stream.cc:193] A protocol message was rejected because it was too big (more than 535822336 bytes). To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h. Error: IO error (yb/rpc/outbound_call.cc:423): Unable to list snapshots: Invalid response, missing fields`
1.0
yb-admin list_snapshots payload size limited - Jira Link: [DB-2362](https://yugabyte.atlassian.net/browse/DB-2362) There is a hardcoded limit that effectively gates rpc_max_message_size at the protobuf bytes limit. `yb-admin -rpc_max_message_size 1500000000 list_snapshots` `tcmalloc: large alloc 1256480768 bytes == 0x2bac000 @ 0x7f8621f5c6d3 0x7f8621f7d09b 0x7f8629ae5e58 0x7f8629b63d28 0x7f8629aea522 0x7f8629b54cdc 0x7f8629b58a56 0x7f8629b595b6 0x7f8624d80b7b 0x7f8624d84c8a 0x7f8629b1587c 0x7f862639c3df 0x7f8620dcb694 [libprotobuf ERROR google/protobuf/io/coded_stream.cc:193] A protocol message was rejected because it was too big (more than 535822336 bytes). To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h. Error: IO error (yb/rpc/outbound_call.cc:423): Unable to list snapshots: Invalid response, missing fields`
priority
yb admin list snapshots payload size limited jira link there is a hardcoded limit that effectively gates rpc max message size at the protobuf bytes limit yb admin rpc max message size list snapshots tcmalloc large alloc bytes a protocol message was rejected because it was too big more than bytes to increase the limit or to disable these warnings see codedinputstream settotalbyteslimit in google protobuf io coded stream h error io error yb rpc outbound call cc unable to list snapshots invalid response missing fields
1
55,590
3,073,804,213
IssuesEvent
2015-08-20 00:40:47
RobotiumTech/robotium
https://api.github.com/repos/RobotiumTech/robotium
closed
Unable to entertext in Webview page using robotium 4.2
bug imported invalid Priority-Medium
_From [sinha.ka...@gmail.com](https://code.google.com/u/103828779966021760852/) on July 26, 2013 11:17:22_ package com.calculator.test; import org.junit.Test; import android.test.ActivityInstrumentationTestCase2; import com.jayway.android.robotium.solo.*; import com.jayway.android.robotium.solo.By; import com.jayway.android.robotium.solo.Solo; @SuppressWarnings("rawtypes") public class SMPage extends ActivityInstrumentationTestCase2{ private static final String LAUNCHER_ACTIVITY_FULL_CLASSNAME = "com.aha.android.app.login.WelcomePageActivity"; private Solo solo; private static Class launcherActivityClass; static{ try { launcherActivityClass = Class.forName(LAUNCHER_ACTIVITY_FULL_CLASSNAME); } catch (ClassNotFoundException e) { throw new RuntimeException(e); } } @SuppressWarnings("unchecked") public SMPage() { super(launcherActivityClass); } @Override public void setUp() throws Exception { //setUp() is run before a test case is started. //This is where the solo object is created. solo = new Solo(getInstrumentation(), getActivity()); } @Override public void tearDown() throws Exception { //tearDown() is run after a test case has finished. //finishOpenedActivities() will finish all the activities that have been opened during the test execution. solo.finishOpenedActivities(); } @Test public void testLaunch() throws Exception { solo.clickOnText("Login"); solo.clickOnText("Forgot your password?"); Thread.sleep(50000); solo.typeTextInWebElement(By.name("email"), "sinha.kanak@gmail.com"); solo.clickOnText("Submit"); Thread.sleep(5000 } } Not able to enter text in the webview page . Screen shot attached of the page. Using robotium 4.2. **Attachment:** [password.pdf](http://code.google.com/p/robotium/issues/detail?id=493) _Original issue: http://code.google.com/p/robotium/issues/detail?id=493_
1.0
Unable to entertext in Webview page using robotium 4.2 - _From [sinha.ka...@gmail.com](https://code.google.com/u/103828779966021760852/) on July 26, 2013 11:17:22_ package com.calculator.test; import org.junit.Test; import android.test.ActivityInstrumentationTestCase2; import com.jayway.android.robotium.solo.*; import com.jayway.android.robotium.solo.By; import com.jayway.android.robotium.solo.Solo; @SuppressWarnings("rawtypes") public class SMPage extends ActivityInstrumentationTestCase2{ private static final String LAUNCHER_ACTIVITY_FULL_CLASSNAME = "com.aha.android.app.login.WelcomePageActivity"; private Solo solo; private static Class launcherActivityClass; static{ try { launcherActivityClass = Class.forName(LAUNCHER_ACTIVITY_FULL_CLASSNAME); } catch (ClassNotFoundException e) { throw new RuntimeException(e); } } @SuppressWarnings("unchecked") public SMPage() { super(launcherActivityClass); } @Override public void setUp() throws Exception { //setUp() is run before a test case is started. //This is where the solo object is created. solo = new Solo(getInstrumentation(), getActivity()); } @Override public void tearDown() throws Exception { //tearDown() is run after a test case has finished. //finishOpenedActivities() will finish all the activities that have been opened during the test execution. solo.finishOpenedActivities(); } @Test public void testLaunch() throws Exception { solo.clickOnText("Login"); solo.clickOnText("Forgot your password?"); Thread.sleep(50000); solo.typeTextInWebElement(By.name("email"), "sinha.kanak@gmail.com"); solo.clickOnText("Submit"); Thread.sleep(5000 } } Not able to enter text in the webview page . Screen shot attached of the page. Using robotium 4.2. **Attachment:** [password.pdf](http://code.google.com/p/robotium/issues/detail?id=493) _Original issue: http://code.google.com/p/robotium/issues/detail?id=493_
priority
unable to entertext in webview page using robotium from on july package com calculator test import org junit test import android test import com jayway android robotium solo import com jayway android robotium solo by import com jayway android robotium solo solo suppresswarnings rawtypes public class smpage extends private static final string launcher activity full classname com aha android app login welcomepageactivity private solo solo private static class launcheractivityclass static try launcheractivityclass class forname launcher activity full classname catch classnotfoundexception e throw new runtimeexception e suppresswarnings unchecked public smpage super launcheractivityclass override public void setup throws exception setup is run before a test case is started this is where the solo object is created solo new solo getinstrumentation getactivity override public void teardown throws exception teardown is run after a test case has finished finishopenedactivities will finish all the activities that have been opened during the test execution solo finishopenedactivities test public void testlaunch throws exception solo clickontext login solo clickontext forgot your password thread sleep solo typetextinwebelement by name email sinha kanak gmail com solo clickontext submit thread sleep not able to enter text in the webview page screen shot attached of the page using robotium attachment original issue
1
285,547
8,766,458,207
IssuesEvent
2018-12-17 16:52:14
spacetelescope/jwql
https://api.github.com/repos/spacetelescope/jwql
closed
Build script execution status display on dashboard
Medium Priority Web Application enhancement
The dashboard on the web application should display the success/failure of the various `cron` jobs (e.g. `monitor_filesystem`) so that the JWQL team can easily identify scripts that are failing or have not executed when they should have. Here is an example from what WFC3/Quicklook uses: <img width="676" alt="screen shot 2018-09-17 at 16 05 56" src="https://user-images.githubusercontent.com/2250769/45647348-aa01be80-ba93-11e8-8ea1-2ffe314f4a8f.png"> Each row in the table corresponds to a particular `cron` job. A user can click on the name of the monitor to display the contents of the corresponding log file. The color indicates how recently the script executed successfully; bright green means that the script executed successfully that day, light green means the script executed successfully several days ago, red indicates that the script crashed.
1.0
Build script execution status display on dashboard - The dashboard on the web application should display the success/failure of the various `cron` jobs (e.g. `monitor_filesystem`) so that the JWQL team can easily identify scripts that are failing or have not executed when they should have. Here is an example from what WFC3/Quicklook uses: <img width="676" alt="screen shot 2018-09-17 at 16 05 56" src="https://user-images.githubusercontent.com/2250769/45647348-aa01be80-ba93-11e8-8ea1-2ffe314f4a8f.png"> Each row in the table corresponds to a particular `cron` job. A user can click on the name of the monitor to display the contents of the corresponding log file. The color indicates how recently the script executed successfully; bright green means that the script executed successfully that day, light green means the script executed successfully several days ago, red indicates that the script crashed.
priority
build script execution status display on dashboard the dashboard on the web application should display the success failure of the various cron jobs e g monitor filesystem so that the jwql team can easily identify scripts that are failing or have not executed when they should have here is an example from what quicklook uses img width alt screen shot at src each row in the table corresponds to a particular cron job a user can click on the name of the monitor to display the contents of the corresponding log file the color indicates how recently the script executed successfully bright green means that the script executed successfully that day light green means the script executed successfully several days ago red indicates that the script crashed
1
140,312
5,400,079,574
IssuesEvent
2017-02-27 21:06:55
canadainc/sunnah10
https://api.github.com/repos/canadainc/sunnah10
closed
Implement individuals table
Admin auto-migrated enhancement logic maintainability Priority-Medium ui
``` Just like Quran10 ``` Original issue reported on code.google.com by `canadai...@gmail.com` on 21 Feb 2015 at 5:09
1.0
Implement individuals table - ``` Just like Quran10 ``` Original issue reported on code.google.com by `canadai...@gmail.com` on 21 Feb 2015 at 5:09
priority
implement individuals table just like original issue reported on code google com by canadai gmail com on feb at
1
234,761
7,726,077,604
IssuesEvent
2018-05-24 20:02:47
DarkPacks/SevTech-Ages
https://api.github.com/repos/DarkPacks/SevTech-Ages
closed
Lava crystal gets eaten inside ie alloy kilns
Category: Mod Priority: Medium Status: Review Needed Type: Bug
## Issue / Bug if you place a lava crystal inside the alloy kiln as a fuel source, the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp. ## Expected Behavior When placing a lava crystal inside a fuel source it should drain 50lp and register it as 1 coal's worth of fuel, as stated in the blood magic's sanguine scientiem. ![image](https://user-images.githubusercontent.com/13581892/39109113-546f154e-4680-11e8-82ba-b433619b4da3.png) ![image](https://user-images.githubusercontent.com/13581892/39109131-642f7a82-4680-11e8-896d-82c6b4480307.png) ![image](https://user-images.githubusercontent.com/13581892/39109152-75c4c57c-4680-11e8-9fdb-375981c60883.png) ## Possible Solution i have no idea how to solve it... i just know how it's supposed to work D= ## Steps to Reproduce 1. Place lava crystal in fuel slot in alloy kiln 2. Add in valid alloyable metals 3. watch lava crystal get eaten D= ## Client Information ![image](https://user-images.githubusercontent.com/13581892/39109248-e0d3a43c-4680-11e8-9241-ea00a8e5e772.png) * Modpack Version: 3.0.6 * Java Version: see image * Launcher Used: twitch launcher * Memory Allocated: see image * Server/LAN/Single Player: single player * Resourcepack Enabled?: no * Optifine Installed?: no
1.0
Lava crystal gets eaten inside ie alloy kilns - ## Issue / Bug if you place a lava crystal inside the alloy kiln as a fuel source, the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp. ## Expected Behavior When placing a lava crystal inside a fuel source it should drain 50lp and register it as 1 coal's worth of fuel, as stated in the blood magic's sanguine scientiem. ![image](https://user-images.githubusercontent.com/13581892/39109113-546f154e-4680-11e8-82ba-b433619b4da3.png) ![image](https://user-images.githubusercontent.com/13581892/39109131-642f7a82-4680-11e8-896d-82c6b4480307.png) ![image](https://user-images.githubusercontent.com/13581892/39109152-75c4c57c-4680-11e8-9fdb-375981c60883.png) ## Possible Solution i have no idea how to solve it... i just know how it's supposed to work D= ## Steps to Reproduce 1. Place lava crystal in fuel slot in alloy kiln 2. Add in valid alloyable metals 3. watch lava crystal get eaten D= ## Client Information ![image](https://user-images.githubusercontent.com/13581892/39109248-e0d3a43c-4680-11e8-9241-ea00a8e5e772.png) * Modpack Version: 3.0.6 * Java Version: see image * Launcher Used: twitch launcher * Memory Allocated: see image * Server/LAN/Single Player: single player * Resourcepack Enabled?: no * Optifine Installed?: no
priority
lava crystal gets eaten inside ie alloy kilns issue bug if you place a lava crystal inside the alloy kiln as a fuel source the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp expected behavior when placing a lava crystal inside a fuel source it should drain and register it as coal s worth of fuel as stated in the blood magic s sanguine scientiem possible solution i have no idea how to solve it i just know how it s supposed to work d steps to reproduce place lava crystal in fuel slot in alloy kiln add in valid alloyable metals watch lava crystal get eaten d client information modpack version java version see image launcher used twitch launcher memory allocated see image server lan single player single player resourcepack enabled no optifine installed no
1
78,357
3,509,587,559
IssuesEvent
2016-01-08 23:38:29
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
opened
Prince Malchezaar (BB #1135)
migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** nort333 **Original Date:** 11.09.2015 19:27:01 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** new **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1135 <hr> When useing his weapon, disarm spells not working on him.
1.0
Prince Malchezaar (BB #1135) - This issue was migrated from bitbucket. **Original Reporter:** nort333 **Original Date:** 11.09.2015 19:27:01 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** new **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1135 <hr> When useing his weapon, disarm spells not working on him.
priority
prince malchezaar bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state new direct link when useing his weapon disarm spells not working on him
1
534,472
15,624,224,300
IssuesEvent
2021-03-21 01:10:14
LujainKhalaf/Soengram
https://api.github.com/repos/LujainKhalaf/Soengram
closed
Add static analysis packages to CI/CD pipeline
1 point med-risk priority-medium task
We want to add various static analysis packages to our CI/CD pipeline. The packages should be added to `requirements.txt` and should be run automatically on each commit. - Then black as the auto formatter https://github.com/psf/black#the-black-code-style - Flake8 as the linter https://flake8.pycqa.org/en/latest/ - Radon for complexity analysis https://pypi.org/project/radon/ - Then pytest for unit and functional tests as well as code coverage https://docs.pytest.org/en/stable/ - mypy for static type checking http://mypy-lang.org/
1.0
Add static analysis packages to CI/CD pipeline - We want to add various static analysis packages to our CI/CD pipeline. The packages should be added to `requirements.txt` and should be run automatically on each commit. - Then black as the auto formatter https://github.com/psf/black#the-black-code-style - Flake8 as the linter https://flake8.pycqa.org/en/latest/ - Radon for complexity analysis https://pypi.org/project/radon/ - Then pytest for unit and functional tests as well as code coverage https://docs.pytest.org/en/stable/ - mypy for static type checking http://mypy-lang.org/
priority
add static analysis packages to ci cd pipeline we want to add various static analysis packages to our ci cd pipeline the packages should be added to requirements txt and should be run automatically on each commit then black as the auto formatter as the linter radon for complexity analysis then pytest for unit and functional tests as well as code coverage mypy for static type checking
1
359,619
10,678,761,500
IssuesEvent
2019-10-21 17:55:30
AY1920S1-CS2113T-T12-3/main
https://api.github.com/repos/AY1920S1-CS2113T-T12-3/main
closed
As a user, I want to be able to save prices and locations of ingredients
priority.Medium type.Story
so that I do not have to input them every time
1.0
As a user, I want to be able to save prices and locations of ingredients - so that I do not have to input them every time
priority
as a user i want to be able to save prices and locations of ingredients so that i do not have to input them every time
1
284,181
8,736,339,430
IssuesEvent
2018-12-11 19:15:54
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
It isn't possible to see the material specific values in a mixed material zone in Label plot
bug crash likelihood medium priority reviewed severity high wrong results
Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 3002 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: It isn't possible to see the material specific values in a mixed material zone in Label plot Assigned to: Alister Maguire Category: Target version: 2.13.1 Author: Eric Brugger Start: 01/29/2018 Due date: % Done: 100 Estimated time: 4.0 Created: 01/29/2018 08:17 pm Updated: 02/26/2018 05:00 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.12.3 Impact: Expected Use: OS: All Support Group: Any Description: Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working. Comments: pdbdatabase.C has a mixed scalar. Thanks for the tip about pdbdatabase.C containing mixed material variables. I tested VisIt 2.12.3 and it did the right thing. I will ask Matt for a data file. Matt gave me his file and when I force interface reconstruction, a Pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece. When I do a label plot of materials, it splits the cells and labels each piece with the correct material label. When I do a label plot of the mixed material variable it doesn't split the cell and displays the whole zone value. So, it is as Matt described, almost everything works correctly, just this one case doesn't. The data file he gave me is in ~brugger1/visit_obrien_bug on kickit. I suspect that the database generated by pdbdatabase.C should show the same behavior. Ok, specmix_ucd.silo has mixed variable values for variables "p" and "d" Hey everyone,Ive committed two revisions to the trunk.Revision 32458:This revision fixes a bug with the label plot not being able to correctly label mixed variables. Ive only been able to test on one dataset, but the issue should be resolved in general.M src/plots/Label/avtLabelFilter.CM src/plots/Label/avtLabelFilter.hRevision 32459:I added in a couple of tests to the label plot test file. These new tests will check to make sure that cells with mixed variables are being labeled correctly.A test/baseline/plots/label/label_11_00.pngA test/baseline/plots/label/label_11_01.pngM test/tests/plots/label.pyBest,Alister
1.0
It isn't possible to see the material specific values in a mixed material zone in Label plot - Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 3002 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: It isn't possible to see the material specific values in a mixed material zone in Label plot Assigned to: Alister Maguire Category: Target version: 2.13.1 Author: Eric Brugger Start: 01/29/2018 Due date: % Done: 100 Estimated time: 4.0 Created: 01/29/2018 08:17 pm Updated: 02/26/2018 05:00 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.12.3 Impact: Expected Use: OS: All Support Group: Any Description: Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working. Comments: pdbdatabase.C has a mixed scalar. Thanks for the tip about pdbdatabase.C containing mixed material variables. I tested VisIt 2.12.3 and it did the right thing. I will ask Matt for a data file. Matt gave me his file and when I force interface reconstruction, a Pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece. When I do a label plot of materials, it splits the cells and labels each piece with the correct material label. When I do a label plot of the mixed material variable it doesn't split the cell and displays the whole zone value. So, it is as Matt described, almost everything works correctly, just this one case doesn't. The data file he gave me is in ~brugger1/visit_obrien_bug on kickit. I suspect that the database generated by pdbdatabase.C should show the same behavior. Ok, specmix_ucd.silo has mixed variable values for variables "p" and "d" Hey everyone,Ive committed two revisions to the trunk.Revision 32458:This revision fixes a bug with the label plot not being able to correctly label mixed variables. Ive only been able to test on one dataset, but the issue should be resolved in general.M src/plots/Label/avtLabelFilter.CM src/plots/Label/avtLabelFilter.hRevision 32459:I added in a couple of tests to the label plot test file. These new tests will check to make sure that cells with mixed variables are being labeled correctly.A test/baseline/plots/label/label_11_00.pngA test/baseline/plots/label/label_11_01.pngM test/tests/plots/label.pyBest,Alister
priority
it isn t possible to see the material specific values in a mixed material zone in label plot matt obrien reported that with code he has mixed materials with material specific values normally visit always displays the whole zone values unless some material selection is being performed if there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values neither of these mechanisms appear to be working i couldn t verify this since none of our test data sets appear to have per material values for variables or at least none that i could find the first step is to modify the silo generator to put out material specific values for the density that are markedly different so that we can test this then we should get this working redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject it isn t possible to see the material specific values in a mixed material zone in label plot assigned to alister maguire category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description matt obrien reported that with code he has mixed materials with material specific values normally visit always displays the whole zone values unless some material selection is being performed if there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values neither of these mechanisms appear to be working i couldn t verify this since none of our test data sets appear to have per material values for variables or at least none that i could find the first step is to modify the silo generator to put out material specific values for the density that are markedly different so that we can test this then we should get this working comments pdbdatabase c has a mixed scalar thanks for the tip about pdbdatabase c containing mixed material variables i tested visit and it did the right thing i will ask matt for a data file matt gave me his file and when i force interface reconstruction a pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece when i do a label plot of materials it splits the cells and labels each piece with the correct material label when i do a label plot of the mixed material variable it doesn t split the cell and displays the whole zone value so it is as matt described almost everything works correctly just this one case doesn t the data file he gave me is in visit obrien bug on kickit i suspect that the database generated by pdbdatabase c should show the same behavior ok specmix ucd silo has mixed variable values for variables p and d hey everyone ive committed two revisions to the trunk revision this revision fixes a bug with the label plot not being able to correctly label mixed variables ive only been able to test on one dataset but the issue should be resolved in general m src plots label avtlabelfilter cm src plots label avtlabelfilter hrevision i added in a couple of tests to the label plot test file these new tests will check to make sure that cells with mixed variables are being labeled correctly a test baseline plots label label pnga test baseline plots label label pngm test tests plots label pybest alister
1
670,443
22,690,059,057
IssuesEvent
2022-07-04 18:45:14
belav/csharpier
https://api.github.com/repos/belav/csharpier
closed
Extra space before curly brace in array initializer
type:bug area:formatting priority:medium
Here is what I expect: ``` public class MyClass { public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 }; } ``` Actual code: ``` public class MyClass { public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 }; } ``` Version: 0.18.0
1.0
Extra space before curly brace in array initializer - Here is what I expect: ``` public class MyClass { public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 }; } ``` Actual code: ``` public class MyClass { public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 }; } ``` Version: 0.18.0
priority
extra space before curly brace in array initializer here is what i expect public class myclass public int arr get set actual code public class myclass public int arr get set version
1
614,813
19,190,249,118
IssuesEvent
2021-12-05 21:43:05
oresat/oresat-dxwifi-software
https://api.github.com/repos/oresat/oresat-dxwifi-software
opened
Add device tree overlay
medium priority
Device tree overlay should project based. So need to move the DxWiFi device tree from `oresat-linux` repo to here and add it to the `oresat-dxwifi-tx` Debian package.
1.0
Add device tree overlay - Device tree overlay should project based. So need to move the DxWiFi device tree from `oresat-linux` repo to here and add it to the `oresat-dxwifi-tx` Debian package.
priority
add device tree overlay device tree overlay should project based so need to move the dxwifi device tree from oresat linux repo to here and add it to the oresat dxwifi tx debian package
1
583,526
17,391,377,207
IssuesEvent
2021-08-02 07:51:57
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
opened
[Розширений пошук] Star rating control is misaligned
Priority: Medium bug
**Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit) **Reproducible:** always **Build found:** last commit **Preconditions** 1. Go to https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Click on 'Розширений пошук' icon 2. Click on 'List view' icon **Actual result** Star control is misaligned to the right side of the page. ![зображення_viber_2021-08-02_10-24-36-314](https://user-images.githubusercontent.com/87877499/127822310-75cc7ba5-b261-4b63-86c9-0aadc542794d.jpg) **Expected result** Star rating control is located below Club's logo and and summary in the left corner. ![expected result1](https://user-images.githubusercontent.com/87877499/127823410-7fa404d8-ca88-416e-8912-cd262c05148f.jpg) **User story and test case links** E.g.: "User story #
1.0
[Розширений пошук] Star rating control is misaligned - **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit) **Reproducible:** always **Build found:** last commit **Preconditions** 1. Go to https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Click on 'Розширений пошук' icon 2. Click on 'List view' icon **Actual result** Star control is misaligned to the right side of the page. ![зображення_viber_2021-08-02_10-24-36-314](https://user-images.githubusercontent.com/87877499/127822310-75cc7ba5-b261-4b63-86c9-0aadc542794d.jpg) **Expected result** Star rating control is located below Club's logo and and summary in the left corner. ![expected result1](https://user-images.githubusercontent.com/87877499/127823410-7fa404d8-ca88-416e-8912-cd262c05148f.jpg) **User story and test case links** E.g.: "User story #
priority
star rating control is misaligned environment windows google chrome version official build bit reproducible always build found last commit preconditions go to steps to reproduce click on розширений пошук icon click on list view icon actual result star control is misaligned to the right side of the page expected result star rating control is located below club s logo and and summary in the left corner user story and test case links e g user story
1
317,241
9,662,405,999
IssuesEvent
2019-05-20 20:44:28
GingerWalnut/SQBeyondPublic
https://api.github.com/repos/GingerWalnut/SQBeyondPublic
closed
Combat Dummy Bug?
medium priority
So I was building a capital ship on Mardos late last night (5-5-19) and I took a shuttle down to the floor of my hangar and logged off for the night (around 9-10pm PST). There were no hostile mobs in sight and the area was well lit. There were also very few players online at that hour. Once I logged back on today at around 4pm PST, I got a greeting saying I had died with all my items and 28 levels because my combat dummy was slain, despite being nowhere near combat. Im not sure if this is a bug or my own fault? Help would be appreciated. If needed I can explain further in game. (IGN: BoyWonderFTW and base referred to is on Mardos under Isolators faction)
1.0
Combat Dummy Bug? - So I was building a capital ship on Mardos late last night (5-5-19) and I took a shuttle down to the floor of my hangar and logged off for the night (around 9-10pm PST). There were no hostile mobs in sight and the area was well lit. There were also very few players online at that hour. Once I logged back on today at around 4pm PST, I got a greeting saying I had died with all my items and 28 levels because my combat dummy was slain, despite being nowhere near combat. Im not sure if this is a bug or my own fault? Help would be appreciated. If needed I can explain further in game. (IGN: BoyWonderFTW and base referred to is on Mardos under Isolators faction)
priority
combat dummy bug so i was building a capital ship on mardos late last night and i took a shuttle down to the floor of my hangar and logged off for the night around pst there were no hostile mobs in sight and the area was well lit there were also very few players online at that hour once i logged back on today at around pst i got a greeting saying i had died with all my items and levels because my combat dummy was slain despite being nowhere near combat im not sure if this is a bug or my own fault help would be appreciated if needed i can explain further in game ign boywonderftw and base referred to is on mardos under isolators faction
1
763,981
26,780,318,066
IssuesEvent
2023-01-31 20:34:35
pystardust/ani-cli
https://api.github.com/repos/pystardust/ani-cli
opened
How do I change providers?
type: bug priority 2: medium
Used to be able to choose what website (Provider) to scrape from with "-r" option, but it is now used for episode selection. how would I be able to scrape a different provider without editing the script?
1.0
How do I change providers? - Used to be able to choose what website (Provider) to scrape from with "-r" option, but it is now used for episode selection. how would I be able to scrape a different provider without editing the script?
priority
how do i change providers used to be able to choose what website provider to scrape from with r option but it is now used for episode selection how would i be able to scrape a different provider without editing the script
1
657,664
21,799,779,597
IssuesEvent
2022-05-16 02:54:00
Wiredcraft/pipelines
https://api.github.com/repos/Wiredcraft/pipelines
closed
Add icons/update UI title
Priority: Medium Type: Enhancement
I got a whole bunch of `pipelines` setup - but they are all the same look & feel. Can we; - specify a custom icon? - tune the title? See example; 5 completely different setups on misc domains - very hard to distinguish. <img width="1232" alt="screen shot 2018-03-23 at 09 26 06" src="https://user-images.githubusercontent.com/631984/37806748-935e3378-2e7c-11e8-8eb2-02d1276929d9.png">
1.0
Add icons/update UI title - I got a whole bunch of `pipelines` setup - but they are all the same look & feel. Can we; - specify a custom icon? - tune the title? See example; 5 completely different setups on misc domains - very hard to distinguish. <img width="1232" alt="screen shot 2018-03-23 at 09 26 06" src="https://user-images.githubusercontent.com/631984/37806748-935e3378-2e7c-11e8-8eb2-02d1276929d9.png">
priority
add icons update ui title i got a whole bunch of pipelines setup but they are all the same look feel can we specify a custom icon tune the title see example completely different setups on misc domains very hard to distinguish img width alt screen shot at src
1
133,290
5,200,305,700
IssuesEvent
2017-01-23 23:25:04
mapbox/mapbox-gl-js
https://api.github.com/repos/mapbox/mapbox-gl-js
closed
Issue an error when map canvas is larger than allowed by gl.MAX_RENDERBUFFER_SIZE
medium priority starter task
Once the map canvas is larger than `gl.MAX_RENDERBUFFER_SIZE / 2`, scaling behaviour starts to occur. I do not think we should invest significant resources into accommodating this case. See https://github.com/mapbox/mapbox-gl-js/issues/2673 and https://github.com/mapbox/mapbox-gl-js/issues/2863
1.0
Issue an error when map canvas is larger than allowed by gl.MAX_RENDERBUFFER_SIZE - Once the map canvas is larger than `gl.MAX_RENDERBUFFER_SIZE / 2`, scaling behaviour starts to occur. I do not think we should invest significant resources into accommodating this case. See https://github.com/mapbox/mapbox-gl-js/issues/2673 and https://github.com/mapbox/mapbox-gl-js/issues/2863
priority
issue an error when map canvas is larger than allowed by gl max renderbuffer size once the map canvas is larger than gl max renderbuffer size scaling behaviour starts to occur i do not think we should invest significant resources into accommodating this case see and
1
25,799
2,683,992,682
IssuesEvent
2015-03-28 15:06:43
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
closed
-new_console глючит
2–5 stars bug imported Priority-Medium
_From [sim....@gmail.com](https://code.google.com/u/105258257765487351754/) on January 28, 2011 03:16:27_ При запуске фара с ключем -new_console он запускается в странном укороченном по ширине виде, и при этом еще страшно тормозит клавиатура. **Attachment:** [Буфер обмена-1.jpg](http://code.google.com/p/conemu-maximus5/issues/detail?id=362) _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=362_
1.0
-new_console глючит - _From [sim....@gmail.com](https://code.google.com/u/105258257765487351754/) on January 28, 2011 03:16:27_ При запуске фара с ключем -new_console он запускается в странном укороченном по ширине виде, и при этом еще страшно тормозит клавиатура. **Attachment:** [Буфер обмена-1.jpg](http://code.google.com/p/conemu-maximus5/issues/detail?id=362) _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=362_
priority
new console глючит from on january при запуске фара с ключем new console он запускается в странном укороченном по ширине виде и при этом еще страшно тормозит клавиатура attachment original issue
1
65,672
3,237,350,334
IssuesEvent
2015-10-14 11:23:04
awesome-raccoons/gqt
https://api.github.com/repos/awesome-raccoons/gqt
opened
Weird things happen for high zoom levels
bug medium priority
To reproduce: Plot POLYGON((0 0, 10 0, 10 10, 0 0)) and press Ctr-+ more than 34 times (the real fun is around 38 presses). I guess this is caused by floating point errors, so it's probably hard to fix it. A good enough fix is to just stop zooming before this becomes an issue.
1.0
Weird things happen for high zoom levels - To reproduce: Plot POLYGON((0 0, 10 0, 10 10, 0 0)) and press Ctr-+ more than 34 times (the real fun is around 38 presses). I guess this is caused by floating point errors, so it's probably hard to fix it. A good enough fix is to just stop zooming before this becomes an issue.
priority
weird things happen for high zoom levels to reproduce plot polygon and press ctr more than times the real fun is around presses i guess this is caused by floating point errors so it s probably hard to fix it a good enough fix is to just stop zooming before this becomes an issue
1
336,105
10,171,474,751
IssuesEvent
2019-08-08 08:29:05
pmem/issues
https://api.github.com/repos/pmem/issues
closed
Tests fail because of can't a read alignment of devdax
Exposure: Medium OS: Linux Priority: 2 high State: To be verified Type: Bug Type: External Bug
<!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> ## Environment Information - PMDK package version(s): 1.6-237-g3ca1cd956 - OS(es) version(s): Fedora 29 - ndctl version(s): 65 - kernel version(s): 5.1.12 <!-- fill in also other useful environment data --> ## Please provide a reproduction of the bug: A lot of PMDK tests fail for the same reason. Below I put an example: ``` $ ./RUNTESTS obj_basic_integration/ -s TEST7 obj_basic_integration/TEST7: SETUP (check/pmem/debug) obj_basic_integration/TEST7 crashed (signal 6). err7.log below. err7.log below. out7.log below. obj_basic_integration/TEST7 out7.log obj_basic_integration/TEST7: START: obj_basic_integration obj_basic_integration/TEST7 out7.log ./obj_basic_integration /dev/dax1.5 pmem7.log below. obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:236 out_init] pid 28580: program: /home/jenkins/greg/pmdk/src/test/obj_basic_integration/obj_basic_integration obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:56 libpmem_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:784 pmem_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:419 pmem_init_funcs] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:376 pmem_cpuinfo_to_funcs] clflushopt supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:388 pmem_cpuinfo_to_funcs] clwb supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:325 use_avx512f_memcpy_memset] avx512f supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:329 use_avx512f_memcpy_memset] PMEM_AVX512F not set or not == 1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:483 pmem_init_funcs] using clwb obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:69 libpmem_fini] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:101 util_mmap_fini] pmemblk7.log below. obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1 obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:130 libpmemblk_fini] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:101 util_mmap_fini] pmemlog7.log below. obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1 obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:130 libpmemlog_fini] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:101 util_mmap_fini] Last 30 lines of pmemobj7.log below (whole file has 42 lines). obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil) obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:301 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x8e8ed4 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x7f5d47644900 name fallocate.at_create indexes 0x7ffc08eac318 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:124 util_remote_init] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1328 pmemobj_createU] path /dev/dax1.5 layout basic poolsize 0 mode 600 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1298 obj_get_nlanes] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3355 util_pool_create] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 attr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3115 util_pool_create_uuids] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 pattr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 remote 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:2124 util_poolset_create_set] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:559 util_file_open] path "/dev/dax1.5" size 0x7ffc08eac118 minsize 0 flags 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:67 device_dax_size] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:1703 util_poolset_single] path /dev/dax1.5 filesize 6339690496 create 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:291 util_file_device_dax_alignment] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:212 device_dax_alignment] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [file_posix.c:230 device_dax_alignment] open "/sys/dev/char/252:11/device/align": No such file or directory obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [set.c:1752 util_poolset_single] assertion failure: rep->part[0].alignment (0x0) != 0 (0x0) obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:325 obj_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:139 util_remote_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [mmap.c:101 util_mmap_fini] pmempool7.log below. obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:238 out_init] libpmempool version 1.3 obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:69 libpmempool_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:124 util_remote_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:85 libpmempool_fini] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:139 util_remote_fini] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:101 util_mmap_fini] rpmem7.log below. obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.3 obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:61 librpmem_init] obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:77 librpmem_fini] trace7.log below. obj_basic_integration/TEST7 trace7.log {obj_basic_integration.c:653 main} obj_basic_integration/TEST7: START: obj_basic_integration obj_basic_integration/TEST7 trace7.log ./obj_basic_integration /dev/dax1.5 RUNTESTS: stopping: obj_basic_integration//TEST7 failed, TEST=check FS=pmem BUILD=debug ``` ## How often bug is revealed: (always, often, rare): always ## Actual behavior: As above. ## Expected behavior: Tests should pass. ## Details <!-- fill this out --> ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
1.0
Tests fail because of can't a read alignment of devdax - <!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> ## Environment Information - PMDK package version(s): 1.6-237-g3ca1cd956 - OS(es) version(s): Fedora 29 - ndctl version(s): 65 - kernel version(s): 5.1.12 <!-- fill in also other useful environment data --> ## Please provide a reproduction of the bug: A lot of PMDK tests fail for the same reason. Below I put an example: ``` $ ./RUNTESTS obj_basic_integration/ -s TEST7 obj_basic_integration/TEST7: SETUP (check/pmem/debug) obj_basic_integration/TEST7 crashed (signal 6). err7.log below. err7.log below. out7.log below. obj_basic_integration/TEST7 out7.log obj_basic_integration/TEST7: START: obj_basic_integration obj_basic_integration/TEST7 out7.log ./obj_basic_integration /dev/dax1.5 pmem7.log below. obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:236 out_init] pid 28580: program: /home/jenkins/greg/pmdk/src/test/obj_basic_integration/obj_basic_integration obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:56 libpmem_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:784 pmem_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:419 pmem_init_funcs] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:376 pmem_cpuinfo_to_funcs] clflushopt supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:388 pmem_cpuinfo_to_funcs] clwb supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:325 use_avx512f_memcpy_memset] avx512f supported obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:329 use_avx512f_memcpy_memset] PMEM_AVX512F not set or not == 1 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:483 pmem_init_funcs] using clwb obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2 obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:69 libpmem_fini] obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:101 util_mmap_fini] pmemblk7.log below. obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1 obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:130 libpmemblk_fini] obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:101 util_mmap_fini] pmemlog7.log below. obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1 obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:130 libpmemlog_fini] obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:101 util_mmap_fini] Last 30 lines of pmemobj7.log below (whole file has 42 lines). obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil) obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:301 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x8e8ed4 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x7f5d47644900 name fallocate.at_create indexes 0x7ffc08eac318 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:124 util_remote_init] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1328 pmemobj_createU] path /dev/dax1.5 layout basic poolsize 0 mode 600 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1298 obj_get_nlanes] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3355 util_pool_create] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 attr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3115 util_pool_create_uuids] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 pattr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 remote 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:2124 util_poolset_create_set] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:559 util_file_open] path "/dev/dax1.5" size 0x7ffc08eac118 minsize 0 flags 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:67 device_dax_size] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:1703 util_poolset_single] path /dev/dax1.5 filesize 6339690496 create 0 obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:291 util_file_device_dax_alignment] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:212 device_dax_alignment] path "/dev/dax1.5" obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [file_posix.c:230 device_dax_alignment] open "/sys/dev/char/252:11/device/align": No such file or directory obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [set.c:1752 util_poolset_single] assertion failure: rep->part[0].alignment (0x0) != 0 (0x0) obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:325 obj_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:139 util_remote_fini] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [mmap.c:101 util_mmap_fini] pmempool7.log below. obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:238 out_init] libpmempool version 1.3 obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:67 util_mmap_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:69 libpmempool_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:124 util_remote_init] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:85 libpmempool_fini] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:139 util_remote_fini] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload] obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:101 util_mmap_fini] rpmem7.log below. obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.3 obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956 obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:61 librpmem_init] obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:77 librpmem_fini] trace7.log below. obj_basic_integration/TEST7 trace7.log {obj_basic_integration.c:653 main} obj_basic_integration/TEST7: START: obj_basic_integration obj_basic_integration/TEST7 trace7.log ./obj_basic_integration /dev/dax1.5 RUNTESTS: stopping: obj_basic_integration//TEST7 failed, TEST=check FS=pmem BUILD=debug ``` ## How often bug is revealed: (always, often, rare): always ## Actual behavior: As above. ## Expected behavior: Tests should pass. ## Details <!-- fill this out --> ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
priority
tests fail because of can t a read alignment of devdax before creating new issue ensure that similar issue wasn t already created search note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report remember this is just a minimal template you can extend it with data you think may be useful environment information pmdk package version s os es version s fedora ndctl version s kernel version s please provide a reproduction of the bug a lot of pmdk tests fail for the same reason below i put an example runtests obj basic integration s obj basic integration setup check pmem debug obj basic integration crashed signal log below log below log below obj basic integration log obj basic integration start obj basic integration obj basic integration log obj basic integration dev log below obj basic integration log pid program home jenkins greg pmdk src test obj basic integration obj basic integration obj basic integration log libpmem version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log clflush supported obj basic integration log clflushopt supported obj basic integration log clwb supported obj basic integration log avx supported obj basic integration log pmem avx not set or not obj basic integration log supported obj basic integration log pmem not set or not obj basic integration log obj basic integration log region path sys bus nd devices obj basic integration log domain path sys bus nd devices persistence domain obj basic integration log flushing cpu cache obj basic integration log using clwb obj basic integration log using movnt obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmemblk version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmemlog version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log last lines of log below whole file has lines obj basic integration log pop nil obj basic integration log ctl nil ctx nil cfg string fallocate at create obj basic integration log ctl nil ctx nil source name fallocate at create type arg obj basic integration log nodes name fallocate at create indexes obj basic integration log obj basic integration log path dev layout basic poolsize mode obj basic integration log obj basic integration log setp path dev poolsize minsize minpartsize attr nlanes can have rep obj basic integration log setp path dev poolsize minsize minpartsize pattr nlanes can have rep remote obj basic integration log path dev obj basic integration log setp path dev poolsize minsize obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev size minsize flags obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev filesize create obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log open sys dev char device align no such file or directory obj basic integration log assertion failure rep part alignment obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmempool version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log librpmem version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log log below obj basic integration log obj basic integration c main obj basic integration start obj basic integration obj basic integration log obj basic integration dev runtests stopping obj basic integration failed test check fs pmem build debug how often bug is revealed always often rare always actual behavior as above expected behavior tests should pass details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low
1
374,403
11,089,301,542
IssuesEvent
2019-12-14 17:31:30
bounswe/bounswe2019group4
https://api.github.com/repos/bounswe/bounswe2019group4
closed
Backend feature recommendation
Back-End Priority: Medium Type: Development
According to our requirements, system shall recommend new users, articles etc to logged in users according user's actions. Endpoint of recommendations will be /recommendations , and method type will be GET.
1.0
Backend feature recommendation - According to our requirements, system shall recommend new users, articles etc to logged in users according user's actions. Endpoint of recommendations will be /recommendations , and method type will be GET.
priority
backend feature recommendation according to our requirements system shall recommend new users articles etc to logged in users according user s actions endpoint of recommendations will be recommendations and method type will be get
1
744,347
25,939,566,999
IssuesEvent
2022-12-16 17:01:38
asastats/channel
https://api.github.com/repos/asastats/channel
closed
[B2] Wrong NFT purchase prices
bug medium priority addressed
We're going to have a bigger analysys and fixup. Here are the example problems: - - - - - https://discord.com/channels/906917846754418770/908054330265960478/961212380904296459 Wrong last purchase price for FORUM Astro #\2 NFT: last purchase for FORUM Astro #\2 showing 25 ALGO, according to NFTexplorer = 55 ALGO on AlgoGems. Previous sale = 200 ALGO on ALGOxNFT 535631028 - - - - - https://github.com/asastats/channel/issues/154#issuecomment-1171855303 The time looks good for these two different ones. Although the amount for the second shows 1 ALGO when in fact it was sold for 7. ![image](https://user-images.githubusercontent.com/97535612/176810648-3305ef44-c053-4490-b7ee-486a8f67aa6b.png) Here's the group txn for that sale: https://algoexplorer.io/tx/group/%2F0aPItSoO2OM%2F9gf7CUk6qQsfpMnQlviXwJxczlJlh8%3D - - - - -
1.0
[B2] Wrong NFT purchase prices - We're going to have a bigger analysys and fixup. Here are the example problems: - - - - - https://discord.com/channels/906917846754418770/908054330265960478/961212380904296459 Wrong last purchase price for FORUM Astro #\2 NFT: last purchase for FORUM Astro #\2 showing 25 ALGO, according to NFTexplorer = 55 ALGO on AlgoGems. Previous sale = 200 ALGO on ALGOxNFT 535631028 - - - - - https://github.com/asastats/channel/issues/154#issuecomment-1171855303 The time looks good for these two different ones. Although the amount for the second shows 1 ALGO when in fact it was sold for 7. ![image](https://user-images.githubusercontent.com/97535612/176810648-3305ef44-c053-4490-b7ee-486a8f67aa6b.png) Here's the group txn for that sale: https://algoexplorer.io/tx/group/%2F0aPItSoO2OM%2F9gf7CUk6qQsfpMnQlviXwJxczlJlh8%3D - - - - -
priority
wrong nft purchase prices we re going to have a bigger analysys and fixup here are the example problems wrong last purchase price for forum astro nft last purchase for forum astro showing algo according to nftexplorer algo on algogems previous sale algo on algoxnft the time looks good for these two different ones although the amount for the second shows algo when in fact it was sold for here s the group txn for that sale
1
622,854
19,657,893,066
IssuesEvent
2022-01-10 14:20:51
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
Delete & Select all icons shows up when there is no video
feature-enhancement priority-medium feature-video feature Stale
### Scope When even there is no video uploaded by the users it still displays the delete & select all icon on the top under profile->Video or anywhere we display the video tab. See the Screenshot: https://nimbusweb.me/nimbus-screenshots/5873fd55b7f1d88d819f65dc0fe36755 ### Acceptance Criteria Delete & select all icons should not show when there is no video. Support Ticket: https://secure.helpscout.net/conversation/1618101603/161750/ **Jira issue** : [PROD-750] [PROD-750]: https://buddyboss.atlassian.net/browse/PROD-750?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
1.0
Delete & Select all icons shows up when there is no video - ### Scope When even there is no video uploaded by the users it still displays the delete & select all icon on the top under profile->Video or anywhere we display the video tab. See the Screenshot: https://nimbusweb.me/nimbus-screenshots/5873fd55b7f1d88d819f65dc0fe36755 ### Acceptance Criteria Delete & select all icons should not show when there is no video. Support Ticket: https://secure.helpscout.net/conversation/1618101603/161750/ **Jira issue** : [PROD-750] [PROD-750]: https://buddyboss.atlassian.net/browse/PROD-750?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
priority
delete select all icons shows up when there is no video scope when even there is no video uploaded by the users it still displays the delete select all icon on the top under profile video or anywhere we display the video tab see the screenshot acceptance criteria delete select all icons should not show when there is no video support ticket jira issue
1
36,246
2,797,417,779
IssuesEvent
2015-05-12 13:44:42
twogee/ant-http
https://api.github.com/repos/twogee/ant-http
closed
[CLOSED] Setup build process with local properties file based on username
auto-migrated Milestone-1.0.1 Priority-Medium Project-ant-http Type-Task
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday May 11, 2015 at 22:05 GMT_ _Originally opened as https://github.com/twogee/missing-link/issues/3_ ---- ``` Setup build process with local properties file based on username * Include a sample properties file with sensitive data not included * Build project into a single jar * Build source jar * Build javadocs * Build api (javadoc) zip * Build distribution folder with binary, source and javadocs * Build distribution zip ``` Original issue reported on code.google.com by `alex.she...@gmail.com` on 13 Mar 2011 at 6:02
1.0
[CLOSED] Setup build process with local properties file based on username - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday May 11, 2015 at 22:05 GMT_ _Originally opened as https://github.com/twogee/missing-link/issues/3_ ---- ``` Setup build process with local properties file based on username * Include a sample properties file with sensitive data not included * Build project into a single jar * Build source jar * Build javadocs * Build api (javadoc) zip * Build distribution folder with binary, source and javadocs * Build distribution zip ``` Original issue reported on code.google.com by `alex.she...@gmail.com` on 13 Mar 2011 at 6:02
priority
setup build process with local properties file based on username issue by monday may at gmt originally opened as setup build process with local properties file based on username include a sample properties file with sensitive data not included build project into a single jar build source jar build javadocs build api javadoc zip build distribution folder with binary source and javadocs build distribution zip original issue reported on code google com by alex she gmail com on mar at
1
823,350
30,991,955,238
IssuesEvent
2023-08-09 05:46:21
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
opened
[YSQL] Update query which uses UNNEST function that results in Hash join takes 20X time for the same query without UNNEST
area/ysql priority/medium status/awaiting-triage
### Description To update a single row of a table(schema details below) using UNNEST it is taking close to ~85ms where as when same query is executed without UNNEST, it results an update with a simple index scan and completes in ~5ms. **The issue is observed when UNNEST results in hash join. UNNEST with nested loop seems to be taking same time when executed without UNNEST.** This issue was resulting in high latency observed with HammerDB Stored Procedures. For details have a look at this document https://docs.google.com/document/d/1n1kfmNi2FeixmmkIhceSw89-qZsdKFj5rACl9vOQy8k/edit?usp=sharing **Update query using UNNEST resulting in hash join** ``` yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 FROM UNNEST(ARRAY[1], ARRAY[1]) AS ids(o_id, d_id) WHERE col_bigint_1=ids.o_id AND col_bigint_2=ids.d_id; QUERY PLAN ------------------------------------------------------------------------------------------------------------------------------------------------------ Update on pkey_rangescan_fulltable_10column_1 (cost=2.50..110.03 rows=2 width=248) (actual time=83.385..83.385 rows=0 loops=1) -> Hash Join (cost=2.50..110.03 rows=2 width=248) (actual time=7.749..79.554 rows=1 loops=1) Hash Cond: ((pkey_rangescan_fulltable_10column_1.col_bigint_1 = ids.o_id) AND (pkey_rangescan_fulltable_10column_1.col_bigint_2 = ids.d_id)) -> Seq Scan on pkey_rangescan_fulltable_10column_1 (cost=0.00..100.00 rows=1000 width=216) (actual time=7.714..78.906 rows=10000 loops=1) -> Hash (cost=1.00..1.00 rows=100 width=40) (actual time=0.023..0.023 rows=1 loops=1) Buckets: 1024 Batches: 1 Memory Usage: 9kB -> Function Scan on ids (cost=0.01..1.00 rows=100 width=40) (actual time=0.017..0.017 rows=1 loops=1) Planning Time: 0.128 ms Execution Time: 83.454 ms Peak Memory Usage: 272 kB (10 rows) Time: 84.511 ms ``` **Update query without UNNEST** ``` yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 WHERE col_bigint_1=1 AND col_bigint_2=1; QUERY PLAN ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ -------- Update on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.192..1.192 rows=0 loops=1) -> Index Scan using pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.174..1.175 rows=1 l oops=1) Index Cond: ((col_bigint_1 = 1) AND (col_bigint_2 = 1)) Planning Time: 0.076 ms Execution Time: 3.109 ms Peak Memory Usage: 24 kB (6 rows) Time: 4.634 ms ``` **Table schema** ``` yugabyte=# \d+ pkey_rangescan_fulltable_10column_1 Table "public.pkey_rangescan_fulltable_10column_1" Column | Type | Collation | Nullable | Default | Storage | Stats target | Description -----------------+--------+-----------+----------+---------+---------+--------------+------------- col_bigint_id_1 | bigint | | not null | | plain | | col_bigint_1 | bigint | | | | plain | | col_bigint_2 | bigint | | | | plain | | col_bigint_3 | bigint | | | | plain | | col_bigint_4 | bigint | | | | plain | | col_bigint_5 | bigint | | | | plain | | col_bigint_6 | bigint | | | | plain | | col_bigint_7 | bigint | | | | plain | | col_bigint_8 | bigint | | | | plain | | col_bigint_9 | bigint | | | | plain | | Indexes: "pkey_rangescan_fulltable_10column_1_pkey" PRIMARY KEY, lsm (col_bigint_id_1 ASC) "pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx" lsm (col_bigint_1 HASH, col_bigint_2 ASC) ``` ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information.
1.0
[YSQL] Update query which uses UNNEST function that results in Hash join takes 20X time for the same query without UNNEST - ### Description To update a single row of a table(schema details below) using UNNEST it is taking close to ~85ms where as when same query is executed without UNNEST, it results an update with a simple index scan and completes in ~5ms. **The issue is observed when UNNEST results in hash join. UNNEST with nested loop seems to be taking same time when executed without UNNEST.** This issue was resulting in high latency observed with HammerDB Stored Procedures. For details have a look at this document https://docs.google.com/document/d/1n1kfmNi2FeixmmkIhceSw89-qZsdKFj5rACl9vOQy8k/edit?usp=sharing **Update query using UNNEST resulting in hash join** ``` yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 FROM UNNEST(ARRAY[1], ARRAY[1]) AS ids(o_id, d_id) WHERE col_bigint_1=ids.o_id AND col_bigint_2=ids.d_id; QUERY PLAN ------------------------------------------------------------------------------------------------------------------------------------------------------ Update on pkey_rangescan_fulltable_10column_1 (cost=2.50..110.03 rows=2 width=248) (actual time=83.385..83.385 rows=0 loops=1) -> Hash Join (cost=2.50..110.03 rows=2 width=248) (actual time=7.749..79.554 rows=1 loops=1) Hash Cond: ((pkey_rangescan_fulltable_10column_1.col_bigint_1 = ids.o_id) AND (pkey_rangescan_fulltable_10column_1.col_bigint_2 = ids.d_id)) -> Seq Scan on pkey_rangescan_fulltable_10column_1 (cost=0.00..100.00 rows=1000 width=216) (actual time=7.714..78.906 rows=10000 loops=1) -> Hash (cost=1.00..1.00 rows=100 width=40) (actual time=0.023..0.023 rows=1 loops=1) Buckets: 1024 Batches: 1 Memory Usage: 9kB -> Function Scan on ids (cost=0.01..1.00 rows=100 width=40) (actual time=0.017..0.017 rows=1 loops=1) Planning Time: 0.128 ms Execution Time: 83.454 ms Peak Memory Usage: 272 kB (10 rows) Time: 84.511 ms ``` **Update query without UNNEST** ``` yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 WHERE col_bigint_1=1 AND col_bigint_2=1; QUERY PLAN ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ -------- Update on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.192..1.192 rows=0 loops=1) -> Index Scan using pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.174..1.175 rows=1 l oops=1) Index Cond: ((col_bigint_1 = 1) AND (col_bigint_2 = 1)) Planning Time: 0.076 ms Execution Time: 3.109 ms Peak Memory Usage: 24 kB (6 rows) Time: 4.634 ms ``` **Table schema** ``` yugabyte=# \d+ pkey_rangescan_fulltable_10column_1 Table "public.pkey_rangescan_fulltable_10column_1" Column | Type | Collation | Nullable | Default | Storage | Stats target | Description -----------------+--------+-----------+----------+---------+---------+--------------+------------- col_bigint_id_1 | bigint | | not null | | plain | | col_bigint_1 | bigint | | | | plain | | col_bigint_2 | bigint | | | | plain | | col_bigint_3 | bigint | | | | plain | | col_bigint_4 | bigint | | | | plain | | col_bigint_5 | bigint | | | | plain | | col_bigint_6 | bigint | | | | plain | | col_bigint_7 | bigint | | | | plain | | col_bigint_8 | bigint | | | | plain | | col_bigint_9 | bigint | | | | plain | | Indexes: "pkey_rangescan_fulltable_10column_1_pkey" PRIMARY KEY, lsm (col_bigint_id_1 ASC) "pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx" lsm (col_bigint_1 HASH, col_bigint_2 ASC) ``` ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information.
priority
update query which uses unnest function that results in hash join takes time for the same query without unnest description to update a single row of a table schema details below using unnest it is taking close to where as when same query is executed without unnest it results an update with a simple index scan and completes in the issue is observed when unnest results in hash join unnest with nested loop seems to be taking same time when executed without unnest this issue was resulting in high latency observed with hammerdb stored procedures for details have a look at this document update query using unnest resulting in hash join yugabyte explain analyse update pkey rangescan fulltable set col bigint col bigint from unnest array array as ids o id d id where col bigint ids o id and col bigint ids d id query plan update on pkey rangescan fulltable cost rows width actual time rows loops hash join cost rows width actual time rows loops hash cond pkey rangescan fulltable col bigint ids o id and pkey rangescan fulltable col bigint ids d id seq scan on pkey rangescan fulltable cost rows width actual time rows loops hash cost rows width actual time rows loops buckets batches memory usage function scan on ids cost rows width actual time rows loops planning time ms execution time ms peak memory usage kb rows time ms update query without unnest yugabyte explain analyse update pkey rangescan fulltable set col bigint col bigint where col bigint and col bigint query plan update on pkey rangescan fulltable cost rows width actual time rows loops index scan using pkey rangescan fulltable col bigint col bigint idx on pkey rangescan fulltable cost rows width actual time rows l oops index cond col bigint and col bigint planning time ms execution time ms peak memory usage kb rows time ms table schema yugabyte d pkey rangescan fulltable table public pkey rangescan fulltable column type collation nullable default storage stats target description col bigint id bigint not null plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain indexes pkey rangescan fulltable pkey primary key lsm col bigint id asc pkey rangescan fulltable col bigint col bigint idx lsm col bigint hash col bigint asc warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information
1
20,580
2,622,853,814
IssuesEvent
2015-03-04 08:06:40
max99x/pagemon-chrome-ext
https://api.github.com/repos/max99x/pagemon-chrome-ext
closed
user needs to modify the monitored URL
auto-migrated Priority-Medium
``` What steps will reproduce the problem? Please include a URL. Some URLs will be changed automatically and so it can't be monitored correctly. The link is: http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district What is the expected output? What do you see instead? The link is turned to be: http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district/district 1_10.html What version of the Chrome are you using? On what operating system? 30.0.1599.101 Win7 64 ``` Original issue reported on code.google.com by `wong....@gmail.com` on 25 Oct 2013 at 10:27 * Merged into: #97
1.0
user needs to modify the monitored URL - ``` What steps will reproduce the problem? Please include a URL. Some URLs will be changed automatically and so it can't be monitored correctly. The link is: http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district What is the expected output? What do you see instead? The link is turned to be: http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district/district 1_10.html What version of the Chrome are you using? On what operating system? 30.0.1599.101 Win7 64 ``` Original issue reported on code.google.com by `wong....@gmail.com` on 25 Oct 2013 at 10:27 * Merged into: #97
priority
user needs to modify the monitored url what steps will reproduce the problem please include a url some urls will be changed automatically and so it can t be monitored correctly the link is what is the expected output what do you see instead the link is turned to be html what version of the chrome are you using on what operating system original issue reported on code google com by wong gmail com on oct at merged into
1
632,846
20,236,681,772
IssuesEvent
2022-02-14 03:20:48
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
You can't pet the cat/etc
Priority: 1-Urgent Type: Feature Difficulty: 2-Medium
## Description If you try to interact with say a caracal and probably other animals, you do not pet them <!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.--> Not sure if this should be only visible to the one petting (and getting pet, I guess), or if it should also be global, though.
1.0
You can't pet the cat/etc - ## Description If you try to interact with say a caracal and probably other animals, you do not pet them <!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.--> Not sure if this should be only visible to the one petting (and getting pet, I guess), or if it should also be global, though.
priority
you can t pet the cat etc description if you try to interact with say a caracal and probably other animals you do not pet them not sure if this should be only visible to the one petting and getting pet i guess or if it should also be global though
1
427,509
12,396,155,945
IssuesEvent
2020-05-20 19:59:59
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
closed
Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User
Priority: Medium Product: caseflow-queue Team: Echo 🐬 Type: Bug
## Description The Assigned To field should display the correct user ID on the search cases page ## Acceptance criteria - [ ] The assigned to field on the search cases page displays the correct user assigned to the case ## Background/context/resources During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots ## Technical notes Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search. ## Screenshots ![image](https://user-images.githubusercontent.com/59883609/81325124-ecf81780-9065-11ea-98bc-f3b27c3bcd0e.png)
1.0
Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User - ## Description The Assigned To field should display the correct user ID on the search cases page ## Acceptance criteria - [ ] The assigned to field on the search cases page displays the correct user assigned to the case ## Background/context/resources During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots ## Technical notes Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search. ## Screenshots ![image](https://user-images.githubusercontent.com/59883609/81325124-ecf81780-9065-11ea-98bc-f3b27c3bcd0e.png)
priority
case movement non blocking tasks assigned to field not displaying correct user description the assigned to field should display the correct user id on the search cases page acceptance criteria the assigned to field on the search cases page displays the correct user assigned to the case background context resources during dogfooding assigning a case to an judge legacy reassigning to burnham assigned case to instead of css when searching the case again check the for screenshots technical notes reproduce before digging into hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs this would leave assigned to unchanged in case search screenshots
1
129,740
5,102,586,693
IssuesEvent
2017-01-04 18:47:12
TauCetiStation/TauCetiClassic
https://api.github.com/repos/TauCetiStation/TauCetiClassic
closed
Показывает плохое состояние здоровья при 100хп
bug priority: medium
Originally reported by: **Anonymous** --- Дело было вечером. Админов не было. Находился в районе получаса со сломанной ногой в плохом и крите(в дальнейшем состоянии). Привезли на станцию,все вроде бы залатали, но теперь первое - никого не слышно,хотя иногда фразы проскакивают второе - не могу встать с каталки третье -показывает плохое состояние здоровья при 100 хп четвертое - меня никто не может поднять Ps: не гоуст Pps: под конец уже начали маяться херней,так как онлайн был 10 человек,половина из которых сидела и говорила в LOOC --- - Bitbucket: https://bitbucket.org/taucetistation/tauceti/issue/154
1.0
Показывает плохое состояние здоровья при 100хп - Originally reported by: **Anonymous** --- Дело было вечером. Админов не было. Находился в районе получаса со сломанной ногой в плохом и крите(в дальнейшем состоянии). Привезли на станцию,все вроде бы залатали, но теперь первое - никого не слышно,хотя иногда фразы проскакивают второе - не могу встать с каталки третье -показывает плохое состояние здоровья при 100 хп четвертое - меня никто не может поднять Ps: не гоуст Pps: под конец уже начали маяться херней,так как онлайн был 10 человек,половина из которых сидела и говорила в LOOC --- - Bitbucket: https://bitbucket.org/taucetistation/tauceti/issue/154
priority
показывает плохое состояние здоровья при originally reported by anonymous дело было вечером админов не было находился в районе получаса со сломанной ногой в плохом и крите в дальнейшем состоянии привезли на станцию все вроде бы залатали но теперь первое никого не слышно хотя иногда фразы проскакивают второе не могу встать с каталки третье показывает плохое состояние здоровья при хп четвертое меня никто не может поднять ps не гоуст pps под конец уже начали маяться херней так как онлайн был человек половина из которых сидела и говорила в looc bitbucket
1
196,324
6,926,823,712
IssuesEvent
2017-11-30 20:32:48
AZMAG/map-ATP
https://api.github.com/repos/AZMAG/map-ATP
closed
Remove the date field completely from the main form
Priority: Medium
Removing the validation and just send in the current date.
1.0
Remove the date field completely from the main form - Removing the validation and just send in the current date.
priority
remove the date field completely from the main form removing the validation and just send in the current date
1
548,161
16,058,980,059
IssuesEvent
2021-04-23 09:43:43
input-output-hk/cardano-node
https://api.github.com/repos/input-output-hk/cardano-node
closed
Add instructions to README on how to run node with testnet and mainnet
byron priority medium
As was fairly [pointed out on Reddit](https://www.reddit.com/r/cardano/comments/f879vw/docker_image_for_cardanonode_160_is_here/): there no instructions on > ... how to connect the current code-base to either Byron mainnet or testnet, do you have some docs on that or is it going to be more of a trial and error type of thing?
1.0
Add instructions to README on how to run node with testnet and mainnet - As was fairly [pointed out on Reddit](https://www.reddit.com/r/cardano/comments/f879vw/docker_image_for_cardanonode_160_is_here/): there no instructions on > ... how to connect the current code-base to either Byron mainnet or testnet, do you have some docs on that or is it going to be more of a trial and error type of thing?
priority
add instructions to readme on how to run node with testnet and mainnet as was fairly there no instructions on how to connect the current code base to either byron mainnet or testnet do you have some docs on that or is it going to be more of a trial and error type of thing
1
657,841
21,869,624,559
IssuesEvent
2022-05-19 03:12:27
pixley/TimelineBuilder
https://api.github.com/repos/pixley/TimelineBuilder
opened
Parallel calendars
type: feature status: to do priority: medium
Allow for dynamic switching between calendar system views on one timeline. For example, a user could switch between Absalom Reckoning and the Imperial Calendar for their Pathfinder timeline, and the dates should change accordingly.
1.0
Parallel calendars - Allow for dynamic switching between calendar system views on one timeline. For example, a user could switch between Absalom Reckoning and the Imperial Calendar for their Pathfinder timeline, and the dates should change accordingly.
priority
parallel calendars allow for dynamic switching between calendar system views on one timeline for example a user could switch between absalom reckoning and the imperial calendar for their pathfinder timeline and the dates should change accordingly
1
47,266
2,974,605,205
IssuesEvent
2015-07-15 02:18:13
Reimashi/jotai
https://api.github.com/repos/Reimashi/jotai
closed
Laptop battery monitoring support
auto-migrated Priority-Medium Type-Enhancement
``` What steps will reproduce the problem? 1. Downloaded 2. Started and checked all options available in View/Sensors menu 3. I can not see the fan speed and many others as presented on your web site. Please see the attached file What is the expected output? What do you see instead? I would like to see the meters for all the hardware, as it is presented on the website What version of the product are you using? On what operating system? I am using the latest version on a Windows 7 OS Please provide any additional information below. ``` Original issue reported on code.google.com by `linasel` on 23 Feb 2010 at 10:12 Attachments: * [Hardware monitor.jpg](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-4/comment-0/Hardware monitor.jpg)
1.0
Laptop battery monitoring support - ``` What steps will reproduce the problem? 1. Downloaded 2. Started and checked all options available in View/Sensors menu 3. I can not see the fan speed and many others as presented on your web site. Please see the attached file What is the expected output? What do you see instead? I would like to see the meters for all the hardware, as it is presented on the website What version of the product are you using? On what operating system? I am using the latest version on a Windows 7 OS Please provide any additional information below. ``` Original issue reported on code.google.com by `linasel` on 23 Feb 2010 at 10:12 Attachments: * [Hardware monitor.jpg](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-4/comment-0/Hardware monitor.jpg)
priority
laptop battery monitoring support what steps will reproduce the problem downloaded started and checked all options available in view sensors menu i can not see the fan speed and many others as presented on your web site please see the attached file what is the expected output what do you see instead i would like to see the meters for all the hardware as it is presented on the website what version of the product are you using on what operating system i am using the latest version on a windows os please provide any additional information below original issue reported on code google com by linasel on feb at attachments monitor jpg
1
645,903
21,032,355,684
IssuesEvent
2022-03-31 02:44:24
hackforla/tdm-calculator
https://api.github.com/repos/hackforla/tdm-calculator
reopened
Add icons for Accordion having external links
role: front-end level: medium priority: MUST HAVE p-Feature - Tool Tips
### Overview For any text having links to external websites, we should be adding an icon that indicates that. ### Action Items - [ ] For external links add the icon that shows its going to an external website (https://designsystem.digital.gov/components/link/ ) ![image](https://user-images.githubusercontent.com/20568596/153102023-5dc4e0d1-2961-4aa1-a73b-4b794790eace.png) - [ ] Add the image for the external links on the AIN/APN row on page 1 - [ ] Add the image for the external links on the Bike Share on page 2 - [ ] Add the image for the external links on the About Us page ### Resources/Instructions REPLACE THIS TEXT -If there is a website which has documentation that helps with this issue provide the link(s) here.
1.0
Add icons for Accordion having external links - ### Overview For any text having links to external websites, we should be adding an icon that indicates that. ### Action Items - [ ] For external links add the icon that shows its going to an external website (https://designsystem.digital.gov/components/link/ ) ![image](https://user-images.githubusercontent.com/20568596/153102023-5dc4e0d1-2961-4aa1-a73b-4b794790eace.png) - [ ] Add the image for the external links on the AIN/APN row on page 1 - [ ] Add the image for the external links on the Bike Share on page 2 - [ ] Add the image for the external links on the About Us page ### Resources/Instructions REPLACE THIS TEXT -If there is a website which has documentation that helps with this issue provide the link(s) here.
priority
add icons for accordion having external links overview for any text having links to external websites we should be adding an icon that indicates that action items for external links add the icon that shows its going to an external website add the image for the external links on the ain apn row on page add the image for the external links on the bike share on page add the image for the external links on the about us page resources instructions replace this text if there is a website which has documentation that helps with this issue provide the link s here
1
673,666
23,026,459,415
IssuesEvent
2022-07-22 09:40:02
apache/incubator-devlake
https://api.github.com/repos/apache/incubator-devlake
closed
Extract file-level data from git repository
type/feature-request priority/medium need-follow-up
## User story As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution. ## Goal The key is to match `bug` to `component`. - If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`; - If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`; ## Proposed Solution - [ ] Add a `commit_files` table in the domain layer | field | type | length | description | key | | ------------- | -------- | -------- | ----------- | -----| | `commit_sha` | char | 40 | | FK | | `file_name` | varchar | 255 | | | | `status` | varchar | 255 | | | | `additions` | int | | | | | `deletions` | int | | | | | `changes` | int | | | | | `file_component` | varchar | 255 | | | a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.** b. how to assign `file_component` value to a commit file? It varies in different programming languages. - In Go, a component can be found by folder names; - while in other languages, it might need user input to define the `file_component`. - [ ] Allow users to define the `file_component` in `.env` by REGEX ## An example of commit files For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files: ![image](https://user-images.githubusercontent.com/14050754/171377048-59230b96-be40-4168-93d4-775bf9b33a7a.png) ## Belongs to issue This is a sub-issue of both #949 and #1771 ## Additional context N/A
1.0
Extract file-level data from git repository - ## User story As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution. ## Goal The key is to match `bug` to `component`. - If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`; - If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`; ## Proposed Solution - [ ] Add a `commit_files` table in the domain layer | field | type | length | description | key | | ------------- | -------- | -------- | ----------- | -----| | `commit_sha` | char | 40 | | FK | | `file_name` | varchar | 255 | | | | `status` | varchar | 255 | | | | `additions` | int | | | | | `deletions` | int | | | | | `changes` | int | | | | | `file_component` | varchar | 255 | | | a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.** b. how to assign `file_component` value to a commit file? It varies in different programming languages. - In Go, a component can be found by folder names; - while in other languages, it might need user input to define the `file_component`. - [ ] Allow users to define the `file_component` in `.env` by REGEX ## An example of commit files For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files: ![image](https://user-images.githubusercontent.com/14050754/171377048-59230b96-be40-4168-93d4-775bf9b33a7a.png) ## Belongs to issue This is a sub-issue of both #949 and #1771 ## Additional context N/A
priority
extract file level data from git repository user story as a tech lead i want to see the bug distribution in low level components from code files as a guide for the future framework evolution goal the key is to match bug to component if we want to find the overall distribution of all bugs we can do issues bug join prs join commits join commit files component if we want to find the overall distribution of bugs fixed in a specific tag we can do ref diff issues bug join prs join commits join commit files component proposed solution add a commit files table in the domain layer field type length description key commit sha char fk file name varchar status varchar additions int deletions int changes int file component varchar a all the columns except file component can be retrieved from note not from or gitlab api b how to assign file component value to a commit file it varies in different programming languages in go a component can be found by folder names while in other languages it might need user input to define the file component allow users to define the file component in env by regex an example of commit files for instance this touches files belongs to issue this is a sub issue of both and additional context n a
1
744,418
25,942,244,804
IssuesEvent
2022-12-16 19:41:03
zowe/zowe-cli
https://api.github.com/repos/zowe/zowe-cli
closed
Command "zowe profiles create" prints the password that should be secret
enhancement for-review priority-medium community-upvoted
I am using Zowe CLI in a CI pipeline. As a part of the pipeline, I call following command: ``` zowe profiles create zosmf-profile zowe \ --host $ZOS_HOST --port $ZOS_ZOSMF_PORT \ --user "$USERID" --pass "$PASSWORD" ``` In the log of the job is the following output: ``` Profile created successfully! Path: /var/lib/jenkins_agent/workspace/Verify Zowe Prereqs/profiles/zosmf/zowe.yaml host: myhost port: 443 user: MYUSER password: MYPASSWORD rejectUnauthorized: true protocol: https ``` I am not happy that my password is printed and all other users can see it.
1.0
Command "zowe profiles create" prints the password that should be secret - I am using Zowe CLI in a CI pipeline. As a part of the pipeline, I call following command: ``` zowe profiles create zosmf-profile zowe \ --host $ZOS_HOST --port $ZOS_ZOSMF_PORT \ --user "$USERID" --pass "$PASSWORD" ``` In the log of the job is the following output: ``` Profile created successfully! Path: /var/lib/jenkins_agent/workspace/Verify Zowe Prereqs/profiles/zosmf/zowe.yaml host: myhost port: 443 user: MYUSER password: MYPASSWORD rejectUnauthorized: true protocol: https ``` I am not happy that my password is printed and all other users can see it.
priority
command zowe profiles create prints the password that should be secret i am using zowe cli in a ci pipeline as a part of the pipeline i call following command zowe profiles create zosmf profile zowe host zos host port zos zosmf port user userid pass password in the log of the job is the following output profile created successfully path var lib jenkins agent workspace verify zowe prereqs profiles zosmf zowe yaml host myhost port user myuser password mypassword rejectunauthorized true protocol https i am not happy that my password is printed and all other users can see it
1
56,835
3,081,206,364
IssuesEvent
2015-08-22 13:47:43
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Create stats database schema if none is present on server
015a bug imported Priority-Medium
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on February 10, 2011 16:06:23_ When a server needs a sqlite stats database, if one is not present, we need to create a new one, tables and all. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=57_
1.0
Create stats database schema if none is present on server - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on February 10, 2011 16:06:23_ When a server needs a sqlite stats database, if one is not present, we need to create a new one, tables and all. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=57_
priority
create stats database schema if none is present on server from on february when a server needs a sqlite stats database if one is not present we need to create a new one tables and all original issue
1
237,759
7,763,897,616
IssuesEvent
2018-06-01 18:14:30
ansible/awx
https://api.github.com/repos/ansible/awx
closed
[RFE] [nl] Job Status graph could use 'dd-mm-yyyy' date format (currently uses 'mm/dd')
component:ui priority:medium state:needs_info type:enhancement
##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: --> - UI ##### SUMMARY <!-- Briefly describe the problem. --> In Dutch (Netherlands) google-chrome, Job Status graph could use 'dd-mm-yyyy' date format, as per [1] and [2] ##### ENVIRONMENT * AWX version: 1.1.0 * AWX install method: docker on linux * Ansible version: 2.3.1.0 * Operating System: Fedora 25 * Web Browser: google-chrome ### Steps To Reproduce: <!-- For bugs, please show exactly how to reproduce the problem. For new features, show how the feature would be used. --> 1. In Dutch google-chrome, do a successful login and login into Tower 2. Post login, observe 'date' format on X axis (Time) of Job Status graph ### Expected Results: <!-- For bug reports, what did you expect to happen when running the steps above? --> dd-mm-yyyy ### Actual Results: <!-- For bug reports, what actually happened? --> mm/dd ### Additional Information: <!-- Include any links to sosreport, database dumps, screenshots or other information. --> [1] https://en.wikipedia.org/wiki/Date_format_by_country [2] https://en.wikipedia.org/wiki/Date_and_time_notation_in_the_Netherlands [3] Screenshot of the Job Status graph currently showing date in 'mm/dd' format on X axis (Time): ![30736608-0d0ec7b6-9fa1-11e7-8a89-67ebcd8eff17](https://user-images.githubusercontent.com/12151785/36148398-e2838092-10e1-11e8-9764-0c50087cc275.png)
1.0
[RFE] [nl] Job Status graph could use 'dd-mm-yyyy' date format (currently uses 'mm/dd') - ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: --> - UI ##### SUMMARY <!-- Briefly describe the problem. --> In Dutch (Netherlands) google-chrome, Job Status graph could use 'dd-mm-yyyy' date format, as per [1] and [2] ##### ENVIRONMENT * AWX version: 1.1.0 * AWX install method: docker on linux * Ansible version: 2.3.1.0 * Operating System: Fedora 25 * Web Browser: google-chrome ### Steps To Reproduce: <!-- For bugs, please show exactly how to reproduce the problem. For new features, show how the feature would be used. --> 1. In Dutch google-chrome, do a successful login and login into Tower 2. Post login, observe 'date' format on X axis (Time) of Job Status graph ### Expected Results: <!-- For bug reports, what did you expect to happen when running the steps above? --> dd-mm-yyyy ### Actual Results: <!-- For bug reports, what actually happened? --> mm/dd ### Additional Information: <!-- Include any links to sosreport, database dumps, screenshots or other information. --> [1] https://en.wikipedia.org/wiki/Date_format_by_country [2] https://en.wikipedia.org/wiki/Date_and_time_notation_in_the_Netherlands [3] Screenshot of the Job Status graph currently showing date in 'mm/dd' format on X axis (Time): ![30736608-0d0ec7b6-9fa1-11e7-8a89-67ebcd8eff17](https://user-images.githubusercontent.com/12151785/36148398-e2838092-10e1-11e8-9764-0c50087cc275.png)
priority
job status graph could use dd mm yyyy date format currently uses mm dd issue type feature idea component name ui summary in dutch netherlands google chrome job status graph could use dd mm yyyy date format as per and environment awx version awx install method docker on linux ansible version operating system fedora web browser google chrome steps to reproduce for bugs please show exactly how to reproduce the problem for new features show how the feature would be used in dutch google chrome do a successful login and login into tower post login observe date format on x axis time of job status graph expected results for bug reports what did you expect to happen when running the steps above dd mm yyyy actual results mm dd additional information include any links to sosreport database dumps screenshots or other information screenshot of the job status graph currently showing date in mm dd format on x axis time
1
495,632
14,285,435,322
IssuesEvent
2020-11-23 13:55:09
geocollections/sarv-edit
https://api.github.com/repos/geocollections/sarv-edit
closed
Drillcore box alternative list views
Difficulty: Medium Priority: Low Source: App Status: Available Type: Feature Request
Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared). A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):   ![PNG image](https://user-images.githubusercontent.com/14974887/91879338-a8356380-ec88-11ea-8c78-695d7bfded77.png)
1.0
Drillcore box alternative list views - Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared). A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):   ![PNG image](https://user-images.githubusercontent.com/14974887/91879338-a8356380-ec88-11ea-8c78-695d7bfded77.png)
priority
drillcore box alternative list views drillcore box filter list view should have separate view with box images one for each box if no image placeholder should be shown similar to file list view where images are shown within table when clicking on box image large image should be shown in separate popup window that will remain open so that two or more box images can be compared a different view is needed for printing box images into pdf two landscape images per portrait sheet  
1
236,049
7,745,634,737
IssuesEvent
2018-05-29 18:58:35
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Have to press E twice to re-open a world object ui you just closed
Medium Priority
This seems to be an error in PlayerHeldObject line ~255, something about the previous interaction object being the same.
1.0
Have to press E twice to re-open a world object ui you just closed - This seems to be an error in PlayerHeldObject line ~255, something about the previous interaction object being the same.
priority
have to press e twice to re open a world object ui you just closed this seems to be an error in playerheldobject line something about the previous interaction object being the same
1
26,545
2,684,862,783
IssuesEvent
2015-03-29 13:08:53
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
opened
Crashes on binary data.
1 star bug imported Priority-Medium
_From [gynvael](https://code.google.com/u/gynvael/) on May 12, 2013 03:29:56_ Required information! OS version: Win7 SP1 x64 ConEmu version: 130427 [64] Far version (if you are using Far Manager): n/a *Bug description* To check the stability of ConEmu I've played a little with outputting binary data and invalid terminal control sequences. In some cases some crashes do appear - usually the crash is in the console application (and not ConEmu itself), but it's due to the ConEmuHk hook. Sometimes ConEmu itself crashes (though tbh I have not managed to reproduce it on the newest ConEmu version; this problem was mostly in older versions). (I've attached the test app - ansifuzz.cpp) Additionally, sometimes a simple: int main(){for(;;)putchar(rand());} freezez ConEmu after pressing ctrl-c to break the output. Note: I have enabled the following In-console options (I think they are relevant): * Inject ConEmuHk * ANSI X3.64 / xterm 256 colors *Steps to reproduction* 1. Download the attached file and run: type last_test.bin 2. It should go for a while and then step. Press ctrl+c - it will output some more text and crash. I didn't analyze the bugs further. **Attachment:** [ansifuzz.cpp last_test.bin](http://code.google.com/p/conemu-maximus5/issues/detail?id=1064) _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1064_
1.0
Crashes on binary data. - _From [gynvael](https://code.google.com/u/gynvael/) on May 12, 2013 03:29:56_ Required information! OS version: Win7 SP1 x64 ConEmu version: 130427 [64] Far version (if you are using Far Manager): n/a *Bug description* To check the stability of ConEmu I've played a little with outputting binary data and invalid terminal control sequences. In some cases some crashes do appear - usually the crash is in the console application (and not ConEmu itself), but it's due to the ConEmuHk hook. Sometimes ConEmu itself crashes (though tbh I have not managed to reproduce it on the newest ConEmu version; this problem was mostly in older versions). (I've attached the test app - ansifuzz.cpp) Additionally, sometimes a simple: int main(){for(;;)putchar(rand());} freezez ConEmu after pressing ctrl-c to break the output. Note: I have enabled the following In-console options (I think they are relevant): * Inject ConEmuHk * ANSI X3.64 / xterm 256 colors *Steps to reproduction* 1. Download the attached file and run: type last_test.bin 2. It should go for a while and then step. Press ctrl+c - it will output some more text and crash. I didn't analyze the bugs further. **Attachment:** [ansifuzz.cpp last_test.bin](http://code.google.com/p/conemu-maximus5/issues/detail?id=1064) _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1064_
priority
crashes on binary data from on may required information os version conemu version far version if you are using far manager n a bug description to check the stability of conemu i ve played a little with outputting binary data and invalid terminal control sequences in some cases some crashes do appear usually the crash is in the console application and not conemu itself but it s due to the conemuhk hook sometimes conemu itself crashes though tbh i have not managed to reproduce it on the newest conemu version this problem was mostly in older versions i ve attached the test app ansifuzz cpp additionally sometimes a simple int main for putchar rand freezez conemu after pressing ctrl c to break the output note i have enabled the following in console options i think they are relevant inject conemuhk ansi xterm colors steps to reproduction download the attached file and run type last test bin it should go for a while and then step press ctrl c it will output some more text and crash i didn t analyze the bugs further attachment original issue
1
817,561
30,644,288,791
IssuesEvent
2023-07-25 02:22:59
diba-io/bitmask-core
https://api.github.com/repos/diba-io/bitmask-core
closed
Return Genesis Timestamp in RGB Struct
priority-medium
I'm pretty sure RGB embeds a timestamp in contract genesis, or at least, it used to. We need that.
1.0
Return Genesis Timestamp in RGB Struct - I'm pretty sure RGB embeds a timestamp in contract genesis, or at least, it used to. We need that.
priority
return genesis timestamp in rgb struct i m pretty sure rgb embeds a timestamp in contract genesis or at least it used to we need that
1
585,894
17,537,624,676
IssuesEvent
2021-08-12 08:21:39
sjeong12/My-Editor
https://api.github.com/repos/sjeong12/My-Editor
closed
개발도구 조사 및 선정
doc priority: medium
## 개요 다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정 - 유저 인터랙션에 대한 반응속도 100ms 미만을 목표로 - SPA로 구성 - 다양한 상태관리 필요 ## 작업내용 - [x] 요구사항에 맞는 개발 도구 조사 - [x] 각 개발도구 성능 비교 및 선정
1.0
개발도구 조사 및 선정 - ## 개요 다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정 - 유저 인터랙션에 대한 반응속도 100ms 미만을 목표로 - SPA로 구성 - 다양한 상태관리 필요 ## 작업내용 - [x] 요구사항에 맞는 개발 도구 조사 - [x] 각 개발도구 성능 비교 및 선정
priority
개발도구 조사 및 선정 개요 다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정 유저 인터랙션에 대한 반응속도 미만을 목표로 spa로 구성 다양한 상태관리 필요 작업내용 요구사항에 맞는 개발 도구 조사 각 개발도구 성능 비교 및 선정
1
206,046
7,108,221,723
IssuesEvent
2018-01-16 22:59:19
qlicker/qlicker
https://api.github.com/repos/qlicker/qlicker
closed
course/:id/results route for students
Medium priority enhancement
This should show the same view as for an instructor, except with only the row for the current student. It means modifying the ClassParticipationList element to subscribe to the correct data. Currently, **have set the route for a student to redirect to their login route**, but should update this Presumably, this is originally meant to be handled by the route /results/:studentId/:courseId, but it currently doesn't work (assume that StudentResultPage needs to be updated. **I've commented this route out**, which is called in sessionClickHandler of student/courses.jsx - I don't think that an overview of the course results is what we want anyway, when clicking on a session item.
1.0
course/:id/results route for students - This should show the same view as for an instructor, except with only the row for the current student. It means modifying the ClassParticipationList element to subscribe to the correct data. Currently, **have set the route for a student to redirect to their login route**, but should update this Presumably, this is originally meant to be handled by the route /results/:studentId/:courseId, but it currently doesn't work (assume that StudentResultPage needs to be updated. **I've commented this route out**, which is called in sessionClickHandler of student/courses.jsx - I don't think that an overview of the course results is what we want anyway, when clicking on a session item.
priority
course id results route for students this should show the same view as for an instructor except with only the row for the current student it means modifying the classparticipationlist element to subscribe to the correct data currently have set the route for a student to redirect to their login route but should update this presumably this is originally meant to be handled by the route results studentid courseid but it currently doesn t work assume that studentresultpage needs to be updated i ve commented this route out which is called in sessionclickhandler of student courses jsx i don t think that an overview of the course results is what we want anyway when clicking on a session item
1
105,001
4,227,399,840
IssuesEvent
2016-07-03 05:43:57
OperationCode/operationcode
https://api.github.com/repos/OperationCode/operationcode
closed
[Code Schools] Add The Guild of Software Architects
Priority: Medium Status: In Progress Type: Maintenance
/codeschools The Guild of Software Architects 6170 Research Rd, Suite 107 Frisco,TX 75033
1.0
[Code Schools] Add The Guild of Software Architects - /codeschools The Guild of Software Architects 6170 Research Rd, Suite 107 Frisco,TX 75033
priority
add the guild of software architects codeschools the guild of software architects research rd suite frisco tx
1
296,721
9,125,640,227
IssuesEvent
2019-02-24 15:25:47
pixijs/pixi.js
https://api.github.com/repos/pixijs/pixi.js
closed
Double RenderTexture destroy
Plugin: cacheAsBitmap Priority: Medium Resolution: Won't Fix Type: Bug
This demo hacks the destroyTexture method and adds counter to it. If you place a breakpoint in it, you see that every renderTexture is destroyed two times: one time when the BaseTexture is destroyed and TextureManager disposes the texture, second time when the FrameBuffer is destroyed. https://exponenta.github.io/gstatsjs/exampless/pixi/
1.0
Double RenderTexture destroy - This demo hacks the destroyTexture method and adds counter to it. If you place a breakpoint in it, you see that every renderTexture is destroyed two times: one time when the BaseTexture is destroyed and TextureManager disposes the texture, second time when the FrameBuffer is destroyed. https://exponenta.github.io/gstatsjs/exampless/pixi/
priority
double rendertexture destroy this demo hacks the destroytexture method and adds counter to it if you place a breakpoint in it you see that every rendertexture is destroyed two times one time when the basetexture is destroyed and texturemanager disposes the texture second time when the framebuffer is destroyed
1
7,677
2,603,742,773
IssuesEvent
2015-02-24 17:41:39
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
proposal on sneaky units
auto-migrated Priority-Medium Type-Enhancement
``` You may have noticed, when there is a cloacked enemy unit, it's not quite invisible, but there is a slight background distortion in place. People easily localize cloacked units when they are moving, but of course can not destroy them until tecnically "detected". I suggest a new state Unit::isSneaking() which is always true when the cloacked unit is moving and turns false as soon as the unit stops or uncloakes. When isSneaking()==true you can access the unit's type and all movement/order related properties like position, but cannot access any of the other values like player, and ofcourse, trying to attack it would result in an error. Some callback will help greatly, like onUnitSneaking. When the unit stops onUnitCloak will be called again. ``` ----- Original issue reported on code.google.com by `goo...@teabix.com` on 18 Nov 2009 at 2:17
1.0
proposal on sneaky units - ``` You may have noticed, when there is a cloacked enemy unit, it's not quite invisible, but there is a slight background distortion in place. People easily localize cloacked units when they are moving, but of course can not destroy them until tecnically "detected". I suggest a new state Unit::isSneaking() which is always true when the cloacked unit is moving and turns false as soon as the unit stops or uncloakes. When isSneaking()==true you can access the unit's type and all movement/order related properties like position, but cannot access any of the other values like player, and ofcourse, trying to attack it would result in an error. Some callback will help greatly, like onUnitSneaking. When the unit stops onUnitCloak will be called again. ``` ----- Original issue reported on code.google.com by `goo...@teabix.com` on 18 Nov 2009 at 2:17
priority
proposal on sneaky units you may have noticed when there is a cloacked enemy unit it s not quite invisible but there is a slight background distortion in place people easily localize cloacked units when they are moving but of course can not destroy them until tecnically detected i suggest a new state unit issneaking which is always true when the cloacked unit is moving and turns false as soon as the unit stops or uncloakes when issneaking true you can access the unit s type and all movement order related properties like position but cannot access any of the other values like player and ofcourse trying to attack it would result in an error some callback will help greatly like onunitsneaking when the unit stops onunitcloak will be called again original issue reported on code google com by goo teabix com on nov at
1
676,068
23,115,327,470
IssuesEvent
2022-07-27 16:09:05
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
refactor platform.findPr to return the newest PRs first
priority-3-medium type:refactor status:ready
### Describe the proposed change(s). refactor according to the discussion https://github.com/renovatebot/renovate/discussions/16803
1.0
refactor platform.findPr to return the newest PRs first - ### Describe the proposed change(s). refactor according to the discussion https://github.com/renovatebot/renovate/discussions/16803
priority
refactor platform findpr to return the newest prs first describe the proposed change s refactor according to the discussion
1
246,314
7,894,564,684
IssuesEvent
2018-06-28 22:00:12
enforcer574/smashclub
https://api.github.com/repos/enforcer574/smashclub
opened
Smash Ultimate UI Refresh
Complexity: Medium Priority: 3 - Medium Type: User Request
Considerations for UI update: - Change header fonts to something more "exciting" - Replace character images with SmashUltimate images
1.0
Smash Ultimate UI Refresh - Considerations for UI update: - Change header fonts to something more "exciting" - Replace character images with SmashUltimate images
priority
smash ultimate ui refresh considerations for ui update change header fonts to something more exciting replace character images with smashultimate images
1
639,636
20,760,691,595
IssuesEvent
2022-03-15 15:54:53
AY2122S2-CS2103T-T12-4/tp
https://api.github.com/repos/AY2122S2-CS2103T-T12-4/tp
opened
Show tasks by date
priority.Medium
As a user, I can see my tasks grouped by date so that I can better plan my schedule.
1.0
Show tasks by date - As a user, I can see my tasks grouped by date so that I can better plan my schedule.
priority
show tasks by date as a user i can see my tasks grouped by date so that i can better plan my schedule
1
217,642
7,325,571,507
IssuesEvent
2018-03-03 10:43:14
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
opened
pfstats pointer and connection errors seen in logs
Priority: Medium Type: Bug
> Mar 3 05:42:15 packetfence pfstats: panic: runtime error: invalid memory address or nil pointer dereference > Mar 3 05:42:15 packetfence pfstats: [signal SIGSEGV: segmentation violation code=0x1 addr=0x30 pc=0x61f597] > Mar 3 05:42:15 packetfence pfstats: goroutine 34 [running]: > Mar 3 05:42:15 packetfence pfstats: github.com/inverse-inc/packetfence/go/log.LoggerWContext(0x0, 0x0, 0x0, 0x0, 0x0, 0xc4200a4140, 0xc42020f93b) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/log/log.go:128 +0x37 > Mar 3 05:42:15 packetfence pfstats: main.gauge.Send(0xc42020f900, 0x43, 0x696b00, 0x8fb040) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:53 +0x103 > Mar 3 05:42:15 packetfence pfstats: main.forward(0x79d0e0, 0xc42024a000) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:218 +0x5c2 > Mar 3 05:42:15 packetfence pfstats: created by main.main > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:356 +0x789 > Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=eror msg="write udp 127.0.0.1:51946->127.0.0.1:8125: write: connection refused" pid=21267 > Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=info msg="Starting stats server" pid=21267 >
1.0
pfstats pointer and connection errors seen in logs - > Mar 3 05:42:15 packetfence pfstats: panic: runtime error: invalid memory address or nil pointer dereference > Mar 3 05:42:15 packetfence pfstats: [signal SIGSEGV: segmentation violation code=0x1 addr=0x30 pc=0x61f597] > Mar 3 05:42:15 packetfence pfstats: goroutine 34 [running]: > Mar 3 05:42:15 packetfence pfstats: github.com/inverse-inc/packetfence/go/log.LoggerWContext(0x0, 0x0, 0x0, 0x0, 0x0, 0xc4200a4140, 0xc42020f93b) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/log/log.go:128 +0x37 > Mar 3 05:42:15 packetfence pfstats: main.gauge.Send(0xc42020f900, 0x43, 0x696b00, 0x8fb040) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:53 +0x103 > Mar 3 05:42:15 packetfence pfstats: main.forward(0x79d0e0, 0xc42024a000) > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:218 +0x5c2 > Mar 3 05:42:15 packetfence pfstats: created by main.main > Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:356 +0x789 > Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=eror msg="write udp 127.0.0.1:51946->127.0.0.1:8125: write: connection refused" pid=21267 > Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=info msg="Starting stats server" pid=21267 >
priority
pfstats pointer and connection errors seen in logs mar packetfence pfstats panic runtime error invalid memory address or nil pointer dereference mar packetfence pfstats mar packetfence pfstats goroutine mar packetfence pfstats github com inverse inc packetfence go log loggerwcontext mar packetfence pfstats root gospace src github com inverse inc packetfence go log log go mar packetfence pfstats main gauge send mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence pfstats main forward mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence pfstats created by main main mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence usr local pf bin pfstats t lvl eror msg write udp write connection refused pid mar packetfence usr local pf bin pfstats t lvl info msg starting stats server pid
1
552,075
16,194,264,574
IssuesEvent
2021-05-04 12:49:35
EvanQuan/Chubberino
https://api.github.com/repos/EvanQuan/Chubberino
opened
Refund heisters on quitting mid-heists
enhancement high priority medium effort
Currently if the program ends in the middle of any heists, all heisters lose all cheese they wagered. Should return all cheese on quitting, preferrable with a message.
1.0
Refund heisters on quitting mid-heists - Currently if the program ends in the middle of any heists, all heisters lose all cheese they wagered. Should return all cheese on quitting, preferrable with a message.
priority
refund heisters on quitting mid heists currently if the program ends in the middle of any heists all heisters lose all cheese they wagered should return all cheese on quitting preferrable with a message
1
388,670
11,490,579,134
IssuesEvent
2020-02-11 17:19:55
radical-cybertools/radical.pilot
https://api.github.com/repos/radical-cybertools/radical.pilot
closed
Netiface requirement for 0.10.4
layer:rct priority:medium topic:deployment type:bug
When installing RP from pip on MacOS 10.14 I see the following: ``` radical-pilot 0.50.22 has requirement netifaces==0.10.4, but you'll have netifaces 0.10.9 which is incompatible. ``` Local tests with EnTK perform as expected but I am not sure whether this would be a problem with more "realistic" runs. Is requirement for `==0.10.4` a hard requirement?
1.0
Netiface requirement for 0.10.4 - When installing RP from pip on MacOS 10.14 I see the following: ``` radical-pilot 0.50.22 has requirement netifaces==0.10.4, but you'll have netifaces 0.10.9 which is incompatible. ``` Local tests with EnTK perform as expected but I am not sure whether this would be a problem with more "realistic" runs. Is requirement for `==0.10.4` a hard requirement?
priority
netiface requirement for when installing rp from pip on macos i see the following radical pilot has requirement netifaces but you ll have netifaces which is incompatible local tests with entk perform as expected but i am not sure whether this would be a problem with more realistic runs is requirement for a hard requirement
1
604,225
18,679,852,017
IssuesEvent
2021-11-01 03:08:56
AY2122S1-CS2113T-T12-3/tp
https://api.github.com/repos/AY2122S1-CS2113T-T12-3/tp
closed
[PE-D] When the excel files has some problems, the app just wipes all the excel entries.
type.Task priority.Medium severity.Medium
It is better to ask the user to backup the file or make a copy (eg. StonksXD_Entries.csv.temp) of that excel entry instead of just wiping the corrupted file. ![Untitled1.png](https://raw.githubusercontent.com/YEOWEIHNGWHYELAB/ped/main/files/8ec50e56-87bd-4c9a-9de4-c8851a3b972e.png) ![Untitled.png](https://raw.githubusercontent.com/YEOWEIHNGWHYELAB/ped/main/files/4d4a43e8-6e68-4cb1-bba5-f09505cecb8c.png) <!--session: 1635497124781-460394b5-e63f-406f-b23c-56ce5c43e5d1--> <!--Version: Web v3.4.1--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: YEOWEIHNGWHYELAB/ped#3
1.0
[PE-D] When the excel files has some problems, the app just wipes all the excel entries. - It is better to ask the user to backup the file or make a copy (eg. StonksXD_Entries.csv.temp) of that excel entry instead of just wiping the corrupted file. ![Untitled1.png](https://raw.githubusercontent.com/YEOWEIHNGWHYELAB/ped/main/files/8ec50e56-87bd-4c9a-9de4-c8851a3b972e.png) ![Untitled.png](https://raw.githubusercontent.com/YEOWEIHNGWHYELAB/ped/main/files/4d4a43e8-6e68-4cb1-bba5-f09505cecb8c.png) <!--session: 1635497124781-460394b5-e63f-406f-b23c-56ce5c43e5d1--> <!--Version: Web v3.4.1--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: YEOWEIHNGWHYELAB/ped#3
priority
when the excel files has some problems the app just wipes all the excel entries it is better to ask the user to backup the file or make a copy eg stonksxd entries csv temp of that excel entry instead of just wiping the corrupted file labels type functionalitybug severity medium original yeoweihngwhyelab ped
1
362,200
10,724,196,528
IssuesEvent
2019-10-28 00:08:45
qlcchain/go-qlc
https://api.github.com/repos/qlcchain/go-qlc
closed
badger may lock if gqlc exit abnormally
Priority: Medium Stale Status: Review Needed Type: Bug
### Description of the issue badger may lock if gqlc exit abnormally ### Issue-Type - [x] bug report - [ ] feature request - [ ] Documentation improvement ### Checklist - [ ] Running latest version of code. - [ ] This issue has not been reported earlier. ### Your environment * OS * Go version * Release tag/commit of the code ### Expected behavior What should happen? ### Actual behavior What is actually happening? ### Steps to reproduce 1. Provide detailed steps 2. how to reproduce this issue ### Any extra info ( for eg. code snippet to reproduce, logs, etc. ) If necessary, provide some extra information like code-snippets or error-logs.
1.0
badger may lock if gqlc exit abnormally - ### Description of the issue badger may lock if gqlc exit abnormally ### Issue-Type - [x] bug report - [ ] feature request - [ ] Documentation improvement ### Checklist - [ ] Running latest version of code. - [ ] This issue has not been reported earlier. ### Your environment * OS * Go version * Release tag/commit of the code ### Expected behavior What should happen? ### Actual behavior What is actually happening? ### Steps to reproduce 1. Provide detailed steps 2. how to reproduce this issue ### Any extra info ( for eg. code snippet to reproduce, logs, etc. ) If necessary, provide some extra information like code-snippets or error-logs.
priority
badger may lock if gqlc exit abnormally description of the issue badger may lock if gqlc exit abnormally issue type bug report feature request documentation improvement checklist running latest version of code this issue has not been reported earlier your environment os go version release tag commit of the code expected behavior what should happen actual behavior what is actually happening steps to reproduce provide detailed steps how to reproduce this issue any extra info for eg code snippet to reproduce logs etc if necessary provide some extra information like code snippets or error logs
1
690,676
23,669,124,892
IssuesEvent
2022-08-27 04:12:38
ecotiya/ecotiya-portfolio-site
https://api.github.com/repos/ecotiya/ecotiya-portfolio-site
opened
Github Actionsを利用して、AWSにDeployできる仕組みの構築。
enhancement Priority: Medium
<!-- 要望のテンプレート --> ## 概要 Github Actionsを利用して、AWSにDeployできる仕組みの構築。 ECRをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。 Terraformでインフラをコード化する。 ## 目的 基本的にGithub Actionsを利用して、AWSにDeployできる仕組みの構築ができれば良い。 その際にはテストコードを走らせ、自動的にAWSにデプロイしてほしい。勉強用。 ## タスク - [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を踏まえて、疑問・調査事項の解決を実施する。 - [ ] 何をすべきかをタスク化。 ## 疑問・調査事項 - [ ] Github Actionsとは何をするのかを調査。 - [ ] Terraformとは何かを調査。現行は[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)を使用してコード化しているため、必要ないのではないか。 - [ ] letest管理からの脱却。 ## 補足 ・一般的に、TerraformやCloudFormationのソースコードってgithubに上げるべきなのか確認。
1.0
Github Actionsを利用して、AWSにDeployできる仕組みの構築。 - <!-- 要望のテンプレート --> ## 概要 Github Actionsを利用して、AWSにDeployできる仕組みの構築。 ECRをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。 Terraformでインフラをコード化する。 ## 目的 基本的にGithub Actionsを利用して、AWSにDeployできる仕組みの構築ができれば良い。 その際にはテストコードを走らせ、自動的にAWSにデプロイしてほしい。勉強用。 ## タスク - [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を踏まえて、疑問・調査事項の解決を実施する。 - [ ] 何をすべきかをタスク化。 ## 疑問・調査事項 - [ ] Github Actionsとは何をするのかを調査。 - [ ] Terraformとは何かを調査。現行は[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)を使用してコード化しているため、必要ないのではないか。 - [ ] letest管理からの脱却。 ## 補足 ・一般的に、TerraformやCloudFormationのソースコードってgithubに上げるべきなのか確認。
priority
github actionsを利用して、awsにdeployできる仕組みの構築。 概要 github actionsを利用して、awsにdeployできる仕組みの構築。 ecrをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。 terraformでインフラをコード化する。 目的 基本的にgithub actionsを利用して、awsにdeployできる仕組みの構築ができれば良い。 その際にはテストコードを走らせ、自動的にawsにデプロイしてほしい。勉強用。 タスク 何をすべきかをタスク化。 疑問・調査事項 github actionsとは何をするのかを調査。 terraformとは何かを調査。現行は letest管理からの脱却。 補足 ・一般的に、terraformやcloudformationのソースコードってgithubに上げるべきなのか確認。
1
57,205
3,081,248,330
IssuesEvent
2015-08-22 14:39:21
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
settime command lost ability to do unlimited time
bug imported Priority-Medium
_From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on May 02, 2014 20:33:55_ in a game, do: /settime 0 It will end the level. I don't remember, was this intended behavior? _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=428_
1.0
settime command lost ability to do unlimited time - _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on May 02, 2014 20:33:55_ in a game, do: /settime 0 It will end the level. I don't remember, was this intended behavior? _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=428_
priority
settime command lost ability to do unlimited time from on may in a game do settime it will end the level i don t remember was this intended behavior original issue
1
201,417
7,031,175,406
IssuesEvent
2017-12-26 15:38:25
xcodeswift/sake
https://api.github.com/repos/xcodeswift/sake
closed
Fix issue with extra line breaks
difficulty:easy good first issue priority:medium status:ready-development
## Context 🕵️‍♀️ There seems to be an issue when printing the output from the shell commands. Sometimes it adds unnecessary extra lines. ## Proposal 🎉 Fix it!
1.0
Fix issue with extra line breaks - ## Context 🕵️‍♀️ There seems to be an issue when printing the output from the shell commands. Sometimes it adds unnecessary extra lines. ## Proposal 🎉 Fix it!
priority
fix issue with extra line breaks context 🕵️‍♀️ there seems to be an issue when printing the output from the shell commands sometimes it adds unnecessary extra lines proposal 🎉 fix it
1
169,156
6,395,853,074
IssuesEvent
2017-08-04 14:14:19
vigetlabs/GroundRules
https://api.github.com/repos/vigetlabs/GroundRules
reopened
[QA] On Mobile, Text Size for Device Problems Too Big/Long
fed medium-priority mobile
Browser: Safari, Mobile URL: GroundRules.co/devices _Current:_ *Device problem text runs into/past the arrow at the end of the line *Device problem text is not vertically centered _Expected:_ *Device problem text should end before the arrow at the end of the line *Device problem text should be vertically centered with arrow
1.0
[QA] On Mobile, Text Size for Device Problems Too Big/Long - Browser: Safari, Mobile URL: GroundRules.co/devices _Current:_ *Device problem text runs into/past the arrow at the end of the line *Device problem text is not vertically centered _Expected:_ *Device problem text should end before the arrow at the end of the line *Device problem text should be vertically centered with arrow
priority
on mobile text size for device problems too big long browser safari mobile url groundrules co devices current device problem text runs into past the arrow at the end of the line device problem text is not vertically centered expected device problem text should end before the arrow at the end of the line device problem text should be vertically centered with arrow
1
168,396
6,370,773,632
IssuesEvent
2017-08-01 14:49:00
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Add support for conflating a randomly calculated tile using the HG workflow
Category: Core Priority: Medium Status: Ready For Review Type: Feature
Create new bug issues doing single tile conflation as they are found.
1.0
Add support for conflating a randomly calculated tile using the HG workflow - Create new bug issues doing single tile conflation as they are found.
priority
add support for conflating a randomly calculated tile using the hg workflow create new bug issues doing single tile conflation as they are found
1
243,875
7,868,156,147
IssuesEvent
2018-06-23 17:51:23
cdgco/VestaWebInterface
https://api.github.com/repos/cdgco/VestaWebInterface
closed
Edit Backup Exclusions
Backend Priority: Medium Status: Accepted Type: Enhancement
Support for editing of backup exclusions. Data must be uploaded from frontend to backend, added to temp file, then directory linked to user. Frontend framework in place.
1.0
Edit Backup Exclusions - Support for editing of backup exclusions. Data must be uploaded from frontend to backend, added to temp file, then directory linked to user. Frontend framework in place.
priority
edit backup exclusions support for editing of backup exclusions data must be uploaded from frontend to backend added to temp file then directory linked to user frontend framework in place
1
723,768
24,907,429,710
IssuesEvent
2022-10-29 12:41:11
KryptonMC/Krypton
https://api.github.com/repos/KryptonMC/Krypton
closed
Implement a backup system
priority: medium area: server
### Description Introduce a backup system that's built-in to the server so users can guarantee the security of their worlds if something goes wrong. ### Goals The following are goals of this system: - Provide a high-performance and efficient automatic backup systems - Provide the ability to automatically backup your worlds on a timer, and also manually queue backups - Provide hooks in to common backup systems, such as AWS S3, Google Drive, or Microsoft OneDrive. - Provide an API for listening for backups and creating new upload sources. - Possibly using a custom data storage format to more efficiently store data than Anvil, and supporting custom writing formats in the API. ### Non goals The following, however, are not goals of this system: - Replace any of the standard loading or saving logic - Using a custom packing format, such as a replacement for ZIP or TAR. ### Motivation Worlds are very important, they are where most of the game takes place, and securing their usability should be vital. However, sometimes things go wrong, and sometimes we can't guarantee this security. In Krypton, this is especially true, as it is quite unstable, and in my time testing it, I have had to delete and repopulate the region folder of the world I use to test so many times that I've actually forgotten most of them. For vanilla Minecraft, and especially Bukkit, there are numerous solutions that attempt to implement backups in to Minecraft. However, I have had mixed results using them in production environments for servers I have worked for in the past, them being slow, inefficient, or consuming huge amounts of memory. This system is intended to replace any possible third-party solutions.
1.0
Implement a backup system - ### Description Introduce a backup system that's built-in to the server so users can guarantee the security of their worlds if something goes wrong. ### Goals The following are goals of this system: - Provide a high-performance and efficient automatic backup systems - Provide the ability to automatically backup your worlds on a timer, and also manually queue backups - Provide hooks in to common backup systems, such as AWS S3, Google Drive, or Microsoft OneDrive. - Provide an API for listening for backups and creating new upload sources. - Possibly using a custom data storage format to more efficiently store data than Anvil, and supporting custom writing formats in the API. ### Non goals The following, however, are not goals of this system: - Replace any of the standard loading or saving logic - Using a custom packing format, such as a replacement for ZIP or TAR. ### Motivation Worlds are very important, they are where most of the game takes place, and securing their usability should be vital. However, sometimes things go wrong, and sometimes we can't guarantee this security. In Krypton, this is especially true, as it is quite unstable, and in my time testing it, I have had to delete and repopulate the region folder of the world I use to test so many times that I've actually forgotten most of them. For vanilla Minecraft, and especially Bukkit, there are numerous solutions that attempt to implement backups in to Minecraft. However, I have had mixed results using them in production environments for servers I have worked for in the past, them being slow, inefficient, or consuming huge amounts of memory. This system is intended to replace any possible third-party solutions.
priority
implement a backup system description introduce a backup system that s built in to the server so users can guarantee the security of their worlds if something goes wrong goals the following are goals of this system provide a high performance and efficient automatic backup systems provide the ability to automatically backup your worlds on a timer and also manually queue backups provide hooks in to common backup systems such as aws google drive or microsoft onedrive provide an api for listening for backups and creating new upload sources possibly using a custom data storage format to more efficiently store data than anvil and supporting custom writing formats in the api non goals the following however are not goals of this system replace any of the standard loading or saving logic using a custom packing format such as a replacement for zip or tar motivation worlds are very important they are where most of the game takes place and securing their usability should be vital however sometimes things go wrong and sometimes we can t guarantee this security in krypton this is especially true as it is quite unstable and in my time testing it i have had to delete and repopulate the region folder of the world i use to test so many times that i ve actually forgotten most of them for vanilla minecraft and especially bukkit there are numerous solutions that attempt to implement backups in to minecraft however i have had mixed results using them in production environments for servers i have worked for in the past them being slow inefficient or consuming huge amounts of memory this system is intended to replace any possible third party solutions
1
484,832
13,957,956,679
IssuesEvent
2020-10-24 09:32:15
sButtons/sbuttons
https://api.github.com/repos/sButtons/sbuttons
closed
Removing border from splash button
Priority: Medium buttons enhancement
**Is your feature request related to a problem? Please describe.** Splash button is having borders of very odd colors , which don't go good with the gradient inside of them. **Describe the solution you'd like** After removing the borders, it will have a good sleek n elegant look. **How the button currently looks \ How it'll look after removing wierd borders :-** The gradient is animated in this button so it looks good with no borders. ![Screenshot (420)](https://user-images.githubusercontent.com/56513758/96821563-ced07900-1445-11eb-94b8-677d4aad8e2e.png) **Additional notes** I'll make a pull request as soon as possible if assigned.
1.0
Removing border from splash button - **Is your feature request related to a problem? Please describe.** Splash button is having borders of very odd colors , which don't go good with the gradient inside of them. **Describe the solution you'd like** After removing the borders, it will have a good sleek n elegant look. **How the button currently looks \ How it'll look after removing wierd borders :-** The gradient is animated in this button so it looks good with no borders. ![Screenshot (420)](https://user-images.githubusercontent.com/56513758/96821563-ced07900-1445-11eb-94b8-677d4aad8e2e.png) **Additional notes** I'll make a pull request as soon as possible if assigned.
priority
removing border from splash button is your feature request related to a problem please describe splash button is having borders of very odd colors which don t go good with the gradient inside of them describe the solution you d like after removing the borders it will have a good sleek n elegant look how the button currently looks how it ll look after removing wierd borders the gradient is animated in this button so it looks good with no borders additional notes i ll make a pull request as soon as possible if assigned
1
731,994
25,240,318,297
IssuesEvent
2022-11-15 06:43:45
OpenMined/PySyft
https://api.github.com/repos/OpenMined/PySyft
closed
Login with an URL doesn't work without specifying port
Type: Improvement :chart_with_upwards_trend: Priority: 3 - Medium :unamused:
## Description When specifying an URL, we need to add the port to both the URL and port parameters. This is not a big deal as the error output helps with that but it might be a better UX to remove the one or the other. ## How to Reproduce Start a domain with `hagrid launch --tag=latest` and run ``` import numpy as np import syft as sy client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0:8081") client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0") ``` ## Expected Behavior If we already specify the port as a parameter we shouldn't specify it in the URL as well. ## Screenshots This is the output when running the code: ![image](https://user-images.githubusercontent.com/29687796/199264346-b48ec68f-b654-4a74-854d-9bb0e0b62817.png) ## Additional Context This is relevant when connecting to a domain over the Internet.
1.0
Login with an URL doesn't work without specifying port - ## Description When specifying an URL, we need to add the port to both the URL and port parameters. This is not a big deal as the error output helps with that but it might be a better UX to remove the one or the other. ## How to Reproduce Start a domain with `hagrid launch --tag=latest` and run ``` import numpy as np import syft as sy client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0:8081") client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0") ``` ## Expected Behavior If we already specify the port as a parameter we shouldn't specify it in the URL as well. ## Screenshots This is the output when running the code: ![image](https://user-images.githubusercontent.com/29687796/199264346-b48ec68f-b654-4a74-854d-9bb0e0b62817.png) ## Additional Context This is relevant when connecting to a domain over the Internet.
priority
login with an url doesn t work without specifying port description when specifying an url we need to add the port to both the url and port parameters this is not a big deal as the error output helps with that but it might be a better ux to remove the one or the other how to reproduce start a domain with hagrid launch tag latest and run import numpy as np import syft as sy client sy login email info openmined org password changethis port url client sy login email info openmined org password changethis port url expected behavior if we already specify the port as a parameter we shouldn t specify it in the url as well screenshots this is the output when running the code additional context this is relevant when connecting to a domain over the internet
1
384,637
11,395,727,632
IssuesEvent
2020-01-30 12:04:42
luna/enso
https://api.github.com/repos/luna/enso
closed
File Manager — Rust Client Library
Category: Libraries Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement
### Summary Provide a Rust library that can connect to the file manager server (#395) and exposes its API. ### Value Being able to use file manager service is necessary for many components in GUI, like the filesystem tree view. ### Specification * connects to the websocket service * implements the file manager websocket protocol * wraps in in the methods that are easily callable from Rust * can target wasm ### Acceptance Criteria & Test Cases Rust code can access filesystem through the running File Manager server.
1.0
File Manager — Rust Client Library - ### Summary Provide a Rust library that can connect to the file manager server (#395) and exposes its API. ### Value Being able to use file manager service is necessary for many components in GUI, like the filesystem tree view. ### Specification * connects to the websocket service * implements the file manager websocket protocol * wraps in in the methods that are easily callable from Rust * can target wasm ### Acceptance Criteria & Test Cases Rust code can access filesystem through the running File Manager server.
priority
file manager — rust client library summary provide a rust library that can connect to the file manager server and exposes its api value being able to use file manager service is necessary for many components in gui like the filesystem tree view specification connects to the websocket service implements the file manager websocket protocol wraps in in the methods that are easily callable from rust can target wasm acceptance criteria test cases rust code can access filesystem through the running file manager server
1
255,079
8,108,257,694
IssuesEvent
2018-08-14 00:35:10
MARKETProtocol/service-listener
https://api.github.com/repos/MARKETProtocol/service-listener
closed
Implement events for contract removal
Priority: Medium To Be Funded Type: Enhancement
<!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### Description Listen for event `AddressRemovedFromWhiteList` and create a new record in the database for the created contract using `MARKET.js` and the `MARKET` api. ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) - [ ] should soft delete from database using `market-api`(sets is_whitelisted to false) - [ ] unit tests ### Additional Information [comment]: # (Any other information that would be useful, content, screenshots, etc.) point of contact @eswarasai or @travisdmathis
1.0
Implement events for contract removal - <!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### Description Listen for event `AddressRemovedFromWhiteList` and create a new record in the database for the created contract using `MARKET.js` and the `MARKET` api. ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) - [ ] should soft delete from database using `market-api`(sets is_whitelisted to false) - [ ] unit tests ### Additional Information [comment]: # (Any other information that would be useful, content, screenshots, etc.) point of contact @eswarasai or @travisdmathis
priority
implement events for contract removal hello please use the template below for issue ideas or bugs found within market protocol if it is general support you need reach out to us at provide a general summary of the issue in the title above and use relevant fields below to define the problem before you start work please read our contribution and if there is a bounty involved please also see if you have ongoing work from other bounties with us where funding has not been released please do not pick up a new issue we would like to involve as many contributors as possible and parallelize the work flow as much as possible please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline please also note that in order for work to be accepted all code must be accompanied by test cases as well description listen for event addressremovedfromwhitelist and create a new record in the database for the created contract using market js and the market api definition of done any other information that would be useful bullets are helpful should soft delete from database using market api sets is whitelisted to false unit tests additional information any other information that would be useful content screenshots etc point of contact eswarasai or travisdmathis
1
508,980
14,709,968,746
IssuesEvent
2021-01-05 03:50:10
syrus-bot/syrus-bot
https://api.github.com/repos/syrus-bot/syrus-bot
opened
feat(commands:core): finish guild settings commands
feat: core priority: medium
**Is your feature request related to a problem? Please describe.** Currently, modification of guild settings is not handled properly. You cannot set multiple array values, you cannot easily change object values, and things are improperly documented. **Describe the solution you'd like** - [ ] Properly document settings modification - [ ] (?) Add smart recursive lookup nesting - [ ] Extend `GuildSchema` on `@data/MongoDB` to include log channel for #52 **Describe alternatives you've considered** N/A. **Additional context** N/A.
1.0
feat(commands:core): finish guild settings commands - **Is your feature request related to a problem? Please describe.** Currently, modification of guild settings is not handled properly. You cannot set multiple array values, you cannot easily change object values, and things are improperly documented. **Describe the solution you'd like** - [ ] Properly document settings modification - [ ] (?) Add smart recursive lookup nesting - [ ] Extend `GuildSchema` on `@data/MongoDB` to include log channel for #52 **Describe alternatives you've considered** N/A. **Additional context** N/A.
priority
feat commands core finish guild settings commands is your feature request related to a problem please describe currently modification of guild settings is not handled properly you cannot set multiple array values you cannot easily change object values and things are improperly documented describe the solution you d like properly document settings modification add smart recursive lookup nesting extend guildschema on data mongodb to include log channel for describe alternatives you ve considered n a additional context n a
1
185,350
6,721,150,594
IssuesEvent
2017-10-16 10:30:56
CS2103AUG2017-T17-B1/main
https://api.github.com/repos/CS2103AUG2017-T17-B1/main
opened
As a debt collector I want to have the ability to perform validity checks on client's personal infomation(HP, email, address)
priority.medium
so that I can verify that their data is genuine
1.0
As a debt collector I want to have the ability to perform validity checks on client's personal infomation(HP, email, address) - so that I can verify that their data is genuine
priority
as a debt collector i want to have the ability to perform validity checks on client s personal infomation hp email address so that i can verify that their data is genuine
1
593,256
17,953,572,222
IssuesEvent
2021-09-13 03:03:34
francheska-vicente/cssweng
https://api.github.com/repos/francheska-vicente/cssweng
closed
Validation and design changes for mobile number is not present in Check-in form
bug priority: medium severity: medium issue: validation
### Summary: Previously implemented validation checks for mobile number (length, starting number) is absent from this form; the change from contact number to mobile number and the placeholder text is also missing ### Steps to Reproduce: - Proceed to Check-in form - Enter 123 for Mobile Number - Click Check-in ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/132490451-6ea5ecd9-24a6-4757-b80c-dea43b2d5899.png) ### Expected Results: - Error Message ### Actual Results: - Invalid input is accepted | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
1.0
Validation and design changes for mobile number is not present in Check-in form - ### Summary: Previously implemented validation checks for mobile number (length, starting number) is absent from this form; the change from contact number to mobile number and the placeholder text is also missing ### Steps to Reproduce: - Proceed to Check-in form - Enter 123 for Mobile Number - Click Check-in ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/132490451-6ea5ecd9-24a6-4757-b80c-dea43b2d5899.png) ### Expected Results: - Error Message ### Actual Results: - Invalid input is accepted | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
priority
validation and design changes for mobile number is not present in check in form summary previously implemented validation checks for mobile number length starting number is absent from this form the change from contact number to mobile number and the placeholder text is also missing steps to reproduce proceed to check in form enter for mobile number click check in visual proof expected results error message actual results invalid input is accepted additional information platform engine google operating system windows
1
632,031
20,168,894,381
IssuesEvent
2022-02-10 08:33:37
TencentBlueKing/bk-iam-saas
https://api.github.com/repos/TencentBlueKing/bk-iam-saas
closed
[日志] 批量优化日志打印
Type: Enhancement Layer: SaaS Priority: Medium Size: M
背景: 目前SaaS的logger需要重新处理, 打的信息过少, 导致信息缺失毫无意义; 要么去掉, 要么打完整; 操作人和被操作对象, 时间, 操作, 来源入口等等 1. Django 日志, info 日志目前只有部分信息,缺乏更全面的信息 例如, 谁 从 哪个group 删除了 谁 2. 什么地方该打日志, 什么地方不应该打? 需要从现有线上日志分类确定下, 去掉高频无用的日志, 优化关键路径的日志 3. 增加layer字段快速定位, 用于定位哪一层打的日志
1.0
[日志] 批量优化日志打印 - 背景: 目前SaaS的logger需要重新处理, 打的信息过少, 导致信息缺失毫无意义; 要么去掉, 要么打完整; 操作人和被操作对象, 时间, 操作, 来源入口等等 1. Django 日志, info 日志目前只有部分信息,缺乏更全面的信息 例如, 谁 从 哪个group 删除了 谁 2. 什么地方该打日志, 什么地方不应该打? 需要从现有线上日志分类确定下, 去掉高频无用的日志, 优化关键路径的日志 3. 增加layer字段快速定位, 用于定位哪一层打的日志
priority
批量优化日志打印 背景 目前saas的logger需要重新处理 打的信息过少 导致信息缺失毫无意义 要么去掉 要么打完整 操作人和被操作对象 时间 操作 来源入口等等 django 日志 info 日志目前只有部分信息 缺乏更全面的信息 例如 谁 从 哪个group 删除了 谁 什么地方该打日志 什么地方不应该打 需要从现有线上日志分类确定下 去掉高频无用的日志 优化关键路径的日志 增加layer字段快速定位 用于定位哪一层打的日志
1
29,494
2,716,237,964
IssuesEvent
2015-04-10 17:51:34
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
The RemoteStreamDataSource is duplicating the last record
bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.2
_From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 13:48:42_ The RemoteStreamingDataSource is duplicating the last record on a page when the number of records = PageSize - 1: Example: If a page can hold five records, but there is only four records, the fourth record will be duplicated to fill the page. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=416_
1.0
The RemoteStreamDataSource is duplicating the last record - _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 13:48:42_ The RemoteStreamingDataSource is duplicating the last record on a page when the number of records = PageSize - 1: Example: If a page can hold five records, but there is only four records, the fourth record will be duplicated to fill the page. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=416_
priority
the remotestreamdatasource is duplicating the last record from on july the remotestreamingdatasource is duplicating the last record on a page when the number of records pagesize example if a page can hold five records but there is only four records the fourth record will be duplicated to fill the page original issue
1
390,966
11,566,600,838
IssuesEvent
2020-02-20 12:49:50
robotology/human-dynamics-estimation
https://api.github.com/repos/robotology/human-dynamics-estimation
closed
Add option to express net external wrench estimates of dummy source (hands) with orientation of world frame
complexity:medium component:HumanDynamicsEstimation component:HumanWrenchProvider priority:high type:enhancement type:task
Currently, the force-torque measurements from the ftShoes are expressed (both origin and orientation) with respect the human foot frames (`LeftFoot` and `RightFoot`). So, on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117), the net external wrench estimates on `LeftFoot` and `RightFoot` links are correctly obtained in the body frame, and if the covariances are correctly set for the MAP estimator, the measurements and the estimates on `LeftFoot` and `RightFoot` links match closely. So, there is no need to modify the function [extractLinkNetExternalWrenchesFromDynamicVariables()](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) for `LeftFoot` and `RightFoot` links. Now, coming to the case of the links `LeftHand` and `RightHand`, they are considered to be dummy sources of force-torques measurements (set to **0**). The net external wrench estimates for `LeftHand` and `RightHand` obtained on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) are expressed (both origin and orientation) in their body frames. To highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands, it is useful to express them at the origin of the links `LeftHand` and `RightHand` but with the orientation of the world frame. As pointed out by @traversaro this code is best suited on the front end of HDE rather than in the back end of Berdy in iDynTree. One of the problems in achieving this is to know inside `HumanDynamicsEstimator` device which link has a dummy wrench source attached. Currently, this information is present in `HumanWrenchProvider` device https://github.com/robotology/human-dynamics-estimation/blob/feature/visualize-berdy-estimated-wrench/conf/xml/Human.xml#L261 but it is not propagated to `HumnaDynamicsEstimator` device. This issue will track the details related to updating HDE code for expressing the estimated net external wrench of dummy sources with the orientation of world frame. @lrapetti @claudia-lat @traversaro
1.0
Add option to express net external wrench estimates of dummy source (hands) with orientation of world frame - Currently, the force-torque measurements from the ftShoes are expressed (both origin and orientation) with respect the human foot frames (`LeftFoot` and `RightFoot`). So, on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117), the net external wrench estimates on `LeftFoot` and `RightFoot` links are correctly obtained in the body frame, and if the covariances are correctly set for the MAP estimator, the measurements and the estimates on `LeftFoot` and `RightFoot` links match closely. So, there is no need to modify the function [extractLinkNetExternalWrenchesFromDynamicVariables()](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) for `LeftFoot` and `RightFoot` links. Now, coming to the case of the links `LeftHand` and `RightHand`, they are considered to be dummy sources of force-torques measurements (set to **0**). The net external wrench estimates for `LeftHand` and `RightHand` obtained on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) are expressed (both origin and orientation) in their body frames. To highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands, it is useful to express them at the origin of the links `LeftHand` and `RightHand` but with the orientation of the world frame. As pointed out by @traversaro this code is best suited on the front end of HDE rather than in the back end of Berdy in iDynTree. One of the problems in achieving this is to know inside `HumanDynamicsEstimator` device which link has a dummy wrench source attached. Currently, this information is present in `HumanWrenchProvider` device https://github.com/robotology/human-dynamics-estimation/blob/feature/visualize-berdy-estimated-wrench/conf/xml/Human.xml#L261 but it is not propagated to `HumnaDynamicsEstimator` device. This issue will track the details related to updating HDE code for expressing the estimated net external wrench of dummy sources with the orientation of world frame. @lrapetti @claudia-lat @traversaro
priority
add option to express net external wrench estimates of dummy source hands with orientation of world frame currently the force torque measurements from the ftshoes are expressed both origin and orientation with respect the human foot frames leftfoot and rightfoot so on calling the net external wrench estimates on leftfoot and rightfoot links are correctly obtained in the body frame and if the covariances are correctly set for the map estimator the measurements and the estimates on leftfoot and rightfoot links match closely so there is no need to modify the function for leftfoot and rightfoot links now coming to the case of the links lefthand and righthand they are considered to be dummy sources of force torques measurements set to the net external wrench estimates for lefthand and righthand obtained on calling are expressed both origin and orientation in their body frames to highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands it is useful to express them at the origin of the links lefthand and righthand but with the orientation of the world frame as pointed out by traversaro this code is best suited on the front end of hde rather than in the back end of berdy in idyntree one of the problems in achieving this is to know inside humandynamicsestimator device which link has a dummy wrench source attached currently this information is present in humanwrenchprovider device but it is not propagated to humnadynamicsestimator device this issue will track the details related to updating hde code for expressing the estimated net external wrench of dummy sources with the orientation of world frame lrapetti claudia lat traversaro
1
122,649
4,838,540,124
IssuesEvent
2016-11-09 04:05:12
smirkspace/smirkspace
https://api.github.com/repos/smirkspace/smirkspace
opened
Add Analytics
backend enhancement medium priority
Add Analytics such as how many times a button is pressed, how much time spent on a page, etc. Possibly use MixPanel for this.
1.0
Add Analytics - Add Analytics such as how many times a button is pressed, how much time spent on a page, etc. Possibly use MixPanel for this.
priority
add analytics add analytics such as how many times a button is pressed how much time spent on a page etc possibly use mixpanel for this
1
56,247
3,078,627,002
IssuesEvent
2015-08-21 11:38:18
nfprojects/nfengine
https://api.github.com/repos/nfprojects/nfengine
closed
Remove all useless #if ... #else ... #endif sections
bug high priority medium
Some parts of engine have code hidden by #if ... #else ... #endif sequence. Search for all of them and either remove them, or provide different way to determine which section to use (avoid preprocessor macros, we want the engine to be entirely compiled).
1.0
Remove all useless #if ... #else ... #endif sections - Some parts of engine have code hidden by #if ... #else ... #endif sequence. Search for all of them and either remove them, or provide different way to determine which section to use (avoid preprocessor macros, we want the engine to be entirely compiled).
priority
remove all useless if else endif sections some parts of engine have code hidden by if else endif sequence search for all of them and either remove them or provide different way to determine which section to use avoid preprocessor macros we want the engine to be entirely compiled
1
376,160
11,139,228,124
IssuesEvent
2019-12-21 03:06:46
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
Error in values[index] (only 0's may be mixed with negative subscripts) when one variable in df contains only zeroes
Difficulty: ② Intermediate Effort: ② Medium Priority: ♨︎ Critical Type: ☹︎ Bug
First of all, let me say congrats for this awseome package gt you are building! :-) Issue detected: If you have a dataframe more than one variable, and one of them contains only empty values, you end up hitting this error: `Error in values[index] : only 0's may be mixed with negative subscripts` (using R 3.6.x under GNU/Linux 64 bits, in case it matters) Reproducible example: ```r data.frame(x=rnorm(10), y=rep(0,10)) %>% gt() %>% fmt_number( columns = vars(x, y), decimals = 2, suffixing = TRUE ) ``` FYI, If you attempt to make the similar reprex but only with one column (y in the previous example), then no error is triggered and the expected result is produced. ```r data.frame(y=rep(0,10)) %>% gt() %>% fmt_number( columns = vars(y), decimals = 2, suffixing = TRUE ) ```
1.0
Error in values[index] (only 0's may be mixed with negative subscripts) when one variable in df contains only zeroes - First of all, let me say congrats for this awseome package gt you are building! :-) Issue detected: If you have a dataframe more than one variable, and one of them contains only empty values, you end up hitting this error: `Error in values[index] : only 0's may be mixed with negative subscripts` (using R 3.6.x under GNU/Linux 64 bits, in case it matters) Reproducible example: ```r data.frame(x=rnorm(10), y=rep(0,10)) %>% gt() %>% fmt_number( columns = vars(x, y), decimals = 2, suffixing = TRUE ) ``` FYI, If you attempt to make the similar reprex but only with one column (y in the previous example), then no error is triggered and the expected result is produced. ```r data.frame(y=rep(0,10)) %>% gt() %>% fmt_number( columns = vars(y), decimals = 2, suffixing = TRUE ) ```
priority
error in values only s may be mixed with negative subscripts when one variable in df contains only zeroes first of all let me say congrats for this awseome package gt you are building issue detected if you have a dataframe more than one variable and one of them contains only empty values you end up hitting this error error in values only s may be mixed with negative subscripts using r x under gnu linux bits in case it matters reproducible example r data frame x rnorm y rep gt fmt number columns vars x y decimals suffixing true fyi if you attempt to make the similar reprex but only with one column y in the previous example then no error is triggered and the expected result is produced r data frame y rep gt fmt number columns vars y decimals suffixing true
1
44,318
2,903,519,828
IssuesEvent
2015-06-18 13:51:23
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
[Staging & Production] Dataverse: Delete fails when trying to delete a draft file
2 - ready bug: production priority - medium
## Steps 1. Go to a public project with Dataverse authorized (link: https://osf.io/9jtkg/ ) 2. Go to the files tab 3. Switch the Dataverse folder from Released to Draft 4. Wait for the folder to re-load with draft files 5. Click on the red 'x' next to a file to delete it 6. Hit 'OK' to confirm ## Expected That a dataverse draft file can be deleted from the files tab ## Actual There is an error in the file grid stating that 'Delete failed'. Also the console reports a 500 error. ![screenshot 2015-04-13 14 36 20](https://cloud.githubusercontent.com/assets/9381062/7122722/3d3b595c-e1eb-11e4-92ef-bfc4b2bae79b.png)
1.0
[Staging & Production] Dataverse: Delete fails when trying to delete a draft file - ## Steps 1. Go to a public project with Dataverse authorized (link: https://osf.io/9jtkg/ ) 2. Go to the files tab 3. Switch the Dataverse folder from Released to Draft 4. Wait for the folder to re-load with draft files 5. Click on the red 'x' next to a file to delete it 6. Hit 'OK' to confirm ## Expected That a dataverse draft file can be deleted from the files tab ## Actual There is an error in the file grid stating that 'Delete failed'. Also the console reports a 500 error. ![screenshot 2015-04-13 14 36 20](https://cloud.githubusercontent.com/assets/9381062/7122722/3d3b595c-e1eb-11e4-92ef-bfc4b2bae79b.png)
priority
dataverse delete fails when trying to delete a draft file steps go to a public project with dataverse authorized link go to the files tab switch the dataverse folder from released to draft wait for the folder to re load with draft files click on the red x next to a file to delete it hit ok to confirm expected that a dataverse draft file can be deleted from the files tab actual there is an error in the file grid stating that delete failed also the console reports a error
1
26,330
2,684,319,085
IssuesEvent
2015-03-28 21:28:10
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
closed
Crashes when I press Win+Right or Win+Left
1 star bug duplicate imported Priority-Medium
_From [CoreyJJo...@gmail.com](https://code.google.com/u/117850621500591170194/) on March 19, 2013 07:08:08_ Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* ... Crashes when I press Win+Right or Win+Left to move the window on Win7 Pro 64-bit *Steps to reproduction* 1. 2. 3. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=993_
1.0
Crashes when I press Win+Right or Win+Left - _From [CoreyJJo...@gmail.com](https://code.google.com/u/117850621500591170194/) on March 19, 2013 07:08:08_ Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* ... Crashes when I press Win+Right or Win+Left to move the window on Win7 Pro 64-bit *Steps to reproduction* 1. 2. 3. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=993_
priority
crashes when i press win right or win left from on march required information os version winxp vista sp conemu version far version if you are using far manager bug description crashes when i press win right or win left to move the window on pro bit steps to reproduction original issue
1
223,118
7,446,984,076
IssuesEvent
2018-03-28 10:55:35
vanilla-framework/vanilla-framework
https://api.github.com/repos/vanilla-framework/vanilla-framework
closed
Margin top on br in Firefox
Priority: Medium Type: Bug
Go to: https://snapcraft.io/spotify There is unnecessary spacing on bullet points, because Vanilla adds top margin to `br`. Related to: #1626 Also reported in snapcraft: https://github.com/canonical-websites/snapcraft.io/issues/405 What you see in Chromium: | What you see in Firefox: -|- ![snapcraft io spotify description chromium](https://user-images.githubusercontent.com/19801137/37516792-40c21190-2907-11e8-9450-2f44b1592885.png) | ![snapcraft io spotify description firefox](https://user-images.githubusercontent.com/19801137/37516793-4123d93e-2907-11e8-803b-fa22cd963f89.png) What you should see: The description in Firefox should not have all the extra line spacing.
1.0
Margin top on br in Firefox - Go to: https://snapcraft.io/spotify There is unnecessary spacing on bullet points, because Vanilla adds top margin to `br`. Related to: #1626 Also reported in snapcraft: https://github.com/canonical-websites/snapcraft.io/issues/405 What you see in Chromium: | What you see in Firefox: -|- ![snapcraft io spotify description chromium](https://user-images.githubusercontent.com/19801137/37516792-40c21190-2907-11e8-9450-2f44b1592885.png) | ![snapcraft io spotify description firefox](https://user-images.githubusercontent.com/19801137/37516793-4123d93e-2907-11e8-803b-fa22cd963f89.png) What you should see: The description in Firefox should not have all the extra line spacing.
priority
margin top on br in firefox go to there is unnecessary spacing on bullet points because vanilla adds top margin to br related to also reported in snapcraft what you see in chromium what you see in firefox what you should see the description in firefox should not have all the extra line spacing
1
759,903
26,617,638,635
IssuesEvent
2023-01-24 08:45:23
apache/airflow
https://api.github.com/repos/apache/airflow
closed
graph disappears during run time when using branch_task and a dynamic classic operator
kind:bug good first issue priority:medium area:UI
### Apache Airflow version 2.5.1 ### What happened when using a dynamically generated task that gets the expand data from xcom after a branch_task the graph doesn't render. It reappears once the dag run is finished. tried with BashOperator and a KubernetesPodOperator. the developer console in the browser shows the error: `Uncaught TypeError: Cannot read properties of undefined (reading 'length') at z (graph.1c0596dfced26c638bfe.js:2:17499) at graph.1c0596dfced26c638bfe.js:2:17654 at Array.map (<anonymous>) at z (graph.1c0596dfced26c638bfe.js:2:17646) at graph.1c0596dfced26c638bfe.js:2:26602 at graph.1c0596dfced26c638bfe.js:2:26655 at graph.1c0596dfced26c638bfe.js:2:26661 at graph.1c0596dfced26c638bfe.js:2:222 at graph.1c0596dfced26c638bfe.js:2:227 z @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 z @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 ` grid view renders fine. ### What you think should happen instead graph should be rendered. ### How to reproduce ```@dag('branch_dynamic', schedule_interval=None, default_args=default_args, catchup=False) def branch_dynamic_flow(): @branch_task def choose_path(): return 'b' @task def a(): print('a') @task def get_args(): return ['echo 1', 'echo 2'] b = BashOperator.partial(task_id="b").expand(bash_command=get_args()) path = choose_path() path >> a() path >> b ``` ### Operating System red hat ### Versions of Apache Airflow Providers apache-airflow-providers-cncf-kubernetes | 5.1.1 | Kubernetes ### Deployment Official Apache Airflow Helm Chart ### Deployment details _No response_ ### Anything else _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
1.0
graph disappears during run time when using branch_task and a dynamic classic operator - ### Apache Airflow version 2.5.1 ### What happened when using a dynamically generated task that gets the expand data from xcom after a branch_task the graph doesn't render. It reappears once the dag run is finished. tried with BashOperator and a KubernetesPodOperator. the developer console in the browser shows the error: `Uncaught TypeError: Cannot read properties of undefined (reading 'length') at z (graph.1c0596dfced26c638bfe.js:2:17499) at graph.1c0596dfced26c638bfe.js:2:17654 at Array.map (<anonymous>) at z (graph.1c0596dfced26c638bfe.js:2:17646) at graph.1c0596dfced26c638bfe.js:2:26602 at graph.1c0596dfced26c638bfe.js:2:26655 at graph.1c0596dfced26c638bfe.js:2:26661 at graph.1c0596dfced26c638bfe.js:2:222 at graph.1c0596dfced26c638bfe.js:2:227 z @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 z @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 (anonymous) @ graph.1c0596dfced26c638bfe.js:2 ` grid view renders fine. ### What you think should happen instead graph should be rendered. ### How to reproduce ```@dag('branch_dynamic', schedule_interval=None, default_args=default_args, catchup=False) def branch_dynamic_flow(): @branch_task def choose_path(): return 'b' @task def a(): print('a') @task def get_args(): return ['echo 1', 'echo 2'] b = BashOperator.partial(task_id="b").expand(bash_command=get_args()) path = choose_path() path >> a() path >> b ``` ### Operating System red hat ### Versions of Apache Airflow Providers apache-airflow-providers-cncf-kubernetes | 5.1.1 | Kubernetes ### Deployment Official Apache Airflow Helm Chart ### Deployment details _No response_ ### Anything else _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
priority
graph disappears during run time when using branch task and a dynamic classic operator apache airflow version what happened when using a dynamically generated task that gets the expand data from xcom after a branch task the graph doesn t render it reappears once the dag run is finished tried with bashoperator and a kubernetespodoperator the developer console in the browser shows the error uncaught typeerror cannot read properties of undefined reading length at z graph js at graph js at array map at z graph js at graph js at graph js at graph js at graph js at graph js z graph js anonymous graph js z graph js anonymous graph js anonymous graph js anonymous graph js anonymous graph js anonymous graph js grid view renders fine what you think should happen instead graph should be rendered how to reproduce dag branch dynamic schedule interval none default args default args catchup false def branch dynamic flow branch task def choose path return b task def a print a task def get args return b bashoperator partial task id b expand bash command get args path choose path path a path b operating system red hat versions of apache airflow providers apache airflow providers cncf kubernetes kubernetes deployment official apache airflow helm chart deployment details no response anything else no response are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
1
438,847
12,652,489,481
IssuesEvent
2020-06-17 03:45:29
Open-Source-Medical-Supplies/react-projects-gallery
https://api.github.com/repos/Open-Source-Medical-Supplies/react-projects-gallery
closed
Full card: semantics per Kimly
correction medium priority
1. 'Reviewed By', text should be, "Medical Status:" 2. the text from the tags 'Reviewed By' does need to be shown, ex 'Affiliated with: UW Health' 3. Text of Medical Supply Category should, be bold or somehow different from the Title Header Style, and link back to category supply page 4. In Materials List, underneath Material Name, should be hyper formatted text from 'Detail', base 'Bill of Materials'
1.0
Full card: semantics per Kimly - 1. 'Reviewed By', text should be, "Medical Status:" 2. the text from the tags 'Reviewed By' does need to be shown, ex 'Affiliated with: UW Health' 3. Text of Medical Supply Category should, be bold or somehow different from the Title Header Style, and link back to category supply page 4. In Materials List, underneath Material Name, should be hyper formatted text from 'Detail', base 'Bill of Materials'
priority
full card semantics per kimly reviewed by text should be medical status the text from the tags reviewed by does need to be shown ex affiliated with uw health text of medical supply category should be bold or somehow different from the title header style and link back to category supply page in materials list underneath material name should be hyper formatted text from detail base bill of materials
1
670,180
22,678,837,531
IssuesEvent
2022-07-04 08:03:54
input-output-hk/cardano-graphql
https://api.github.com/repos/input-output-hk/cardano-graphql
closed
Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes.
BUG SEVERITY:LOW PRIORITY:MEDIUM
### Environment Testnet https://gimbalabs.com/dandelion/endpoints/graphql-api ### Steps to reproduce the bug Send well formatted transaction with: ```graphql mutation submitTransaction( $transaction: String! ) { submitTransaction(transaction: $transaction) { hash } } ``` I got the error: `Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes.` But the transaction is actually send to network: https://explorer.cardano-testnet.iohkdev.io/en/transaction?id=6f5762b2e52cd8f81380e197daae2ab6bded8db34e62346c3434085c0b6b27cd ### What is the expected behavior? No false negative error.
1.0
Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes. - ### Environment Testnet https://gimbalabs.com/dandelion/endpoints/graphql-api ### Steps to reproduce the bug Send well formatted transaction with: ```graphql mutation submitTransaction( $transaction: String! ) { submitTransaction(transaction: $transaction) { hash } } ``` I got the error: `Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes.` But the transaction is actually send to network: https://explorer.cardano-testnet.iohkdev.io/en/transaction?id=6f5762b2e52cd8f81380e197daae2ab6bded8db34e62346c3434085c0b6b27cd ### What is the expected behavior? No false negative error.
priority
deserialization failed in transaction because invalid cbor not enough bytes expect bytes but received bytes environment testnet steps to reproduce the bug send well formatted transaction with graphql mutation submittransaction transaction string submittransaction transaction transaction hash i got the error deserialization failed in transaction because invalid cbor not enough bytes expect bytes but received bytes but the transaction is actually send to network what is the expected behavior no false negative error
1
638,830
20,739,650,985
IssuesEvent
2022-03-14 16:31:03
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
closed
MD - Incorrect observation status of "X" in SimpleReport
onboarding-ops receiver data-issue Medium Priority
MD is also reporting receiving "X" in OBR-25 under observation status instead of P, F or C Sample messages Complete Care Severna Park 2/9/22 5:36 AM ```3a12a0df-d1ab-4556-a477-54ce808c188f``` sanchez pediatrics 2/10/22 11:26 AM ```ae25a69f-5726-48f3-b48d-1192d743e022``` ResultsLab,LLC 2/14/22 5:30 AM ```7a9e18e3-63c0-435a-9da5-a44d3fa90833``` Brightview White Marsh 2/8/22 2:16 PM ```ae25a69f-5726-48f3-b48d-1192d743e022```
1.0
MD - Incorrect observation status of "X" in SimpleReport - MD is also reporting receiving "X" in OBR-25 under observation status instead of P, F or C Sample messages Complete Care Severna Park 2/9/22 5:36 AM ```3a12a0df-d1ab-4556-a477-54ce808c188f``` sanchez pediatrics 2/10/22 11:26 AM ```ae25a69f-5726-48f3-b48d-1192d743e022``` ResultsLab,LLC 2/14/22 5:30 AM ```7a9e18e3-63c0-435a-9da5-a44d3fa90833``` Brightview White Marsh 2/8/22 2:16 PM ```ae25a69f-5726-48f3-b48d-1192d743e022```
priority
md incorrect observation status of x in simplereport md is also reporting receiving x in obr under observation status instead of p f or c sample messages complete care severna park am sanchez pediatrics am resultslab llc am brightview white marsh pm
1
302,676
9,285,331,388
IssuesEvent
2019-03-21 06:37:18
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
tests/kernel/threads/no-multithreading/testcase.yam: Not Booting
area: NIOS2 bug priority: medium
**Describe the bug** Failed to boot up on altera_max10 board **To Reproduce** Steps to reproduce the behavior: 1. mkdir build; cd build 2. cmake -DBOARD=altera_max10 3. make 4. make run **Expected behavior** Should able to boot zephyr and show log on console output. **Screenshots or console output** no console output. OS: fedora29 Toolchain: zephyr Commit SHA or Version used: 6368ed4
1.0
tests/kernel/threads/no-multithreading/testcase.yam: Not Booting - **Describe the bug** Failed to boot up on altera_max10 board **To Reproduce** Steps to reproduce the behavior: 1. mkdir build; cd build 2. cmake -DBOARD=altera_max10 3. make 4. make run **Expected behavior** Should able to boot zephyr and show log on console output. **Screenshots or console output** no console output. OS: fedora29 Toolchain: zephyr Commit SHA or Version used: 6368ed4
priority
tests kernel threads no multithreading testcase yam not booting describe the bug failed to boot up on altera board to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard altera make make run expected behavior should able to boot zephyr and show log on console output screenshots or console output no console output os toolchain zephyr commit sha or version used
1
228,719
7,567,005,138
IssuesEvent
2018-04-22 04:03:34
compodoc/compodoc
https://api.github.com/repos/compodoc/compodoc
closed
[BUG] UsersAdminComponent has a templateUrl, include it. Cannot read template for UsersAdminComponent
1. Type: Bug Priority: Medium Status: Accepted Status: Waiting feedback Time: ~1 hour
##### **Overview of the issue** I am unable to run it for templates the application compiles and runs perfectly. the template is in the same folder as component **I get the following error:** UserPreferencesComponent has a templateUrl, include it [15:13:37] UserProfileComponent has a templateUrl, include it [15:13:37] UserRegistrationComponent has a templateUrl, include it [15:13:37] UserSigninComponent has a templateUrl, include it [15:13:37] UserVerificationComponent has a templateUrl, include it [15:13:37] UsersAdminComponent has a templateUrl, include it [15:13:37] Cannot read template for UsersAdminComponent ##### **Operating System, Node.js, npm, compodoc version(s)** angular/cli: 1.4.6 node: 6.11.3 os: darwin x64 angular/animations: 5.1.2 angular/cdk: 5.0.2 angular/common: 5.1.2 angular/compiler: 5.1.2 angular/core: 5.1.2 angular/forms: 5.1.2 angular/http: 5.1.2 angular/material: 5.0.2 angular/platform-browser: 5.1.2 angular/platform-browser-dynamic: 5.1.2 angular/router: 5.1.2 angular/cli: 1.4.6 angular/compiler-cli: 5.1.2 angular/language-service: 5.1.2 typescript: 2.5.3 ##### **Angular configuration, a `package.json` file in the root folder** **users-admin.component.ts** ``` @Component({ selector: 'app-users-admin', templateUrl: './users-admin.component.html', styleUrls: ['./users-admin.component.scss'] }) export class UsersAdminComponent implements OnInit { ``` **users-admin.component.html** ``` <p> <button mat-raised-button color="primary" [routerLink]="['/users/atsadmin']">Back to Dashboard</button> </p> <p> users-admin works! </p> <div class="master-detail-container"> <div class="master-detail-header"> </div> <div class="master-detail"> <div class="master"> <mat-card> <mat-card-title-group> <mat-card-title>Users</mat-card-title> <mat-card-subtitle>Subtitle</mat-card-subtitle> <img matCardImage src="Side image"> </mat-card-title-group> <img matCardImage src="Card wide image source"> <mat-card-content> <mat-list dense> <mat-list-item *ngFor="let user of users; let i=index" (click)="switchUser(i)"> {{user.firstName}} {{user.lastName}} </mat-list-item> </mat-list> </mat-card-content> <mat-card-actions align="start"> <button mat-button matTooltip="Add User" (click)="onAction(event)">Add User</button> <button mat-button matTooltip="Delete User" (click)="onAction(event)">Delete Users</button> </mat-card-actions> <mat-card-footer> Footer </mat-card-footer> </mat-card> </div> <div class="detail"> <mat-card> <mat-card-title-group> <mat-card-title>User Details</mat-card-title> <mat-card-subtitle>details</mat-card-subtitle> <img matCardImage src="Side image"> </mat-card-title-group> <img matCardImage src="Card wide image source"> <mat-card-content> <form *ngIf="currentUser" [formGroup]="userDetailsFormGroup"> <mat-form-field> <input matInput formControlName="firstName" type="text" placeholder="First Name" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['firstName'].valid && userDetailsFormGroup.controls['firstName'].touched && userDetailsFormGroup.controls['firstName'].errors.required"> First Name is required </mat-error> </mat-form-field> <mat-form-field> <input matInput formControlName="lastName" type="text" placeholder="Last Name" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['lastName'].valid && userDetailsFormGroup.controls['lastName'].touched && userDetailsFormGroup.controls['lastName'].errors.required"> Last Name is required </mat-error> </mat-form-field> <mat-form-field> <mat-select formControlName="roles" placeholder="Roles" multiple required> <mat-option *ngFor="let role of rolesList" [value]=role.key>{{ role.displayedValue }}</mat-option> </mat-select> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['roles'].valid && userDetailsFormGroup.controls['roles'].touched && userDetailsFormGroup.controls['roles'].errors.required"> Role is required </mat-error> </mat-form-field> <mat-form-field> <input matInput formControlName="email" type="email" placeholder="Email" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid && userDetailsFormGroup.controls['email'].touched && userDetailsFormGroup.controls['email'].errors.required"> Email is required </mat-error> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid && userDetailsFormGroup.controls['email'].touched && userDetailsFormGroup.controls['email'].errors.email"> Email is not valid </mat-error> </mat-form-field> <button mat-raised-button matTooltip="Add Course to this user" (click)="openSelectItemDialog()">Add Course</button> </form> </mat-card-content> <mat-card-actions align="start"> <button mat-button matTooltip="Save User" (click)="saveUser(event)">Save User</button> <button mat-button matTooltip="Delete User" (click)="deleteUser(event)">Delete User</button> </mat-card-actions> <mat-card-footer> Footer </mat-card-footer> </mat-card> </div> <!-- end detail --> </div> <!-- end master detail --> </div> <!-- end master detail container--> ``` <!-- This is mandatory for bug reports. This will help us to replicate the scenario. --> ##### **Compodoc installed globally or locally ?** Locally ##### **Motivation for or Use Case** <Unable to resolve templates ##### **Reproduce the error** <!-- an unambiguous set of steps to reproduce the error. --> ##### **Related issues** <!-- has a similar issue been reported before? --> ##### **Suggest a Fix** no idea <!-- Love compodoc? Please consider supporting our collective: 👉 https://opencollective.com/compodoc/donate -->
1.0
[BUG] UsersAdminComponent has a templateUrl, include it. Cannot read template for UsersAdminComponent - ##### **Overview of the issue** I am unable to run it for templates the application compiles and runs perfectly. the template is in the same folder as component **I get the following error:** UserPreferencesComponent has a templateUrl, include it [15:13:37] UserProfileComponent has a templateUrl, include it [15:13:37] UserRegistrationComponent has a templateUrl, include it [15:13:37] UserSigninComponent has a templateUrl, include it [15:13:37] UserVerificationComponent has a templateUrl, include it [15:13:37] UsersAdminComponent has a templateUrl, include it [15:13:37] Cannot read template for UsersAdminComponent ##### **Operating System, Node.js, npm, compodoc version(s)** angular/cli: 1.4.6 node: 6.11.3 os: darwin x64 angular/animations: 5.1.2 angular/cdk: 5.0.2 angular/common: 5.1.2 angular/compiler: 5.1.2 angular/core: 5.1.2 angular/forms: 5.1.2 angular/http: 5.1.2 angular/material: 5.0.2 angular/platform-browser: 5.1.2 angular/platform-browser-dynamic: 5.1.2 angular/router: 5.1.2 angular/cli: 1.4.6 angular/compiler-cli: 5.1.2 angular/language-service: 5.1.2 typescript: 2.5.3 ##### **Angular configuration, a `package.json` file in the root folder** **users-admin.component.ts** ``` @Component({ selector: 'app-users-admin', templateUrl: './users-admin.component.html', styleUrls: ['./users-admin.component.scss'] }) export class UsersAdminComponent implements OnInit { ``` **users-admin.component.html** ``` <p> <button mat-raised-button color="primary" [routerLink]="['/users/atsadmin']">Back to Dashboard</button> </p> <p> users-admin works! </p> <div class="master-detail-container"> <div class="master-detail-header"> </div> <div class="master-detail"> <div class="master"> <mat-card> <mat-card-title-group> <mat-card-title>Users</mat-card-title> <mat-card-subtitle>Subtitle</mat-card-subtitle> <img matCardImage src="Side image"> </mat-card-title-group> <img matCardImage src="Card wide image source"> <mat-card-content> <mat-list dense> <mat-list-item *ngFor="let user of users; let i=index" (click)="switchUser(i)"> {{user.firstName}} {{user.lastName}} </mat-list-item> </mat-list> </mat-card-content> <mat-card-actions align="start"> <button mat-button matTooltip="Add User" (click)="onAction(event)">Add User</button> <button mat-button matTooltip="Delete User" (click)="onAction(event)">Delete Users</button> </mat-card-actions> <mat-card-footer> Footer </mat-card-footer> </mat-card> </div> <div class="detail"> <mat-card> <mat-card-title-group> <mat-card-title>User Details</mat-card-title> <mat-card-subtitle>details</mat-card-subtitle> <img matCardImage src="Side image"> </mat-card-title-group> <img matCardImage src="Card wide image source"> <mat-card-content> <form *ngIf="currentUser" [formGroup]="userDetailsFormGroup"> <mat-form-field> <input matInput formControlName="firstName" type="text" placeholder="First Name" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['firstName'].valid && userDetailsFormGroup.controls['firstName'].touched && userDetailsFormGroup.controls['firstName'].errors.required"> First Name is required </mat-error> </mat-form-field> <mat-form-field> <input matInput formControlName="lastName" type="text" placeholder="Last Name" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['lastName'].valid && userDetailsFormGroup.controls['lastName'].touched && userDetailsFormGroup.controls['lastName'].errors.required"> Last Name is required </mat-error> </mat-form-field> <mat-form-field> <mat-select formControlName="roles" placeholder="Roles" multiple required> <mat-option *ngFor="let role of rolesList" [value]=role.key>{{ role.displayedValue }}</mat-option> </mat-select> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['roles'].valid && userDetailsFormGroup.controls['roles'].touched && userDetailsFormGroup.controls['roles'].errors.required"> Role is required </mat-error> </mat-form-field> <mat-form-field> <input matInput formControlName="email" type="email" placeholder="Email" required> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid && userDetailsFormGroup.controls['email'].touched && userDetailsFormGroup.controls['email'].errors.required"> Email is required </mat-error> <mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid && userDetailsFormGroup.controls['email'].touched && userDetailsFormGroup.controls['email'].errors.email"> Email is not valid </mat-error> </mat-form-field> <button mat-raised-button matTooltip="Add Course to this user" (click)="openSelectItemDialog()">Add Course</button> </form> </mat-card-content> <mat-card-actions align="start"> <button mat-button matTooltip="Save User" (click)="saveUser(event)">Save User</button> <button mat-button matTooltip="Delete User" (click)="deleteUser(event)">Delete User</button> </mat-card-actions> <mat-card-footer> Footer </mat-card-footer> </mat-card> </div> <!-- end detail --> </div> <!-- end master detail --> </div> <!-- end master detail container--> ``` <!-- This is mandatory for bug reports. This will help us to replicate the scenario. --> ##### **Compodoc installed globally or locally ?** Locally ##### **Motivation for or Use Case** <Unable to resolve templates ##### **Reproduce the error** <!-- an unambiguous set of steps to reproduce the error. --> ##### **Related issues** <!-- has a similar issue been reported before? --> ##### **Suggest a Fix** no idea <!-- Love compodoc? Please consider supporting our collective: 👉 https://opencollective.com/compodoc/donate -->
priority
usersadmincomponent has a templateurl include it cannot read template for usersadmincomponent overview of the issue i am unable to run it for templates the application compiles and runs perfectly the template is in the same folder as component i get the following error userpreferencescomponent has a templateurl include it userprofilecomponent has a templateurl include it userregistrationcomponent has a templateurl include it usersignincomponent has a templateurl include it userverificationcomponent has a templateurl include it usersadmincomponent has a templateurl include it cannot read template for usersadmincomponent operating system node js npm compodoc version s angular cli node os darwin angular animations angular cdk angular common angular compiler angular core angular forms angular http angular material angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli angular language service typescript angular configuration a package json file in the root folder users admin component ts component selector app users admin templateurl users admin component html styleurls export class usersadmincomponent implements oninit users admin component html back to dashboard users admin works users subtitle user firstname user lastname add user delete users footer user details details mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required first name is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required last name is required role displayedvalue mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required role is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required email is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors email email is not valid add course save user delete user footer compodoc installed globally or locally locally motivation for or use case unable to resolve templates reproduce the error related issues suggest a fix no idea love compodoc please consider supporting our collective 👉
1
112,676
4,535,786,650
IssuesEvent
2016-09-08 18:25:03
newamericafoundation/newamerica-data-projects
https://api.github.com/repos/newamericafoundation/newamerica-data-projects
closed
Title panel heading css
medium-priority
Font seemed to have gotten smaller and padding was added ` .title-panel__title { font-size: 3rem; } .title-panel__title { text-align: center; font-weight: 700; font-size: 2rem; line-height: 1.2; padding: 0px 0; }`
1.0
Title panel heading css - Font seemed to have gotten smaller and padding was added ` .title-panel__title { font-size: 3rem; } .title-panel__title { text-align: center; font-weight: 700; font-size: 2rem; line-height: 1.2; padding: 0px 0; }`
priority
title panel heading css font seemed to have gotten smaller and padding was added title panel title font size title panel title text align center font weight font size line height padding
1