Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12 values | text_combine stringlengths 96 259k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
257,968 | 8,149,305,678 | IssuesEvent | 2018-08-22 09:12:58 | Xceptance/neodymium-library | https://api.github.com/repos/Xceptance/neodymium-library | closed | Attachments in Allure report missing | Medium Priority bug doneInDevelop | Selenide attachments created while testing are currently not attached to Allure report. | 1.0 | Attachments in Allure report missing - Selenide attachments created while testing are currently not attached to Allure report. | priority | attachments in allure report missing selenide attachments created while testing are currently not attached to allure report | 1 |
660,060 | 21,950,461,107 | IssuesEvent | 2022-05-24 07:26:12 | dodona-edu/dodona | https://api.github.com/repos/dodona-edu/dodona | closed | Show transient messages to users | feature medium priority | We want a way to easily show messages to all users (or all course admins/staff/...). Such message would be displayed until it is dismissed or until we disable it.
This could be used to announce new releases, scheduled downtime, etc.
We want to be able to enable/disable this using an interface on Dodona and not doing a deploy. | 1.0 | Show transient messages to users - We want a way to easily show messages to all users (or all course admins/staff/...). Such message would be displayed until it is dismissed or until we disable it.
This could be used to announce new releases, scheduled downtime, etc.
We want to be able to enable/disable this using an interface on Dodona and not doing a deploy. | priority | show transient messages to users we want a way to easily show messages to all users or all course admins staff such message would be displayed until it is dismissed or until we disable it this could be used to announce new releases scheduled downtime etc we want to be able to enable disable this using an interface on dodona and not doing a deploy | 1 |
103,346 | 4,167,530,056 | IssuesEvent | 2016-06-20 09:52:19 | GreatEmerald/RGIC1601 | https://api.github.com/repos/GreatEmerald/RGIC1601 | closed | Implement a metadata chain | Priority 4: Medium | Starting from the ClassifyToZones module, operations that result in new rasters should contain metadata about the parameters used in creating the output. The way to do it, for raster objects, when `input` is an input raster, `output` is the output raster, and you want to add `newvar` set to `newvalue`, is:
```r
metadata(output) = append(metadata(input), list(newvar="newvalue", newvar2="newvalue2"))
```
This metadata should be written to files in the end (put into the Spatial*DataFrame before sending to the ExportToFile module). | 1.0 | Implement a metadata chain - Starting from the ClassifyToZones module, operations that result in new rasters should contain metadata about the parameters used in creating the output. The way to do it, for raster objects, when `input` is an input raster, `output` is the output raster, and you want to add `newvar` set to `newvalue`, is:
```r
metadata(output) = append(metadata(input), list(newvar="newvalue", newvar2="newvalue2"))
```
This metadata should be written to files in the end (put into the Spatial*DataFrame before sending to the ExportToFile module). | priority | implement a metadata chain starting from the classifytozones module operations that result in new rasters should contain metadata about the parameters used in creating the output the way to do it for raster objects when input is an input raster output is the output raster and you want to add newvar set to newvalue is r metadata output append metadata input list newvar newvalue this metadata should be written to files in the end put into the spatial dataframe before sending to the exporttofile module | 1 |
260,749 | 8,214,348,435 | IssuesEvent | 2018-09-04 22:54:20 | wevote/WebApp | https://api.github.com/repos/wevote/WebApp | closed | Force /src/index.html to reload every 12 hours | Difficulty: Medium Priority: 1 | ### Please describe the issue (What happens? What do you expect?)
Add code that sets an expiration to the /src/index.html so that the page has to reload every 12 hours. With a new index.html page, the name of the bundle will change if there has been an update, which will force the download of the latest Javascript bundle. | 1.0 | Force /src/index.html to reload every 12 hours - ### Please describe the issue (What happens? What do you expect?)
Add code that sets an expiration to the /src/index.html so that the page has to reload every 12 hours. With a new index.html page, the name of the bundle will change if there has been an update, which will force the download of the latest Javascript bundle. | priority | force src index html to reload every hours please describe the issue what happens what do you expect add code that sets an expiration to the src index html so that the page has to reload every hours with a new index html page the name of the bundle will change if there has been an update which will force the download of the latest javascript bundle | 1 |
116,114 | 4,696,970,118 | IssuesEvent | 2016-10-12 07:35:54 | PowerlineApp/powerline-mobile | https://api.github.com/repos/PowerlineApp/powerline-mobile | closed | Notification Action Button: Fundraiser/Crowdfunder "Donate" should load Item Detail Screen | bug P2 - Medium Priority | Currently shows newsfeed. Should show item detail screen for the fundraiser. | 1.0 | Notification Action Button: Fundraiser/Crowdfunder "Donate" should load Item Detail Screen - Currently shows newsfeed. Should show item detail screen for the fundraiser. | priority | notification action button fundraiser crowdfunder donate should load item detail screen currently shows newsfeed should show item detail screen for the fundraiser | 1 |
20,342 | 2,622,843,561 | IssuesEvent | 2015-03-04 08:01:25 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | closed | Optionally allow hiding of the removals ("red parts") in a diff | auto-migrated Priority-Medium Type-Enhancement | ```
When viewing a diff, text that has been removed is marked in red and new text
is marked in green. This is nice, but distorts the web page structure and
makes it more difficult to read the diff (we are interested in "what is new?"
more than "what has changed?").
As an option (and maybe by default) diffs should be shown without the text
that has been removed (the parts currently in red).
```
Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Mar 2010 at 7:30 | 1.0 | Optionally allow hiding of the removals ("red parts") in a diff - ```
When viewing a diff, text that has been removed is marked in red and new text
is marked in green. This is nice, but distorts the web page structure and
makes it more difficult to read the diff (we are interested in "what is new?"
more than "what has changed?").
As an option (and maybe by default) diffs should be shown without the text
that has been removed (the parts currently in red).
```
Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Mar 2010 at 7:30 | priority | optionally allow hiding of the removals red parts in a diff when viewing a diff text that has been removed is marked in red and new text is marked in green this is nice but distorts the web page structure and makes it more difficult to read the diff we are interested in what is new more than what has changed as an option and maybe by default diffs should be shown without the text that has been removed the parts currently in red original issue reported on code google com by gkaemp gmail com on mar at | 1 |
499,462 | 14,447,684,186 | IssuesEvent | 2020-12-08 04:26:13 | JuezUN/INGInious | https://api.github.com/repos/JuezUN/INGInious | closed | Changes on subproblems tab | Change request Frontend Medium Priority Plugins Task | - [x] Allow languages in multilang depending on grading environment, that is, allowing VHDL and Verilog for VHDL environment, python for Data Science and the others for multilang.
- [x] Restrict the creation of subproblems to maximum 1.
- [x] Show multiple languages in alphabetic order. | 1.0 | Changes on subproblems tab - - [x] Allow languages in multilang depending on grading environment, that is, allowing VHDL and Verilog for VHDL environment, python for Data Science and the others for multilang.
- [x] Restrict the creation of subproblems to maximum 1.
- [x] Show multiple languages in alphabetic order. | priority | changes on subproblems tab allow languages in multilang depending on grading environment that is allowing vhdl and verilog for vhdl environment python for data science and the others for multilang restrict the creation of subproblems to maximum show multiple languages in alphabetic order | 1 |
29,497 | 2,716,242,341 | IssuesEvent | 2015-04-10 17:53:12 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | The event BeforeRowEdit does not allow access to the widget responsible for row editing | bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.2.0 | _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 17:12:55_
The event BeforeRowEdit does not allow access to the widget responsible for row editing
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=419_ | 1.0 | The event BeforeRowEdit does not allow access to the widget responsible for row editing - _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 17:12:55_
The event BeforeRowEdit does not allow access to the widget responsible for row editing
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=419_ | priority | the event beforerowedit does not allow access to the widget responsible for row editing from on july the event beforerowedit does not allow access to the widget responsible for row editing original issue | 1 |
70,203 | 3,321,165,151 | IssuesEvent | 2015-11-09 06:43:55 | ekux44/LampShade | https://api.github.com/repos/ekux44/LampShade | closed | java.lang.NullPointerException in com.kuxhausen.huemore.MainFragment.onStateChanged | crash medium priority | LampShade version: 79, 82
Android version: 5.0, 5.1, 6.0
Devices: OnePlus One (A0001), Nexus 6 (shamu), Nexus 5 (hammerhead), Nexus 4 (mako), Galaxy S6 Edge (zeroltebmc)
java.lang.NullPointerException: Attempt to invoke virtual method 'com.kuxhausen.huemore.net.DeviceManager com.kuxhausen.huemore.net.ConnectivityService.getDeviceManager()' on a null object reference
at com.kuxhausen.huemore.MainFragment.onStateChanged(MainFragment.java:263)
at com.kuxhausen.huemore.net.DeviceManager.onStateChanged(DeviceManager.java:186)
at com.kuxhausen.huemore.net.hue.HubConnection.reportStateChangeSucess(HubConnection.java:288)
at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:22)
at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:7)
at com.kuxhausen.huemore.net.hue.api.GsonRequest.deliverResponse(GsonRequest.java:83)
at com.android.volley.ExecutorDelivery$ResponseDeliveryRunnable.run(ExecutorDelivery.java:99)
at android.os.Handler.handleCallback(Handler.java:739)
at android.os.Handler.dispatchMessage(Handler.java:95)
at android.os.Looper.loop(Looper.java:148)
at android.app.ActivityThread.main(ActivityThread.java:5417)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616)
| 1.0 | java.lang.NullPointerException in com.kuxhausen.huemore.MainFragment.onStateChanged - LampShade version: 79, 82
Android version: 5.0, 5.1, 6.0
Devices: OnePlus One (A0001), Nexus 6 (shamu), Nexus 5 (hammerhead), Nexus 4 (mako), Galaxy S6 Edge (zeroltebmc)
java.lang.NullPointerException: Attempt to invoke virtual method 'com.kuxhausen.huemore.net.DeviceManager com.kuxhausen.huemore.net.ConnectivityService.getDeviceManager()' on a null object reference
at com.kuxhausen.huemore.MainFragment.onStateChanged(MainFragment.java:263)
at com.kuxhausen.huemore.net.DeviceManager.onStateChanged(DeviceManager.java:186)
at com.kuxhausen.huemore.net.hue.HubConnection.reportStateChangeSucess(HubConnection.java:288)
at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:22)
at com.kuxhausen.huemore.net.hue.api.StateSuccessListener.onResponse(StateSuccessListener.java:7)
at com.kuxhausen.huemore.net.hue.api.GsonRequest.deliverResponse(GsonRequest.java:83)
at com.android.volley.ExecutorDelivery$ResponseDeliveryRunnable.run(ExecutorDelivery.java:99)
at android.os.Handler.handleCallback(Handler.java:739)
at android.os.Handler.dispatchMessage(Handler.java:95)
at android.os.Looper.loop(Looper.java:148)
at android.app.ActivityThread.main(ActivityThread.java:5417)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616)
| priority | java lang nullpointerexception in com kuxhausen huemore mainfragment onstatechanged lampshade version android version devices oneplus one nexus shamu nexus hammerhead nexus mako galaxy edge zeroltebmc java lang nullpointerexception attempt to invoke virtual method com kuxhausen huemore net devicemanager com kuxhausen huemore net connectivityservice getdevicemanager on a null object reference at com kuxhausen huemore mainfragment onstatechanged mainfragment java at com kuxhausen huemore net devicemanager onstatechanged devicemanager java at com kuxhausen huemore net hue hubconnection reportstatechangesucess hubconnection java at com kuxhausen huemore net hue api statesuccesslistener onresponse statesuccesslistener java at com kuxhausen huemore net hue api statesuccesslistener onresponse statesuccesslistener java at com kuxhausen huemore net hue api gsonrequest deliverresponse gsonrequest java at com android volley executordelivery responsedeliveryrunnable run executordelivery java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java | 1 |
268,788 | 8,414,496,734 | IssuesEvent | 2018-10-13 03:10:43 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] Route docs are not generated | 1. Type: Bug Context : routing Priority: Medium Status: Accepted Time: ~1 hour | <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
Routes information is not populated in generated docs.
Used [angular-started](https://github.com/gdi2290/angular-starter) as example for testing.
[Generated docs screenshot](http://prntscr.com/k64f0e)
##### **Operating System, Node.js, npm, compodoc version(s)**
* **Node**: `v9.11.1`
* **OS**: macOS 10.13, Windows 10
* **Compodoc**: `1.1.3`
##### **Angular configuration, a `package.json` file in the root folder**
Use [angular-starter](https://github.com/gdi2290/angular-starter) sources.
##### **Compodoc installed globally or locally ?**
Both
##### **If possible sourcecode of the file where it breaks**
Use [angular-starter](https://github.com/gdi2290/angular-starter) sources.
##### **Reproduce the error**
1. Clone [angular-starter](https://github.com/gdi2290/angular-starter).
2. Run `npm install` or `yarn install`
3. Generate docs using compodoc: `npm run docs:compodoc`
4. View docs: `npm run docs:compodoc:generate`
No route section generated
##### **Related issues**
https://github.com/compodoc/compodoc/issues/257
| 1.0 | [BUG] Route docs are not generated - <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
Routes information is not populated in generated docs.
Used [angular-started](https://github.com/gdi2290/angular-starter) as example for testing.
[Generated docs screenshot](http://prntscr.com/k64f0e)
##### **Operating System, Node.js, npm, compodoc version(s)**
* **Node**: `v9.11.1`
* **OS**: macOS 10.13, Windows 10
* **Compodoc**: `1.1.3`
##### **Angular configuration, a `package.json` file in the root folder**
Use [angular-starter](https://github.com/gdi2290/angular-starter) sources.
##### **Compodoc installed globally or locally ?**
Both
##### **If possible sourcecode of the file where it breaks**
Use [angular-starter](https://github.com/gdi2290/angular-starter) sources.
##### **Reproduce the error**
1. Clone [angular-starter](https://github.com/gdi2290/angular-starter).
2. Run `npm install` or `yarn install`
3. Generate docs using compodoc: `npm run docs:compodoc`
4. View docs: `npm run docs:compodoc:generate`
No route section generated
##### **Related issues**
https://github.com/compodoc/compodoc/issues/257
| priority | route docs are not generated please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue routes information is not populated in generated docs used as example for testing operating system node js npm compodoc version s node os macos windows compodoc angular configuration a package json file in the root folder use sources compodoc installed globally or locally both if possible sourcecode of the file where it breaks use sources reproduce the error clone run npm install or yarn install generate docs using compodoc npm run docs compodoc view docs npm run docs compodoc generate no route section generated related issues | 1 |
296,737 | 9,125,653,180 | IssuesEvent | 2019-02-24 15:31:39 | pixijs/pixi.js | https://api.github.com/repos/pixijs/pixi.js | closed | Prefiltered line technique to dramatically improve the rendering of lines an shapes | Difficulty: Medium Domain: API Plugin: Graphics Priority: Low Resolution: Won't Fix Status: Accepting PRs Type: Feature Request Version: v5.x | My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA)
I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI.
http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
| 1.0 | Prefiltered line technique to dramatically improve the rendering of lines an shapes - My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA)
I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI.
http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
| priority | prefiltered line technique to dramatically improve the rendering of lines an shapes my understanding is that currently pixi uses the stencil buffer to draw graphics objects like lines and shapes since the stencil buffer does not support alpha all lines are very jagged aliased this is fine with aa or fxaa on but both of these come with other sets of issues that may or may not be acceptable for a given project gpu browser support performance for aa and text rendering issues for fxaa i ve stumbled across a different technique which appears to complete solve this issue and if my understanding is right it should be faster than the current method i thought i would share the link in case it is something you guys might want to look into for a future version of pixi | 1 |
231,650 | 7,641,700,393 | IssuesEvent | 2018-05-08 06:30:12 | vmware/harbor | https://api.github.com/repos/vmware/harbor | closed | Sometimes label filter does not close | area/ui priority/medium | Harbor version: v1.5 rc3
Go into project repo, open search filter, select a label, then click refresh button on the right, click search filter to close search filter, the label filter does not close

| 1.0 | Sometimes label filter does not close - Harbor version: v1.5 rc3
Go into project repo, open search filter, select a label, then click refresh button on the right, click search filter to close search filter, the label filter does not close

| priority | sometimes label filter does not close harbor version go into project repo open search filter select a label then click refresh button on the right click search filter to close search filter the label filter does not close | 1 |
704,162 | 24,187,698,110 | IssuesEvent | 2022-09-23 14:37:46 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | cross-compile toolchain variant doesn't working properly with multilib toolchain | bug priority: medium area: Toolchains | **Describe the bug**
In case of `ZEPHYR_TOOLCHAIN_VARIANT=cross-compile` we don't set `SYSROOT_DIR` cmake variable, so we don't execute
block of code which sets `LIBC_LIBRARY_DIR` variable based on the toolchain C flags (including mcpu flag):
https://github.com/zephyrproject-rtos/zephyr/blob/88aa873fbc544b99c1e48e37d5c74662f90b3de0/cmake/compiler/gcc/target.cmake#L74-L88
So, in case of multilib cross-compile toolchain we always use default libraries set no mater which mcpu value we've provided. Obviously it's incorrect.
**Impact**
Implicit & unexpected usage of libraries which are build for different mcpu than was specified.
If we are lucky we may get compile error - for example if we got something not linkable, like 64bit libraries from toolchain and 32 bit object files from rest of zephyr. However we may get successful linkage and use some not optimal libraries (i.e built for processor which use smaller instruction set) or get runtime issues.
**possible solutions**
* add additional environment variable for cross-compile toolchain variant, which will be pointing to correct SYSROOT_DIR path.
* try to extract SYSROOT_DIR path from the `CROSS_COMPILE` variable. I've checked the value of `CROSS_COMPILE` I've previously used - it doesn't seems to be feasible for these toolchains I.e:
```
CROSS_COMPILE=/home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/bin/arc64-snps-elf-
SYSROOT_DIR should be /home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/arc64-unknown-elf/
CROSS_COMPILE=/global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/bin/arc-elf32-
SYSROOT_DIR should be /global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/arc-snps-elf/
```
| 1.0 | cross-compile toolchain variant doesn't working properly with multilib toolchain - **Describe the bug**
In case of `ZEPHYR_TOOLCHAIN_VARIANT=cross-compile` we don't set `SYSROOT_DIR` cmake variable, so we don't execute
block of code which sets `LIBC_LIBRARY_DIR` variable based on the toolchain C flags (including mcpu flag):
https://github.com/zephyrproject-rtos/zephyr/blob/88aa873fbc544b99c1e48e37d5c74662f90b3de0/cmake/compiler/gcc/target.cmake#L74-L88
So, in case of multilib cross-compile toolchain we always use default libraries set no mater which mcpu value we've provided. Obviously it's incorrect.
**Impact**
Implicit & unexpected usage of libraries which are build for different mcpu than was specified.
If we are lucky we may get compile error - for example if we got something not linkable, like 64bit libraries from toolchain and 32 bit object files from rest of zephyr. However we may get successful linkage and use some not optimal libraries (i.e built for processor which use smaller instruction set) or get runtime issues.
**possible solutions**
* add additional environment variable for cross-compile toolchain variant, which will be pointing to correct SYSROOT_DIR path.
* try to extract SYSROOT_DIR path from the `CROSS_COMPILE` variable. I've checked the value of `CROSS_COMPILE` I've previously used - it doesn't seems to be feasible for these toolchains I.e:
```
CROSS_COMPILE=/home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/bin/arc64-snps-elf-
SYSROOT_DIR should be /home/user/dev/arc64-unknown-elf-1205/arc64-unknown-elf/arc64-unknown-elf/
CROSS_COMPILE=/global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/bin/arc-elf32-
SYSROOT_DIR should be /global/tools/Linux/arc_gnu_2021.09_prebuilt_elf32_le/arc-snps-elf/
```
| priority | cross compile toolchain variant doesn t working properly with multilib toolchain describe the bug in case of zephyr toolchain variant cross compile we don t set sysroot dir cmake variable so we don t execute block of code which sets libc library dir variable based on the toolchain c flags including mcpu flag so in case of multilib cross compile toolchain we always use default libraries set no mater which mcpu value we ve provided obviously it s incorrect impact implicit unexpected usage of libraries which are build for different mcpu than was specified if we are lucky we may get compile error for example if we got something not linkable like libraries from toolchain and bit object files from rest of zephyr however we may get successful linkage and use some not optimal libraries i e built for processor which use smaller instruction set or get runtime issues possible solutions add additional environment variable for cross compile toolchain variant which will be pointing to correct sysroot dir path try to extract sysroot dir path from the cross compile variable i ve checked the value of cross compile i ve previously used it doesn t seems to be feasible for these toolchains i e cross compile home user dev unknown elf unknown elf bin snps elf sysroot dir should be home user dev unknown elf unknown elf unknown elf cross compile global tools linux arc gnu prebuilt le bin arc sysroot dir should be global tools linux arc gnu prebuilt le arc snps elf | 1 |
293,471 | 8,996,207,878 | IssuesEvent | 2019-02-02 00:04:08 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | USB Power Event Panic | area: USB bug priority: medium | **Describe the bug**
A panic appears to occur on a USB power event.
**To Reproduce**
As best as I can tell, I've had a tough time reproducing this consistently, but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj.conf seems to cause the problem sometimes.
My prj.conf for the app that causes the fault.
```
CONFIG_FLOAT=y
CONFIG_COMPILER_OPT="-ffast-math"
# GPIO Enable
CONFIG_GPIO=y
# Logging Config
CONFIG_LOG=y
CONFIG_LOG_PRINTK=y
CONFIG_LOG_BACKEND_RTT=y
CONFIG_ISR_STACK_SIZE=4096
# Workqueue stack needs to be larger for some tasks
CONFIG_SYSTEM_WORKQUEUE_STACK_SIZE=8192
## USB
CONFIG_USB=y
CONFIG_USB_DEVICE_STACK=y
CONFIG_USB_DEVICE_PRODUCT="The Boss"
CONFIG_USB_UART_CONSOLE=y
CONFIG_UART_INTERRUPT_DRIVEN=y
CONFIG_UART_LINE_CTRL=y
## RTT and SystemView
CONFIG_USE_SEGGER_RTT=n
CONFIG_RTT_CONSOLE=y
CONFIG_SEGGER_SYSTEMVIEW=n
# Debug help
CONFIG_DEBUG=n
CONFIG_ASSERT=n
CONFIG_OUTPUT_DISASSEMBLY=y
CONFIG_STACK_CANARIES=n
CONFIG_STACK_SENTINEL=n
CONFIG_DEBUG_INFO=y
CONFIG_INIT_STACKS=n
CONFIG_STACK_USAGE=n
# Enable mcumgr.
CONFIG_MCUMGR=y
CONFIG_MCUMGR_SMP_BT=y
#CONFIG_MCUMGR_SMP_SHELL=n
#CONFIG_MCUMGR_SMP_UART=y
CONFIG_MCUMGR_CMD_FS_MGMT=y
CONFIG_MCUMGR_CMD_IMG_MGMT=y
CONFIG_MCUMGR_CMD_OS_MGMT=y
CONFIG_MCUMGR_CMD_STAT_MGMT=y
# Ensure an MCUboot-compatible binary is generated.
CONFIG_BOOTLOADER_MCUBOOT=n
# Setup watchdog support
CONFIG_WATCHDOG=n
# Enable ADC for battery level monitoring
CONFIG_ADC=y
CONFIG_ADC_0=y
CONFIG_ADC_NRFX_SAADC=y
#CONFIG_ADC_NRFX_ADC_CHANNEL_COUNT=1
# Setup flash access and file system settings
CONFIG_FLASH=y
CONFIG_FLASH_PAGE_LAYOUT=y
CONFIG_SOC_FLASH_NRF_RADIO_SYNC=y
CONFIG_MPU_ALLOW_FLASH_WRITE=y
CONFIG_FILE_SYSTEM=y
CONFIG_FILE_SYSTEM_NFFS=y
CONFIG_FS_NFFS_FLASH_DEV_NAME="NRF_FLASH_DRV_NAME"
CONFIG_FS_NFFS_NUM_FILES=4
CONFIG_FS_NFFS_NUM_DIRS=4
CONFIG_FS_NFFS_NUM_INODES=1024
CONFIG_FS_NFFS_NUM_BLOCKS=1024
CONFIG_FS_NFFS_NUM_CACHE_INODES=1
CONFIG_FS_NFFS_NUM_CACHE_BLOCKS=1
CONFIG_FILE_SYSTEM_NFFS=y
CONFIG_NFFS_FILESYSTEM_MAX_AREAS=12
# Settings
CONFIG_SETTINGS=y
CONFIG_SETTINGS_FS_DIR="/nffs/settings"
CONFIG_SETTINGS_FS_FILE="/nffs/settings/run"
# Bluetooth Config
CONFIG_BT=y
#CONFIG_BT_PERIPHERAL=y
#CONFIG_BT_DEBUG_LOG=n
#CONFIG_BT_DEBUG_CONN=y
#CONFIG_BT_DEBUG_SETTINGS=y
#CONFIG_BT_DEBUG_GATT=y
#CONFIG_BT_DEBUG_ATT=y
#CONFIG_BT_DEBUG_L2CAP=y
#CONFIG_BT_DEBUG_HCI_CORE=y
#CONFIG_BT_DEBUG_HCI_DRIVER=y
#CONFIG_BT_DEBUG_MONITOR=n # dumps protocol on UART
#CONFIG_BT_CTLR_PROFILE_ISR=y
CONFIG_BT_DEVICE_NAME="My Device"
#CONFIG_BT_DEVICE_NAME_DYNAMIC=y
#CONFIG_BT_DEVICE_NAME_GATT_WRITABLE=y
#CONFIG_BT_DEVICE_NAME_MAX=64
CONFIG_BT_RX_BUF_COUNT=4
CONFIG_BT_RX_BUF_LEN=251
CONFIG_BT_RX_STACK_SIZE=2048
#CONFIG_BT_SETTINGS=n
CONFIG_BT_L2CAP_TX_BUF_COUNT=10
#CONFIG_BT_L2CAP_TX_FRAG_COUNT=4
CONFIG_BT_L2CAP_TX_MTU=251
#CONFIG_BT_CONN_TX_MAX=10
#CONFIG_BT_ATT_PREPARE_COUNT=10
CONFIG_BT_ATT_TX_MAX=10
#CONFIG_BT_SMP=n
#CONFIG_BT_PRIVACY=n
#CONFIG_BT_SIGNING=n
#CONFIG_BT_CTLR=y
CONFIG_BT_CTLR_TX_BUFFERS=10
CONFIG_BT_CTLR_RX_BUFFERS=4
CONFIG_BT_CTLR_TX_BUFFER_SIZE=251
CONFIG_BT_CTLR_DATA_LENGTH_MAX=251
CONFIG_BT_CTLR_ADVANCED_FEATURES=y
CONFIG_BT_CTLR_WORKER_PRIO=0
CONFIG_BT_CTLR_JOB_PRIO=0
CONFIG_BT_CTLR_PHY_2M=y #toggle suppoort of 2M phy
CONFIG_BT_CTLR_CONN_RSSI=y
CONFIG_BT_CTLR_SCAN_REQ_RSSI=y
# Bluetooth support of mcumgr requires a net_buf user_data size >= 7.
CONFIG_NET_BUF_USER_DATA_SIZE=16
# Required by the `taskstat` command.
CONFIG_THREAD_MONITOR=y
# Enable statistics and statistic names.
CONFIG_STATS=y
CONFIG_STATS_NAMES=y
CONFIG_I2C=y
CONFIG_I2C_NRFX=y
CONFIG_I2C_0=y
CONFIG_I2C_0_NRF_TWIM=y
CONFIG_SPI=y
CONFIG_SPI_NRFX=y
CONFIG_SPI_1=y
CONFIG_SPI_1_NRF_SPIM=y
CONFIG_SENSOR=y
# Magnetometer Sensor
CONFIG_LIS3MDL=y
CONFIG_LIS3MDL_TRIGGER=y
CONFIG_LIS3MDL_THREAD_PRIORITY=0
CONFIG_LIS3MDL_TRIGGER_OWN_THREAD=y
CONFIG_LIS3MDL_TRIGGER_IRQ=n
CONFIG_LIS3MDL_TRIGGER_GLOBAL_THREAD=n
CONFIG_LIS3MDL_DATA_RAW=y
CONFIG_LIS3MDL_DATA_CONVERTED=n
CONFIG_LIS3MDL_ODR="1000"
CONFIG_LIS3MDL_FS=4
# LP5521 LED controller
CONFIG_LED=y
CONFIG_LP5521=y
```
**Expected behavior**
No panics!
**Impact**
This prevents my application from working with mcuboot, which prevents OTA updates.
**Screenshots or console output**
GDB output as I was debugging this problem
```
(gdb) break usb_dc_nrfx_power_event_callback
Breakpoint 1 at 0x14a60: file /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c, line 477.
(gdb) break usb_enable
Breakpoint 2 at 0x126a4: file /home/tburdick/src/zephyr/subsys/usb/usb_device.c, line 999.
(gdb) monitor reset 2
mResets core & peripherals using RESET pin.
(gdb) monitor reset
Resetting target
(gdb) c
Continuing.
Breakpoint 1, usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:477
477 switch (event) {
(gdb) print usbd_ctx
$1 = {status_cb = 0x0, attached = false, ready = false, usb_work = {_reserved = 0x0, handler = 0x0, flags = {0}}, work_queue = {_queue = {data_q = {head = 0x0, tail = 0x0}, {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, poll_events = {{head = 0x0,
next = 0x0}, {tail = 0x0, prev = 0x0}}}}}, drv_lock = {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, owner = 0x0, lock_count = 0, owner_orig_prio = 0}, ep_ctx = {{cfg = {cb = 0x0, max_sz = 0, en = false, addr = 0 '\000',
type = USB_DC_EP_CONTROL}, buf = {len = 0, block = {data = 0x0, id = {pool = 0, level = 0, block = 0}}, data = 0x0, curr = 0x0}, read_complete = false, read_pending = false, write_in_progress = false, write_fragmented = false} <repeats 18 times>}, ctrl_read_len = 0}
(gdb) break _Sys
_SysFatalErrorHandler _SysNmiOnReset
(gdb) break _SysFatalErrorHandler
Breakpoint 3 at 0x156ac: file /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c, line 52.
(gdb) c
Continuing.
Breakpoint 3, _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52
52 if (reason == _NANO_ERR_KERNEL_PANIC) {
(gdb) bt
#0 _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52
#1 0x00015636 in __usage_fault () at /home/tburdick/src/zephyr/arch/arm/core/fault_s.S:143
#2 <signal handler called>
#3 signal_poll_event (event=0x0, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:342
#4 0x000309a4 in _handle_obj_poll_events (events=events@entry=0x20027164 <usbd_ctx+28>, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:371
#5 0x0002f790 in handle_poll_events (state=4, queue=0x2002715c <usbd_ctx+20>) at /home/tburdick/src/zephyr/kernel/queue.c:119
#6 queue_insert (queue=queue@entry=0x2002715c <usbd_ctx+20>, prev=0x0, data=<optimized out>, alloc=alloc@entry=false) at /home/tburdick/src/zephyr/kernel/queue.c:179
#7 0x0002f7ee in k_queue_append (queue=queue@entry=0x2002715c <usbd_ctx+20>, data=<optimized out>) at /home/tburdick/src/zephyr/kernel/queue.c:193
#8 0x000141ac in usbd_evt_put (ev=<optimized out>) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:395
#9 0x00014a88 in usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:501
#10 0x00013a56 in power_event_cb (event=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:275
#11 _power_clock_isr (arg=0x0) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:415
#12 0x0001577a in _isr_wrapper () at /home/tburdick/src/zephyr/arch/arm/core/isr_wrapper.S:120
#13 <signal handler called>
#14 0x17fb7d56 in ?? ()
#15 0x28311824 in ?? ()
Backtrace stopped: previous frame identical to this frame (corrupt stack?)
(gdb)
```
**Environment (please complete the following information):**
- gcc 6.2.1
**Additional context**
It appears that sometimes a usb power event occurs before the usbd_ctx has been initialized from usb_dc_attach. I think that is the cause of the problem, because then the usb_work struct is likely to be uninitialized.
I'd guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an IRQ, this IRQ must be occuring before the subsequent usb_enable is called.
Why that happens beyond my current understanding of zephyr and the usb device driver.
| 1.0 | USB Power Event Panic - **Describe the bug**
A panic appears to occur on a USB power event.
**To Reproduce**
As best as I can tell, I've had a tough time reproducing this consistently, but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj.conf seems to cause the problem sometimes.
My prj.conf for the app that causes the fault.
```
CONFIG_FLOAT=y
CONFIG_COMPILER_OPT="-ffast-math"
# GPIO Enable
CONFIG_GPIO=y
# Logging Config
CONFIG_LOG=y
CONFIG_LOG_PRINTK=y
CONFIG_LOG_BACKEND_RTT=y
CONFIG_ISR_STACK_SIZE=4096
# Workqueue stack needs to be larger for some tasks
CONFIG_SYSTEM_WORKQUEUE_STACK_SIZE=8192
## USB
CONFIG_USB=y
CONFIG_USB_DEVICE_STACK=y
CONFIG_USB_DEVICE_PRODUCT="The Boss"
CONFIG_USB_UART_CONSOLE=y
CONFIG_UART_INTERRUPT_DRIVEN=y
CONFIG_UART_LINE_CTRL=y
## RTT and SystemView
CONFIG_USE_SEGGER_RTT=n
CONFIG_RTT_CONSOLE=y
CONFIG_SEGGER_SYSTEMVIEW=n
# Debug help
CONFIG_DEBUG=n
CONFIG_ASSERT=n
CONFIG_OUTPUT_DISASSEMBLY=y
CONFIG_STACK_CANARIES=n
CONFIG_STACK_SENTINEL=n
CONFIG_DEBUG_INFO=y
CONFIG_INIT_STACKS=n
CONFIG_STACK_USAGE=n
# Enable mcumgr.
CONFIG_MCUMGR=y
CONFIG_MCUMGR_SMP_BT=y
#CONFIG_MCUMGR_SMP_SHELL=n
#CONFIG_MCUMGR_SMP_UART=y
CONFIG_MCUMGR_CMD_FS_MGMT=y
CONFIG_MCUMGR_CMD_IMG_MGMT=y
CONFIG_MCUMGR_CMD_OS_MGMT=y
CONFIG_MCUMGR_CMD_STAT_MGMT=y
# Ensure an MCUboot-compatible binary is generated.
CONFIG_BOOTLOADER_MCUBOOT=n
# Setup watchdog support
CONFIG_WATCHDOG=n
# Enable ADC for battery level monitoring
CONFIG_ADC=y
CONFIG_ADC_0=y
CONFIG_ADC_NRFX_SAADC=y
#CONFIG_ADC_NRFX_ADC_CHANNEL_COUNT=1
# Setup flash access and file system settings
CONFIG_FLASH=y
CONFIG_FLASH_PAGE_LAYOUT=y
CONFIG_SOC_FLASH_NRF_RADIO_SYNC=y
CONFIG_MPU_ALLOW_FLASH_WRITE=y
CONFIG_FILE_SYSTEM=y
CONFIG_FILE_SYSTEM_NFFS=y
CONFIG_FS_NFFS_FLASH_DEV_NAME="NRF_FLASH_DRV_NAME"
CONFIG_FS_NFFS_NUM_FILES=4
CONFIG_FS_NFFS_NUM_DIRS=4
CONFIG_FS_NFFS_NUM_INODES=1024
CONFIG_FS_NFFS_NUM_BLOCKS=1024
CONFIG_FS_NFFS_NUM_CACHE_INODES=1
CONFIG_FS_NFFS_NUM_CACHE_BLOCKS=1
CONFIG_FILE_SYSTEM_NFFS=y
CONFIG_NFFS_FILESYSTEM_MAX_AREAS=12
# Settings
CONFIG_SETTINGS=y
CONFIG_SETTINGS_FS_DIR="/nffs/settings"
CONFIG_SETTINGS_FS_FILE="/nffs/settings/run"
# Bluetooth Config
CONFIG_BT=y
#CONFIG_BT_PERIPHERAL=y
#CONFIG_BT_DEBUG_LOG=n
#CONFIG_BT_DEBUG_CONN=y
#CONFIG_BT_DEBUG_SETTINGS=y
#CONFIG_BT_DEBUG_GATT=y
#CONFIG_BT_DEBUG_ATT=y
#CONFIG_BT_DEBUG_L2CAP=y
#CONFIG_BT_DEBUG_HCI_CORE=y
#CONFIG_BT_DEBUG_HCI_DRIVER=y
#CONFIG_BT_DEBUG_MONITOR=n # dumps protocol on UART
#CONFIG_BT_CTLR_PROFILE_ISR=y
CONFIG_BT_DEVICE_NAME="My Device"
#CONFIG_BT_DEVICE_NAME_DYNAMIC=y
#CONFIG_BT_DEVICE_NAME_GATT_WRITABLE=y
#CONFIG_BT_DEVICE_NAME_MAX=64
CONFIG_BT_RX_BUF_COUNT=4
CONFIG_BT_RX_BUF_LEN=251
CONFIG_BT_RX_STACK_SIZE=2048
#CONFIG_BT_SETTINGS=n
CONFIG_BT_L2CAP_TX_BUF_COUNT=10
#CONFIG_BT_L2CAP_TX_FRAG_COUNT=4
CONFIG_BT_L2CAP_TX_MTU=251
#CONFIG_BT_CONN_TX_MAX=10
#CONFIG_BT_ATT_PREPARE_COUNT=10
CONFIG_BT_ATT_TX_MAX=10
#CONFIG_BT_SMP=n
#CONFIG_BT_PRIVACY=n
#CONFIG_BT_SIGNING=n
#CONFIG_BT_CTLR=y
CONFIG_BT_CTLR_TX_BUFFERS=10
CONFIG_BT_CTLR_RX_BUFFERS=4
CONFIG_BT_CTLR_TX_BUFFER_SIZE=251
CONFIG_BT_CTLR_DATA_LENGTH_MAX=251
CONFIG_BT_CTLR_ADVANCED_FEATURES=y
CONFIG_BT_CTLR_WORKER_PRIO=0
CONFIG_BT_CTLR_JOB_PRIO=0
CONFIG_BT_CTLR_PHY_2M=y #toggle suppoort of 2M phy
CONFIG_BT_CTLR_CONN_RSSI=y
CONFIG_BT_CTLR_SCAN_REQ_RSSI=y
# Bluetooth support of mcumgr requires a net_buf user_data size >= 7.
CONFIG_NET_BUF_USER_DATA_SIZE=16
# Required by the `taskstat` command.
CONFIG_THREAD_MONITOR=y
# Enable statistics and statistic names.
CONFIG_STATS=y
CONFIG_STATS_NAMES=y
CONFIG_I2C=y
CONFIG_I2C_NRFX=y
CONFIG_I2C_0=y
CONFIG_I2C_0_NRF_TWIM=y
CONFIG_SPI=y
CONFIG_SPI_NRFX=y
CONFIG_SPI_1=y
CONFIG_SPI_1_NRF_SPIM=y
CONFIG_SENSOR=y
# Magnetometer Sensor
CONFIG_LIS3MDL=y
CONFIG_LIS3MDL_TRIGGER=y
CONFIG_LIS3MDL_THREAD_PRIORITY=0
CONFIG_LIS3MDL_TRIGGER_OWN_THREAD=y
CONFIG_LIS3MDL_TRIGGER_IRQ=n
CONFIG_LIS3MDL_TRIGGER_GLOBAL_THREAD=n
CONFIG_LIS3MDL_DATA_RAW=y
CONFIG_LIS3MDL_DATA_CONVERTED=n
CONFIG_LIS3MDL_ODR="1000"
CONFIG_LIS3MDL_FS=4
# LP5521 LED controller
CONFIG_LED=y
CONFIG_LP5521=y
```
**Expected behavior**
No panics!
**Impact**
This prevents my application from working with mcuboot, which prevents OTA updates.
**Screenshots or console output**
GDB output as I was debugging this problem
```
(gdb) break usb_dc_nrfx_power_event_callback
Breakpoint 1 at 0x14a60: file /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c, line 477.
(gdb) break usb_enable
Breakpoint 2 at 0x126a4: file /home/tburdick/src/zephyr/subsys/usb/usb_device.c, line 999.
(gdb) monitor reset 2
mResets core & peripherals using RESET pin.
(gdb) monitor reset
Resetting target
(gdb) c
Continuing.
Breakpoint 1, usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:477
477 switch (event) {
(gdb) print usbd_ctx
$1 = {status_cb = 0x0, attached = false, ready = false, usb_work = {_reserved = 0x0, handler = 0x0, flags = {0}}, work_queue = {_queue = {data_q = {head = 0x0, tail = 0x0}, {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, poll_events = {{head = 0x0,
next = 0x0}, {tail = 0x0, prev = 0x0}}}}}, drv_lock = {wait_q = {waitq = {{head = 0x0, next = 0x0}, {tail = 0x0, prev = 0x0}}}, owner = 0x0, lock_count = 0, owner_orig_prio = 0}, ep_ctx = {{cfg = {cb = 0x0, max_sz = 0, en = false, addr = 0 '\000',
type = USB_DC_EP_CONTROL}, buf = {len = 0, block = {data = 0x0, id = {pool = 0, level = 0, block = 0}}, data = 0x0, curr = 0x0}, read_complete = false, read_pending = false, write_in_progress = false, write_fragmented = false} <repeats 18 times>}, ctrl_read_len = 0}
(gdb) break _Sys
_SysFatalErrorHandler _SysNmiOnReset
(gdb) break _SysFatalErrorHandler
Breakpoint 3 at 0x156ac: file /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c, line 52.
(gdb) c
Continuing.
Breakpoint 3, _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52
52 if (reason == _NANO_ERR_KERNEL_PANIC) {
(gdb) bt
#0 _SysFatalErrorHandler (reason=0, pEsf=0x200391f0 <_interrupt_stack+3960>) at /home/tburdick/src/zephyr/arch/arm/core/sys_fatal_error_handler.c:52
#1 0x00015636 in __usage_fault () at /home/tburdick/src/zephyr/arch/arm/core/fault_s.S:143
#2 <signal handler called>
#3 signal_poll_event (event=0x0, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:342
#4 0x000309a4 in _handle_obj_poll_events (events=events@entry=0x20027164 <usbd_ctx+28>, state=state@entry=4) at /home/tburdick/src/zephyr/kernel/poll.c:371
#5 0x0002f790 in handle_poll_events (state=4, queue=0x2002715c <usbd_ctx+20>) at /home/tburdick/src/zephyr/kernel/queue.c:119
#6 queue_insert (queue=queue@entry=0x2002715c <usbd_ctx+20>, prev=0x0, data=<optimized out>, alloc=alloc@entry=false) at /home/tburdick/src/zephyr/kernel/queue.c:179
#7 0x0002f7ee in k_queue_append (queue=queue@entry=0x2002715c <usbd_ctx+20>, data=<optimized out>) at /home/tburdick/src/zephyr/kernel/queue.c:193
#8 0x000141ac in usbd_evt_put (ev=<optimized out>) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:395
#9 0x00014a88 in usb_dc_nrfx_power_event_callback (event=event@entry=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/usb/device/usb_dc_nrfx.c:501
#10 0x00013a56 in power_event_cb (event=NRF_POWER_EVENT_USBDETECTED) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:275
#11 _power_clock_isr (arg=0x0) at /home/tburdick/src/zephyr/drivers/clock_control/nrf_power_clock.c:415
#12 0x0001577a in _isr_wrapper () at /home/tburdick/src/zephyr/arch/arm/core/isr_wrapper.S:120
#13 <signal handler called>
#14 0x17fb7d56 in ?? ()
#15 0x28311824 in ?? ()
Backtrace stopped: previous frame identical to this frame (corrupt stack?)
(gdb)
```
**Environment (please complete the following information):**
- gcc 6.2.1
**Additional context**
It appears that sometimes a usb power event occurs before the usbd_ctx has been initialized from usb_dc_attach. I think that is the cause of the problem, because then the usb_work struct is likely to be uninitialized.
I'd guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an IRQ, this IRQ must be occuring before the subsequent usb_enable is called.
Why that happens beyond my current understanding of zephyr and the usb device driver.
| priority | usb power event panic describe the bug a panic appears to occur on a usb power event to reproduce as best as i can tell i ve had a tough time reproducing this consistently but using mcuboot master with zephyr and an app with a variety of stuff enabled via prj conf seems to cause the problem sometimes my prj conf for the app that causes the fault config float y config compiler opt ffast math gpio enable config gpio y logging config config log y config log printk y config log backend rtt y config isr stack size workqueue stack needs to be larger for some tasks config system workqueue stack size usb config usb y config usb device stack y config usb device product the boss config usb uart console y config uart interrupt driven y config uart line ctrl y rtt and systemview config use segger rtt n config rtt console y config segger systemview n debug help config debug n config assert n config output disassembly y config stack canaries n config stack sentinel n config debug info y config init stacks n config stack usage n enable mcumgr config mcumgr y config mcumgr smp bt y config mcumgr smp shell n config mcumgr smp uart y config mcumgr cmd fs mgmt y config mcumgr cmd img mgmt y config mcumgr cmd os mgmt y config mcumgr cmd stat mgmt y ensure an mcuboot compatible binary is generated config bootloader mcuboot n setup watchdog support config watchdog n enable adc for battery level monitoring config adc y config adc y config adc nrfx saadc y config adc nrfx adc channel count setup flash access and file system settings config flash y config flash page layout y config soc flash nrf radio sync y config mpu allow flash write y config file system y config file system nffs y config fs nffs flash dev name nrf flash drv name config fs nffs num files config fs nffs num dirs config fs nffs num inodes config fs nffs num blocks config fs nffs num cache inodes config fs nffs num cache blocks config file system nffs y config nffs filesystem max areas settings config settings y config settings fs dir nffs settings config settings fs file nffs settings run bluetooth config config bt y config bt peripheral y config bt debug log n config bt debug conn y config bt debug settings y config bt debug gatt y config bt debug att y config bt debug y config bt debug hci core y config bt debug hci driver y config bt debug monitor n dumps protocol on uart config bt ctlr profile isr y config bt device name my device config bt device name dynamic y config bt device name gatt writable y config bt device name max config bt rx buf count config bt rx buf len config bt rx stack size config bt settings n config bt tx buf count config bt tx frag count config bt tx mtu config bt conn tx max config bt att prepare count config bt att tx max config bt smp n config bt privacy n config bt signing n config bt ctlr y config bt ctlr tx buffers config bt ctlr rx buffers config bt ctlr tx buffer size config bt ctlr data length max config bt ctlr advanced features y config bt ctlr worker prio config bt ctlr job prio config bt ctlr phy y toggle suppoort of phy config bt ctlr conn rssi y config bt ctlr scan req rssi y bluetooth support of mcumgr requires a net buf user data size config net buf user data size required by the taskstat command config thread monitor y enable statistics and statistic names config stats y config stats names y config y config nrfx y config y config nrf twim y config spi y config spi nrfx y config spi y config spi nrf spim y config sensor y magnetometer sensor config y config trigger y config thread priority config trigger own thread y config trigger irq n config trigger global thread n config data raw y config data converted n config odr config fs led controller config led y config y expected behavior no panics impact this prevents my application from working with mcuboot which prevents ota updates screenshots or console output gdb output as i was debugging this problem gdb break usb dc nrfx power event callback breakpoint at file home tburdick src zephyr drivers usb device usb dc nrfx c line gdb break usb enable breakpoint at file home tburdick src zephyr subsys usb usb device c line gdb monitor reset mresets core peripherals using reset pin gdb monitor reset resetting target gdb c continuing breakpoint usb dc nrfx power event callback event event entry nrf power event usbdetected at home tburdick src zephyr drivers usb device usb dc nrfx c switch event gdb print usbd ctx status cb attached false ready false usb work reserved handler flags work queue queue data q head tail wait q waitq head next tail prev poll events head next tail prev drv lock wait q waitq head next tail prev owner lock count owner orig prio ep ctx cfg cb max sz en false addr type usb dc ep control buf len block data id pool level block data curr read complete false read pending false write in progress false write fragmented false ctrl read len gdb break sys sysfatalerrorhandler sysnmionreset gdb break sysfatalerrorhandler breakpoint at file home tburdick src zephyr arch arm core sys fatal error handler c line gdb c continuing breakpoint sysfatalerrorhandler reason pesf at home tburdick src zephyr arch arm core sys fatal error handler c if reason nano err kernel panic gdb bt sysfatalerrorhandler reason pesf at home tburdick src zephyr arch arm core sys fatal error handler c in usage fault at home tburdick src zephyr arch arm core fault s s signal poll event event state state entry at home tburdick src zephyr kernel poll c in handle obj poll events events events entry state state entry at home tburdick src zephyr kernel poll c in handle poll events state queue at home tburdick src zephyr kernel queue c queue insert queue queue entry prev data alloc alloc entry false at home tburdick src zephyr kernel queue c in k queue append queue queue entry data at home tburdick src zephyr kernel queue c in usbd evt put ev at home tburdick src zephyr drivers usb device usb dc nrfx c in usb dc nrfx power event callback event event entry nrf power event usbdetected at home tburdick src zephyr drivers usb device usb dc nrfx c in power event cb event nrf power event usbdetected at home tburdick src zephyr drivers clock control nrf power clock c power clock isr arg at home tburdick src zephyr drivers clock control nrf power clock c in isr wrapper at home tburdick src zephyr arch arm core isr wrapper s in in backtrace stopped previous frame identical to this frame corrupt stack gdb environment please complete the following information gcc additional context it appears that sometimes a usb power event occurs before the usbd ctx has been initialized from usb dc attach i think that is the cause of the problem because then the usb work struct is likely to be uninitialized i d guess this happens because the clock device is initialized before the usb device where the callback is initially setup in an irq this irq must be occuring before the subsequent usb enable is called why that happens beyond my current understanding of zephyr and the usb device driver | 1 |
298,769 | 9,201,112,713 | IssuesEvent | 2019-03-07 18:46:54 | trimstray/htrace.sh | https://api.github.com/repos/trimstray/htrace.sh | closed | Wiki: updated 'Examples' chapter | Priority: Medium Status: Completed Type: Enhancement | Wiki [Examples](https://github.com/trimstray/htrace.sh/wiki/Examples) chapter.
- [x] - Check redirects
- [x] - POST method | 1.0 | Wiki: updated 'Examples' chapter - Wiki [Examples](https://github.com/trimstray/htrace.sh/wiki/Examples) chapter.
- [x] - Check redirects
- [x] - POST method | priority | wiki updated examples chapter wiki chapter check redirects post method | 1 |
261,381 | 8,230,606,048 | IssuesEvent | 2018-09-07 13:29:51 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | Work items created in Analytics Insights can't be opened in Chromium and Firefox. | SEV3-medium area/browser-compatibility priority/P2 status/in-progress team/analytics type/bug | The work item created in the Analytics Insights menu can't be opened, see blank screen and Chrome log: "Redirect ... was blocked by CORS policy...". And I can't see it in the Planner backlog.
Scenario:
1) Open 'View complete stack report' from 'Stack Report Recommendations'
2) Select 'Insights' and tab 'Companion Dependency Details'
3) Select a dependency and click 'Create work item'
4) Click 'View here' after the WI has been created
5) I got forwarded to a black page, see picture below
6) In addition the WI is not present in the Planner backlog
Tested on Chrome 56.0.2924.87 Fedora Project (64-bit) and Firefox.


| 1.0 | Work items created in Analytics Insights can't be opened in Chromium and Firefox. - The work item created in the Analytics Insights menu can't be opened, see blank screen and Chrome log: "Redirect ... was blocked by CORS policy...". And I can't see it in the Planner backlog.
Scenario:
1) Open 'View complete stack report' from 'Stack Report Recommendations'
2) Select 'Insights' and tab 'Companion Dependency Details'
3) Select a dependency and click 'Create work item'
4) Click 'View here' after the WI has been created
5) I got forwarded to a black page, see picture below
6) In addition the WI is not present in the Planner backlog
Tested on Chrome 56.0.2924.87 Fedora Project (64-bit) and Firefox.


| priority | work items created in analytics insights can t be opened in chromium and firefox the work item created in the analytics insights menu can t be opened see blank screen and chrome log redirect was blocked by cors policy and i can t see it in the planner backlog scenario open view complete stack report from stack report recommendations select insights and tab companion dependency details select a dependency and click create work item click view here after the wi has been created i got forwarded to a black page see picture below in addition the wi is not present in the planner backlog tested on chrome fedora project bit and firefox | 1 |
753,930 | 26,367,686,186 | IssuesEvent | 2023-01-11 17:52:05 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | yb-admin list_snapshots payload size limited | kind/bug area/docdb priority/medium | Jira Link: [DB-2362](https://yugabyte.atlassian.net/browse/DB-2362)
There is a hardcoded limit that effectively gates rpc_max_message_size at the protobuf bytes limit.
`yb-admin -rpc_max_message_size 1500000000 list_snapshots`
`tcmalloc: large alloc 1256480768 bytes == 0x2bac000 @ 0x7f8621f5c6d3 0x7f8621f7d09b 0x7f8629ae5e58 0x7f8629b63d28 0x7f8629aea522 0x7f8629b54cdc 0x7f8629b58a56 0x7f8629b595b6 0x7f8624d80b7b 0x7f8624d84c8a 0x7f8629b1587c 0x7f862639c3df 0x7f8620dcb694
[libprotobuf ERROR google/protobuf/io/coded_stream.cc:193] A protocol message was rejected because it was too big (more than 535822336 bytes). To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h.
Error: IO error (yb/rpc/outbound_call.cc:423): Unable to list snapshots: Invalid response, missing fields` | 1.0 | yb-admin list_snapshots payload size limited - Jira Link: [DB-2362](https://yugabyte.atlassian.net/browse/DB-2362)
There is a hardcoded limit that effectively gates rpc_max_message_size at the protobuf bytes limit.
`yb-admin -rpc_max_message_size 1500000000 list_snapshots`
`tcmalloc: large alloc 1256480768 bytes == 0x2bac000 @ 0x7f8621f5c6d3 0x7f8621f7d09b 0x7f8629ae5e58 0x7f8629b63d28 0x7f8629aea522 0x7f8629b54cdc 0x7f8629b58a56 0x7f8629b595b6 0x7f8624d80b7b 0x7f8624d84c8a 0x7f8629b1587c 0x7f862639c3df 0x7f8620dcb694
[libprotobuf ERROR google/protobuf/io/coded_stream.cc:193] A protocol message was rejected because it was too big (more than 535822336 bytes). To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h.
Error: IO error (yb/rpc/outbound_call.cc:423): Unable to list snapshots: Invalid response, missing fields` | priority | yb admin list snapshots payload size limited jira link there is a hardcoded limit that effectively gates rpc max message size at the protobuf bytes limit yb admin rpc max message size list snapshots tcmalloc large alloc bytes a protocol message was rejected because it was too big more than bytes to increase the limit or to disable these warnings see codedinputstream settotalbyteslimit in google protobuf io coded stream h error io error yb rpc outbound call cc unable to list snapshots invalid response missing fields | 1 |
55,590 | 3,073,804,213 | IssuesEvent | 2015-08-20 00:40:47 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | Unable to entertext in Webview page using robotium 4.2 | bug imported invalid Priority-Medium | _From [sinha.ka...@gmail.com](https://code.google.com/u/103828779966021760852/) on July 26, 2013 11:17:22_
package com.calculator.test;
import org.junit.Test;
import android.test.ActivityInstrumentationTestCase2;
import com.jayway.android.robotium.solo.*;
import com.jayway.android.robotium.solo.By;
import com.jayway.android.robotium.solo.Solo;
@SuppressWarnings("rawtypes")
public class SMPage extends ActivityInstrumentationTestCase2{
private static final String LAUNCHER_ACTIVITY_FULL_CLASSNAME = "com.aha.android.app.login.WelcomePageActivity";
private Solo solo;
private static Class launcherActivityClass;
static{
try {
launcherActivityClass = Class.forName(LAUNCHER_ACTIVITY_FULL_CLASSNAME);
} catch (ClassNotFoundException e) {
throw new RuntimeException(e);
}
}
@SuppressWarnings("unchecked")
public SMPage() {
super(launcherActivityClass);
}
@Override
public void setUp() throws Exception {
//setUp() is run before a test case is started.
//This is where the solo object is created.
solo = new Solo(getInstrumentation(), getActivity());
}
@Override
public void tearDown() throws Exception {
//tearDown() is run after a test case has finished.
//finishOpenedActivities() will finish all the activities that have been opened during the test execution.
solo.finishOpenedActivities();
}
@Test
public void testLaunch() throws Exception {
solo.clickOnText("Login");
solo.clickOnText("Forgot your password?");
Thread.sleep(50000);
solo.typeTextInWebElement(By.name("email"), "sinha.kanak@gmail.com");
solo.clickOnText("Submit");
Thread.sleep(5000
}
}
Not able to enter text in the webview page . Screen shot attached of the page. Using robotium 4.2.
**Attachment:** [password.pdf](http://code.google.com/p/robotium/issues/detail?id=493)
_Original issue: http://code.google.com/p/robotium/issues/detail?id=493_ | 1.0 | Unable to entertext in Webview page using robotium 4.2 - _From [sinha.ka...@gmail.com](https://code.google.com/u/103828779966021760852/) on July 26, 2013 11:17:22_
package com.calculator.test;
import org.junit.Test;
import android.test.ActivityInstrumentationTestCase2;
import com.jayway.android.robotium.solo.*;
import com.jayway.android.robotium.solo.By;
import com.jayway.android.robotium.solo.Solo;
@SuppressWarnings("rawtypes")
public class SMPage extends ActivityInstrumentationTestCase2{
private static final String LAUNCHER_ACTIVITY_FULL_CLASSNAME = "com.aha.android.app.login.WelcomePageActivity";
private Solo solo;
private static Class launcherActivityClass;
static{
try {
launcherActivityClass = Class.forName(LAUNCHER_ACTIVITY_FULL_CLASSNAME);
} catch (ClassNotFoundException e) {
throw new RuntimeException(e);
}
}
@SuppressWarnings("unchecked")
public SMPage() {
super(launcherActivityClass);
}
@Override
public void setUp() throws Exception {
//setUp() is run before a test case is started.
//This is where the solo object is created.
solo = new Solo(getInstrumentation(), getActivity());
}
@Override
public void tearDown() throws Exception {
//tearDown() is run after a test case has finished.
//finishOpenedActivities() will finish all the activities that have been opened during the test execution.
solo.finishOpenedActivities();
}
@Test
public void testLaunch() throws Exception {
solo.clickOnText("Login");
solo.clickOnText("Forgot your password?");
Thread.sleep(50000);
solo.typeTextInWebElement(By.name("email"), "sinha.kanak@gmail.com");
solo.clickOnText("Submit");
Thread.sleep(5000
}
}
Not able to enter text in the webview page . Screen shot attached of the page. Using robotium 4.2.
**Attachment:** [password.pdf](http://code.google.com/p/robotium/issues/detail?id=493)
_Original issue: http://code.google.com/p/robotium/issues/detail?id=493_ | priority | unable to entertext in webview page using robotium from on july package com calculator test import org junit test import android test import com jayway android robotium solo import com jayway android robotium solo by import com jayway android robotium solo solo suppresswarnings rawtypes public class smpage extends private static final string launcher activity full classname com aha android app login welcomepageactivity private solo solo private static class launcheractivityclass static try launcheractivityclass class forname launcher activity full classname catch classnotfoundexception e throw new runtimeexception e suppresswarnings unchecked public smpage super launcheractivityclass override public void setup throws exception setup is run before a test case is started this is where the solo object is created solo new solo getinstrumentation getactivity override public void teardown throws exception teardown is run after a test case has finished finishopenedactivities will finish all the activities that have been opened during the test execution solo finishopenedactivities test public void testlaunch throws exception solo clickontext login solo clickontext forgot your password thread sleep solo typetextinwebelement by name email sinha kanak gmail com solo clickontext submit thread sleep not able to enter text in the webview page screen shot attached of the page using robotium attachment original issue | 1 |
285,547 | 8,766,458,207 | IssuesEvent | 2018-12-17 16:52:14 | spacetelescope/jwql | https://api.github.com/repos/spacetelescope/jwql | closed | Build script execution status display on dashboard | Medium Priority Web Application enhancement | The dashboard on the web application should display the success/failure of the various `cron` jobs (e.g. `monitor_filesystem`) so that the JWQL team can easily identify scripts that are failing or have not executed when they should have.
Here is an example from what WFC3/Quicklook uses:
<img width="676" alt="screen shot 2018-09-17 at 16 05 56" src="https://user-images.githubusercontent.com/2250769/45647348-aa01be80-ba93-11e8-8ea1-2ffe314f4a8f.png">
Each row in the table corresponds to a particular `cron` job. A user can click on the name of the monitor to display the contents of the corresponding log file. The color indicates how recently the script executed successfully; bright green means that the script executed successfully that day, light green means the script executed successfully several days ago, red indicates that the script crashed. | 1.0 | Build script execution status display on dashboard - The dashboard on the web application should display the success/failure of the various `cron` jobs (e.g. `monitor_filesystem`) so that the JWQL team can easily identify scripts that are failing or have not executed when they should have.
Here is an example from what WFC3/Quicklook uses:
<img width="676" alt="screen shot 2018-09-17 at 16 05 56" src="https://user-images.githubusercontent.com/2250769/45647348-aa01be80-ba93-11e8-8ea1-2ffe314f4a8f.png">
Each row in the table corresponds to a particular `cron` job. A user can click on the name of the monitor to display the contents of the corresponding log file. The color indicates how recently the script executed successfully; bright green means that the script executed successfully that day, light green means the script executed successfully several days ago, red indicates that the script crashed. | priority | build script execution status display on dashboard the dashboard on the web application should display the success failure of the various cron jobs e g monitor filesystem so that the jwql team can easily identify scripts that are failing or have not executed when they should have here is an example from what quicklook uses img width alt screen shot at src each row in the table corresponds to a particular cron job a user can click on the name of the monitor to display the contents of the corresponding log file the color indicates how recently the script executed successfully bright green means that the script executed successfully that day light green means the script executed successfully several days ago red indicates that the script crashed | 1 |
140,312 | 5,400,079,574 | IssuesEvent | 2017-02-27 21:06:55 | canadainc/sunnah10 | https://api.github.com/repos/canadainc/sunnah10 | closed | Implement individuals table | Admin auto-migrated enhancement logic maintainability Priority-Medium ui | ```
Just like Quran10
```
Original issue reported on code.google.com by `canadai...@gmail.com` on 21 Feb 2015 at 5:09
| 1.0 | Implement individuals table - ```
Just like Quran10
```
Original issue reported on code.google.com by `canadai...@gmail.com` on 21 Feb 2015 at 5:09
| priority | implement individuals table just like original issue reported on code google com by canadai gmail com on feb at | 1 |
234,761 | 7,726,077,604 | IssuesEvent | 2018-05-24 20:02:47 | DarkPacks/SevTech-Ages | https://api.github.com/repos/DarkPacks/SevTech-Ages | closed | Lava crystal gets eaten inside ie alloy kilns | Category: Mod Priority: Medium Status: Review Needed Type: Bug | ## Issue / Bug
if you place a lava crystal inside the alloy kiln as a fuel source, the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp.
## Expected Behavior
When placing a lava crystal inside a fuel source it should drain 50lp and register it as 1 coal's worth of fuel, as stated in the blood magic's sanguine scientiem.



## Possible Solution
i have no idea how to solve it... i just know how it's supposed to work D=
## Steps to Reproduce
1. Place lava crystal in fuel slot in alloy kiln
2. Add in valid alloyable metals
3. watch lava crystal get eaten D=
## Client Information

* Modpack Version: 3.0.6
* Java Version: see image
* Launcher Used: twitch launcher
* Memory Allocated: see image
* Server/LAN/Single Player: single player
* Resourcepack Enabled?: no
* Optifine Installed?: no
| 1.0 | Lava crystal gets eaten inside ie alloy kilns - ## Issue / Bug
if you place a lava crystal inside the alloy kiln as a fuel source, the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp.
## Expected Behavior
When placing a lava crystal inside a fuel source it should drain 50lp and register it as 1 coal's worth of fuel, as stated in the blood magic's sanguine scientiem.



## Possible Solution
i have no idea how to solve it... i just know how it's supposed to work D=
## Steps to Reproduce
1. Place lava crystal in fuel slot in alloy kiln
2. Add in valid alloyable metals
3. watch lava crystal get eaten D=
## Client Information

* Modpack Version: 3.0.6
* Java Version: see image
* Launcher Used: twitch launcher
* Memory Allocated: see image
* Server/LAN/Single Player: single player
* Resourcepack Enabled?: no
* Optifine Installed?: no
| priority | lava crystal gets eaten inside ie alloy kilns issue bug if you place a lava crystal inside the alloy kiln as a fuel source the alloy kiln eats the lava crystal instead of behaving like the lava crystal should and draining lp expected behavior when placing a lava crystal inside a fuel source it should drain and register it as coal s worth of fuel as stated in the blood magic s sanguine scientiem possible solution i have no idea how to solve it i just know how it s supposed to work d steps to reproduce place lava crystal in fuel slot in alloy kiln add in valid alloyable metals watch lava crystal get eaten d client information modpack version java version see image launcher used twitch launcher memory allocated see image server lan single player single player resourcepack enabled no optifine installed no | 1 |
78,357 | 3,509,587,559 | IssuesEvent | 2016-01-08 23:38:29 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | opened | Prince Malchezaar (BB #1135) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** nort333
**Original Date:** 11.09.2015 19:27:01 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** new
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1135
<hr>
When useing his weapon, disarm spells not working on him. | 1.0 | Prince Malchezaar (BB #1135) - This issue was migrated from bitbucket.
**Original Reporter:** nort333
**Original Date:** 11.09.2015 19:27:01 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** new
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1135
<hr>
When useing his weapon, disarm spells not working on him. | priority | prince malchezaar bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state new direct link when useing his weapon disarm spells not working on him | 1 |
534,472 | 15,624,224,300 | IssuesEvent | 2021-03-21 01:10:14 | LujainKhalaf/Soengram | https://api.github.com/repos/LujainKhalaf/Soengram | closed | Add static analysis packages to CI/CD pipeline | 1 point med-risk priority-medium task | We want to add various static analysis packages to our CI/CD pipeline.
The packages should be added to `requirements.txt` and should be run automatically on each commit.
- Then black as the auto formatter https://github.com/psf/black#the-black-code-style
- Flake8 as the linter https://flake8.pycqa.org/en/latest/
- Radon for complexity analysis https://pypi.org/project/radon/
- Then pytest for unit and functional tests as well as code coverage https://docs.pytest.org/en/stable/
- mypy for static type checking http://mypy-lang.org/ | 1.0 | Add static analysis packages to CI/CD pipeline - We want to add various static analysis packages to our CI/CD pipeline.
The packages should be added to `requirements.txt` and should be run automatically on each commit.
- Then black as the auto formatter https://github.com/psf/black#the-black-code-style
- Flake8 as the linter https://flake8.pycqa.org/en/latest/
- Radon for complexity analysis https://pypi.org/project/radon/
- Then pytest for unit and functional tests as well as code coverage https://docs.pytest.org/en/stable/
- mypy for static type checking http://mypy-lang.org/ | priority | add static analysis packages to ci cd pipeline we want to add various static analysis packages to our ci cd pipeline the packages should be added to requirements txt and should be run automatically on each commit then black as the auto formatter as the linter radon for complexity analysis then pytest for unit and functional tests as well as code coverage mypy for static type checking | 1 |
359,619 | 10,678,761,500 | IssuesEvent | 2019-10-21 17:55:30 | AY1920S1-CS2113T-T12-3/main | https://api.github.com/repos/AY1920S1-CS2113T-T12-3/main | closed | As a user, I want to be able to save prices and locations of ingredients | priority.Medium type.Story | so that I do not have to input them every time | 1.0 | As a user, I want to be able to save prices and locations of ingredients - so that I do not have to input them every time | priority | as a user i want to be able to save prices and locations of ingredients so that i do not have to input them every time | 1 |
284,181 | 8,736,339,430 | IssuesEvent | 2018-12-11 19:15:54 | aowen87/TicketTester | https://api.github.com/repos/aowen87/TicketTester | closed | It isn't possible to see the material specific values in a mixed material zone in Label plot | bug crash likelihood medium priority reviewed severity high wrong results | Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 3002
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: It isn't possible to see the material specific values in a mixed material zone in Label plot
Assigned to: Alister Maguire
Category:
Target version: 2.13.1
Author: Eric Brugger
Start: 01/29/2018
Due date:
% Done: 100
Estimated time: 4.0
Created: 01/29/2018 08:17 pm
Updated: 02/26/2018 05:00 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.12.3
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working.
Comments:
pdbdatabase.C has a mixed scalar. Thanks for the tip about pdbdatabase.C containing mixed material variables. I tested VisIt 2.12.3 and it did the right thing. I will ask Matt for a data file. Matt gave me his file and when I force interface reconstruction, a Pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece. When I do a label plot of materials, it splits the cells and labels each piece with the correct material label. When I do a label plot of the mixed material variable it doesn't split the cell and displays the whole zone value. So, it is as Matt described, almost everything works correctly, just this one case doesn't. The data file he gave me is in ~brugger1/visit_obrien_bug on kickit. I suspect that the database generated by pdbdatabase.C should show the same behavior. Ok, specmix_ucd.silo has mixed variable values for variables "p" and "d" Hey everyone,Ive committed two revisions to the trunk.Revision 32458:This revision fixes a bug with the label plot not being able to correctly label mixed variables. Ive only been able to test on one dataset, but the issue should be resolved in general.M src/plots/Label/avtLabelFilter.CM src/plots/Label/avtLabelFilter.hRevision 32459:I added in a couple of tests to the label plot test file. These new tests will check to make sure that cells with mixed variables are being labeled correctly.A test/baseline/plots/label/label_11_00.pngA test/baseline/plots/label/label_11_01.pngM test/tests/plots/label.pyBest,Alister
| 1.0 | It isn't possible to see the material specific values in a mixed material zone in Label plot - Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 3002
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: It isn't possible to see the material specific values in a mixed material zone in Label plot
Assigned to: Alister Maguire
Category:
Target version: 2.13.1
Author: Eric Brugger
Start: 01/29/2018
Due date:
% Done: 100
Estimated time: 4.0
Created: 01/29/2018 08:17 pm
Updated: 02/26/2018 05:00 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.12.3
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Matt Obrien reported that with code he has mixed materials with material specific values. Normally VisIt always displays the whole zone values unless some material selection is being performed. If there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values. Neither of these mechanisms appear to be working. I couldn't verify this since none of our test data sets appear to have per material values for variables (or at least none that I could find). The first step is to modify the curv2d.silo generator to put out material specific values for the density that are markedly different so that we can test this. Then we should get this working.
Comments:
pdbdatabase.C has a mixed scalar. Thanks for the tip about pdbdatabase.C containing mixed material variables. I tested VisIt 2.12.3 and it did the right thing. I will ask Matt for a data file. Matt gave me his file and when I force interface reconstruction, a Pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece. When I do a label plot of materials, it splits the cells and labels each piece with the correct material label. When I do a label plot of the mixed material variable it doesn't split the cell and displays the whole zone value. So, it is as Matt described, almost everything works correctly, just this one case doesn't. The data file he gave me is in ~brugger1/visit_obrien_bug on kickit. I suspect that the database generated by pdbdatabase.C should show the same behavior. Ok, specmix_ucd.silo has mixed variable values for variables "p" and "d" Hey everyone,Ive committed two revisions to the trunk.Revision 32458:This revision fixes a bug with the label plot not being able to correctly label mixed variables. Ive only been able to test on one dataset, but the issue should be resolved in general.M src/plots/Label/avtLabelFilter.CM src/plots/Label/avtLabelFilter.hRevision 32459:I added in a couple of tests to the label plot test file. These new tests will check to make sure that cells with mixed variables are being labeled correctly.A test/baseline/plots/label/label_11_00.pngA test/baseline/plots/label/label_11_01.pngM test/tests/plots/label.pyBest,Alister
| priority | it isn t possible to see the material specific values in a mixed material zone in label plot matt obrien reported that with code he has mixed materials with material specific values normally visit always displays the whole zone values unless some material selection is being performed if there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values neither of these mechanisms appear to be working i couldn t verify this since none of our test data sets appear to have per material values for variables or at least none that i could find the first step is to modify the silo generator to put out material specific values for the density that are markedly different so that we can test this then we should get this working redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject it isn t possible to see the material specific values in a mixed material zone in label plot assigned to alister maguire category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description matt obrien reported that with code he has mixed materials with material specific values normally visit always displays the whole zone values unless some material selection is being performed if there are no materials being selected there is a setting that forces material interface reconstruction and allows seeing the per material values neither of these mechanisms appear to be working i couldn t verify this since none of our test data sets appear to have per material values for variables or at least none that i could find the first step is to modify the silo generator to put out material specific values for the density that are markedly different so that we can test this then we should get this working comments pdbdatabase c has a mixed scalar thanks for the tip about pdbdatabase c containing mixed material variables i tested visit and it did the right thing i will ask matt for a data file matt gave me his file and when i force interface reconstruction a pseudocolor plot of the mixed material variable splits the cells and displays the material specific value in each piece when i do a label plot of materials it splits the cells and labels each piece with the correct material label when i do a label plot of the mixed material variable it doesn t split the cell and displays the whole zone value so it is as matt described almost everything works correctly just this one case doesn t the data file he gave me is in visit obrien bug on kickit i suspect that the database generated by pdbdatabase c should show the same behavior ok specmix ucd silo has mixed variable values for variables p and d hey everyone ive committed two revisions to the trunk revision this revision fixes a bug with the label plot not being able to correctly label mixed variables ive only been able to test on one dataset but the issue should be resolved in general m src plots label avtlabelfilter cm src plots label avtlabelfilter hrevision i added in a couple of tests to the label plot test file these new tests will check to make sure that cells with mixed variables are being labeled correctly a test baseline plots label label pnga test baseline plots label label pngm test tests plots label pybest alister | 1 |
670,443 | 22,690,059,057 | IssuesEvent | 2022-07-04 18:45:14 | belav/csharpier | https://api.github.com/repos/belav/csharpier | closed | Extra space before curly brace in array initializer | type:bug area:formatting priority:medium | Here is what I expect:
```
public class MyClass
{
public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 };
}
```
Actual code:
```
public class MyClass
{
public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 };
}
```
Version: 0.18.0 | 1.0 | Extra space before curly brace in array initializer - Here is what I expect:
```
public class MyClass
{
public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 };
}
```
Actual code:
```
public class MyClass
{
public int[] Arr { get; set; } = { 1, 2, 3, 4, 5 };
}
```
Version: 0.18.0 | priority | extra space before curly brace in array initializer here is what i expect public class myclass public int arr get set actual code public class myclass public int arr get set version | 1 |
614,813 | 19,190,249,118 | IssuesEvent | 2021-12-05 21:43:05 | oresat/oresat-dxwifi-software | https://api.github.com/repos/oresat/oresat-dxwifi-software | opened | Add device tree overlay | medium priority | Device tree overlay should project based. So need to move the DxWiFi device tree from `oresat-linux` repo to here and add it to the `oresat-dxwifi-tx` Debian package. | 1.0 | Add device tree overlay - Device tree overlay should project based. So need to move the DxWiFi device tree from `oresat-linux` repo to here and add it to the `oresat-dxwifi-tx` Debian package. | priority | add device tree overlay device tree overlay should project based so need to move the dxwifi device tree from oresat linux repo to here and add it to the oresat dxwifi tx debian package | 1 |
583,526 | 17,391,377,207 | IssuesEvent | 2021-08-02 07:51:57 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Розширений пошук] Star rating control is misaligned | Priority: Medium bug | **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Click on 'Розширений пошук' icon
2. Click on 'List view' icon
**Actual result**
Star control is misaligned to the right side of the page.

**Expected result**
Star rating control is located below Club's logo and and summary in the left corner.

**User story and test case links**
E.g.: "User story #
| 1.0 | [Розширений пошук] Star rating control is misaligned - **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Click on 'Розширений пошук' icon
2. Click on 'List view' icon
**Actual result**
Star control is misaligned to the right side of the page.

**Expected result**
Star rating control is located below Club's logo and and summary in the left corner.

**User story and test case links**
E.g.: "User story #
| priority | star rating control is misaligned environment windows google chrome version official build bit reproducible always build found last commit preconditions go to steps to reproduce click on розширений пошук icon click on list view icon actual result star control is misaligned to the right side of the page expected result star rating control is located below club s logo and and summary in the left corner user story and test case links e g user story | 1 |
317,241 | 9,662,405,999 | IssuesEvent | 2019-05-20 20:44:28 | GingerWalnut/SQBeyondPublic | https://api.github.com/repos/GingerWalnut/SQBeyondPublic | closed | Combat Dummy Bug? | medium priority | So I was building a capital ship on Mardos late last night (5-5-19) and I took a shuttle down to the floor of my hangar and logged off for the night (around 9-10pm PST). There were no hostile mobs in sight and the area was well lit. There were also very few players online at that hour. Once I logged back on today at around 4pm PST, I got a greeting saying I had died with all my items and 28 levels because my combat dummy was slain, despite being nowhere near combat. Im not sure if this is a bug or my own fault? Help would be appreciated. If needed I can explain further in game. (IGN: BoyWonderFTW and base referred to is on Mardos under Isolators faction) | 1.0 | Combat Dummy Bug? - So I was building a capital ship on Mardos late last night (5-5-19) and I took a shuttle down to the floor of my hangar and logged off for the night (around 9-10pm PST). There were no hostile mobs in sight and the area was well lit. There were also very few players online at that hour. Once I logged back on today at around 4pm PST, I got a greeting saying I had died with all my items and 28 levels because my combat dummy was slain, despite being nowhere near combat. Im not sure if this is a bug or my own fault? Help would be appreciated. If needed I can explain further in game. (IGN: BoyWonderFTW and base referred to is on Mardos under Isolators faction) | priority | combat dummy bug so i was building a capital ship on mardos late last night and i took a shuttle down to the floor of my hangar and logged off for the night around pst there were no hostile mobs in sight and the area was well lit there were also very few players online at that hour once i logged back on today at around pst i got a greeting saying i had died with all my items and levels because my combat dummy was slain despite being nowhere near combat im not sure if this is a bug or my own fault help would be appreciated if needed i can explain further in game ign boywonderftw and base referred to is on mardos under isolators faction | 1 |
763,981 | 26,780,318,066 | IssuesEvent | 2023-01-31 20:34:35 | pystardust/ani-cli | https://api.github.com/repos/pystardust/ani-cli | opened | How do I change providers? | type: bug priority 2: medium | Used to be able to choose what website (Provider) to scrape from with "-r" option,
but it is now used for episode selection. how would I be able to scrape a different provider without editing the script? | 1.0 | How do I change providers? - Used to be able to choose what website (Provider) to scrape from with "-r" option,
but it is now used for episode selection. how would I be able to scrape a different provider without editing the script? | priority | how do i change providers used to be able to choose what website provider to scrape from with r option but it is now used for episode selection how would i be able to scrape a different provider without editing the script | 1 |
657,664 | 21,799,779,597 | IssuesEvent | 2022-05-16 02:54:00 | Wiredcraft/pipelines | https://api.github.com/repos/Wiredcraft/pipelines | closed | Add icons/update UI title | Priority: Medium Type: Enhancement | I got a whole bunch of `pipelines` setup - but they are all the same look & feel.
Can we;
- specify a custom icon?
- tune the title?
See example; 5 completely different setups on misc domains - very hard to distinguish.
<img width="1232" alt="screen shot 2018-03-23 at 09 26 06" src="https://user-images.githubusercontent.com/631984/37806748-935e3378-2e7c-11e8-8eb2-02d1276929d9.png">
| 1.0 | Add icons/update UI title - I got a whole bunch of `pipelines` setup - but they are all the same look & feel.
Can we;
- specify a custom icon?
- tune the title?
See example; 5 completely different setups on misc domains - very hard to distinguish.
<img width="1232" alt="screen shot 2018-03-23 at 09 26 06" src="https://user-images.githubusercontent.com/631984/37806748-935e3378-2e7c-11e8-8eb2-02d1276929d9.png">
| priority | add icons update ui title i got a whole bunch of pipelines setup but they are all the same look feel can we specify a custom icon tune the title see example completely different setups on misc domains very hard to distinguish img width alt screen shot at src | 1 |
133,290 | 5,200,305,700 | IssuesEvent | 2017-01-23 23:25:04 | mapbox/mapbox-gl-js | https://api.github.com/repos/mapbox/mapbox-gl-js | closed | Issue an error when map canvas is larger than allowed by gl.MAX_RENDERBUFFER_SIZE | medium priority starter task | Once the map canvas is larger than `gl.MAX_RENDERBUFFER_SIZE / 2`, scaling behaviour starts to occur. I do not think we should invest significant resources into accommodating this case.
See https://github.com/mapbox/mapbox-gl-js/issues/2673 and https://github.com/mapbox/mapbox-gl-js/issues/2863
| 1.0 | Issue an error when map canvas is larger than allowed by gl.MAX_RENDERBUFFER_SIZE - Once the map canvas is larger than `gl.MAX_RENDERBUFFER_SIZE / 2`, scaling behaviour starts to occur. I do not think we should invest significant resources into accommodating this case.
See https://github.com/mapbox/mapbox-gl-js/issues/2673 and https://github.com/mapbox/mapbox-gl-js/issues/2863
| priority | issue an error when map canvas is larger than allowed by gl max renderbuffer size once the map canvas is larger than gl max renderbuffer size scaling behaviour starts to occur i do not think we should invest significant resources into accommodating this case see and | 1 |
25,799 | 2,683,992,682 | IssuesEvent | 2015-03-28 15:06:43 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | -new_console глючит | 2–5 stars bug imported Priority-Medium | _From [sim....@gmail.com](https://code.google.com/u/105258257765487351754/) on January 28, 2011 03:16:27_
При запуске фара с ключем -new_console он запускается в странном укороченном по ширине виде, и при этом еще страшно тормозит клавиатура.
**Attachment:** [Буфер обмена-1.jpg](http://code.google.com/p/conemu-maximus5/issues/detail?id=362)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=362_ | 1.0 | -new_console глючит - _From [sim....@gmail.com](https://code.google.com/u/105258257765487351754/) on January 28, 2011 03:16:27_
При запуске фара с ключем -new_console он запускается в странном укороченном по ширине виде, и при этом еще страшно тормозит клавиатура.
**Attachment:** [Буфер обмена-1.jpg](http://code.google.com/p/conemu-maximus5/issues/detail?id=362)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=362_ | priority | new console глючит from on january при запуске фара с ключем new console он запускается в странном укороченном по ширине виде и при этом еще страшно тормозит клавиатура attachment original issue | 1 |
65,672 | 3,237,350,334 | IssuesEvent | 2015-10-14 11:23:04 | awesome-raccoons/gqt | https://api.github.com/repos/awesome-raccoons/gqt | opened | Weird things happen for high zoom levels | bug medium priority | To reproduce: Plot POLYGON((0 0, 10 0, 10 10, 0 0)) and press Ctr-+ more than 34 times (the real fun is around 38 presses).
I guess this is caused by floating point errors, so it's probably hard to fix it. A good enough fix is to just stop zooming before this becomes an issue. | 1.0 | Weird things happen for high zoom levels - To reproduce: Plot POLYGON((0 0, 10 0, 10 10, 0 0)) and press Ctr-+ more than 34 times (the real fun is around 38 presses).
I guess this is caused by floating point errors, so it's probably hard to fix it. A good enough fix is to just stop zooming before this becomes an issue. | priority | weird things happen for high zoom levels to reproduce plot polygon and press ctr more than times the real fun is around presses i guess this is caused by floating point errors so it s probably hard to fix it a good enough fix is to just stop zooming before this becomes an issue | 1 |
336,105 | 10,171,474,751 | IssuesEvent | 2019-08-08 08:29:05 | pmem/issues | https://api.github.com/repos/pmem/issues | closed | Tests fail because of can't a read alignment of devdax | Exposure: Medium OS: Linux Priority: 2 high State: To be verified Type: Bug Type: External Bug | <!--
Before creating new issue, ensure that similar issue wasn't already created
* Search: https://github.com/pmem/issues/issues
Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report.
Remember this is just a minimal template. You can extend it with data you think may be useful.
-->
## Environment Information
- PMDK package version(s): 1.6-237-g3ca1cd956
- OS(es) version(s): Fedora 29
- ndctl version(s): 65
- kernel version(s): 5.1.12
<!-- fill in also other useful environment data -->
## Please provide a reproduction of the bug:
A lot of PMDK tests fail for the same reason. Below I put an example:
```
$ ./RUNTESTS obj_basic_integration/ -s TEST7
obj_basic_integration/TEST7: SETUP (check/pmem/debug)
obj_basic_integration/TEST7 crashed (signal 6). err7.log below.
err7.log below.
out7.log below.
obj_basic_integration/TEST7 out7.log obj_basic_integration/TEST7: START: obj_basic_integration
obj_basic_integration/TEST7 out7.log ./obj_basic_integration /dev/dax1.5
pmem7.log below.
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:236 out_init] pid 28580: program: /home/jenkins/greg/pmdk/src/test/obj_basic_integration/obj_basic_integration
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:56 libpmem_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:784 pmem_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:419 pmem_init_funcs]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:376 pmem_cpuinfo_to_funcs] clflushopt supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:388 pmem_cpuinfo_to_funcs] clwb supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:325 use_avx512f_memcpy_memset] avx512f supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:329 use_avx512f_memcpy_memset] PMEM_AVX512F not set or not == 1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:483 pmem_init_funcs] using clwb
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:69 libpmem_fini]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:101 util_mmap_fini]
pmemblk7.log below.
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:130 libpmemblk_fini]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:101 util_mmap_fini]
pmemlog7.log below.
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:130 libpmemlog_fini]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:101 util_mmap_fini]
Last 30 lines of pmemobj7.log below (whole file has 42 lines).
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil)
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:301 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x8e8ed4
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x7f5d47644900 name fallocate.at_create indexes 0x7ffc08eac318
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:124 util_remote_init]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1328 pmemobj_createU] path /dev/dax1.5 layout basic poolsize 0 mode 600
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1298 obj_get_nlanes]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3355 util_pool_create] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 attr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3115 util_pool_create_uuids] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 pattr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 remote 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:2124 util_poolset_create_set] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:559 util_file_open] path "/dev/dax1.5" size 0x7ffc08eac118 minsize 0 flags 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:67 device_dax_size] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:1703 util_poolset_single] path /dev/dax1.5 filesize 6339690496 create 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:291 util_file_device_dax_alignment] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:212 device_dax_alignment] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [file_posix.c:230 device_dax_alignment] open "/sys/dev/char/252:11/device/align": No such file or directory
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [set.c:1752 util_poolset_single] assertion failure: rep->part[0].alignment (0x0) != 0 (0x0)
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:325 obj_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:139 util_remote_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [mmap.c:101 util_mmap_fini]
pmempool7.log below.
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:238 out_init] libpmempool version 1.3
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:69 libpmempool_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:124 util_remote_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:85 libpmempool_fini]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:139 util_remote_fini]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:101 util_mmap_fini]
rpmem7.log below.
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.3
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:61 librpmem_init]
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:77 librpmem_fini]
trace7.log below.
obj_basic_integration/TEST7 trace7.log {obj_basic_integration.c:653 main} obj_basic_integration/TEST7: START: obj_basic_integration
obj_basic_integration/TEST7 trace7.log ./obj_basic_integration /dev/dax1.5
RUNTESTS: stopping: obj_basic_integration//TEST7 failed, TEST=check FS=pmem BUILD=debug
```
## How often bug is revealed: (always, often, rare):
always
## Actual behavior:
As above.
## Expected behavior:
Tests should pass.
## Details
<!-- fill this out -->
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible -->
Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
| 1.0 | Tests fail because of can't a read alignment of devdax - <!--
Before creating new issue, ensure that similar issue wasn't already created
* Search: https://github.com/pmem/issues/issues
Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report.
Remember this is just a minimal template. You can extend it with data you think may be useful.
-->
## Environment Information
- PMDK package version(s): 1.6-237-g3ca1cd956
- OS(es) version(s): Fedora 29
- ndctl version(s): 65
- kernel version(s): 5.1.12
<!-- fill in also other useful environment data -->
## Please provide a reproduction of the bug:
A lot of PMDK tests fail for the same reason. Below I put an example:
```
$ ./RUNTESTS obj_basic_integration/ -s TEST7
obj_basic_integration/TEST7: SETUP (check/pmem/debug)
obj_basic_integration/TEST7 crashed (signal 6). err7.log below.
err7.log below.
out7.log below.
obj_basic_integration/TEST7 out7.log obj_basic_integration/TEST7: START: obj_basic_integration
obj_basic_integration/TEST7 out7.log ./obj_basic_integration /dev/dax1.5
pmem7.log below.
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:236 out_init] pid 28580: program: /home/jenkins/greg/pmdk/src/test/obj_basic_integration/obj_basic_integration
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmem7.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:56 libpmem_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:784 pmem_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:419 pmem_init_funcs]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:376 pmem_cpuinfo_to_funcs] clflushopt supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:388 pmem_cpuinfo_to_funcs] clwb supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:325 use_avx512f_memcpy_memset] avx512f supported
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:329 use_avx512f_memcpy_memset] PMEM_AVX512F not set or not == 1
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:483 pmem_init_funcs] using clwb
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [libpmem.c:69 libpmem_fini]
obj_basic_integration/TEST7 pmem7.log <libpmem>: <3> [mmap.c:101 util_mmap_fini]
pmemblk7.log below.
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [libpmemblk.c:130 libpmemblk_fini]
obj_basic_integration/TEST7 pmemblk7.log <libpmemblk>: <3> [mmap.c:101 util_mmap_fini]
pmemlog7.log below.
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [libpmemlog.c:130 libpmemlog_fini]
obj_basic_integration/TEST7 pmemlog7.log <libpmemlog>: <3> [mmap.c:101 util_mmap_fini]
Last 30 lines of pmemobj7.log below (whole file has 42 lines).
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil)
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:301 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x8e8ed4
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x7f5d47644900 name fallocate.at_create indexes 0x7ffc08eac318
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:124 util_remote_init]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1328 pmemobj_createU] path /dev/dax1.5 layout basic poolsize 0 mode 600
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:1298 obj_get_nlanes]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3355 util_pool_create] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 attr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:3115 util_pool_create_uuids] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608 minpartsize 2097152 pattr 0x7ffc08eac290 nlanes 0x7ffc08eac304 can_have_rep 1 remote 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:2124 util_poolset_create_set] setp 0x7ffc08eac308 path /dev/dax1.5 poolsize 0 minsize 8388608
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:559 util_file_open] path "/dev/dax1.5" size 0x7ffc08eac118 minsize 0 flags 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:67 device_dax_size] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:1703 util_poolset_single] path /dev/dax1.5 filesize 6339690496 create 0
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file.c:131 util_file_exists] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:291 util_file_device_dax_alignment] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [file_posix.c:212 device_dax_alignment] path "/dev/dax1.5"
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [file_posix.c:230 device_dax_alignment] open "/sys/dev/char/252:11/device/align": No such file or directory
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <1> [set.c:1752 util_poolset_single] assertion failure: rep->part[0].alignment (0x0) != 0 (0x0)
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [obj.c:325 obj_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:139 util_remote_fini]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmemobj7.log <libpmemobj>: <3> [mmap.c:101 util_mmap_fini]
pmempool7.log below.
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:238 out_init] libpmempool version 1.3
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:67 util_mmap_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:69 libpmempool_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:124 util_remote_init]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [libpmempool.c:85 libpmempool_fini]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:139 util_remote_fini]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [set.c:194 util_remote_unload]
obj_basic_integration/TEST7 pmempool7.log <libpmempool>: <3> [mmap.c:101 util_mmap_fini]
rpmem7.log below.
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:236 out_init] pid 28578: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.3
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git237.g3ca1cd956
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:61 librpmem_init]
obj_basic_integration/TEST7 rpmem7.log <librpmem>: <3> [librpmem.c:77 librpmem_fini]
trace7.log below.
obj_basic_integration/TEST7 trace7.log {obj_basic_integration.c:653 main} obj_basic_integration/TEST7: START: obj_basic_integration
obj_basic_integration/TEST7 trace7.log ./obj_basic_integration /dev/dax1.5
RUNTESTS: stopping: obj_basic_integration//TEST7 failed, TEST=check FS=pmem BUILD=debug
```
## How often bug is revealed: (always, often, rare):
always
## Actual behavior:
As above.
## Expected behavior:
Tests should pass.
## Details
<!-- fill this out -->
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible -->
Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
| priority | tests fail because of can t a read alignment of devdax before creating new issue ensure that similar issue wasn t already created search note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report remember this is just a minimal template you can extend it with data you think may be useful environment information pmdk package version s os es version s fedora ndctl version s kernel version s please provide a reproduction of the bug a lot of pmdk tests fail for the same reason below i put an example runtests obj basic integration s obj basic integration setup check pmem debug obj basic integration crashed signal log below log below log below obj basic integration log obj basic integration start obj basic integration obj basic integration log obj basic integration dev log below obj basic integration log pid program home jenkins greg pmdk src test obj basic integration obj basic integration obj basic integration log libpmem version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log clflush supported obj basic integration log clflushopt supported obj basic integration log clwb supported obj basic integration log avx supported obj basic integration log pmem avx not set or not obj basic integration log supported obj basic integration log pmem not set or not obj basic integration log obj basic integration log region path sys bus nd devices obj basic integration log domain path sys bus nd devices persistence domain obj basic integration log flushing cpu cache obj basic integration log using clwb obj basic integration log using movnt obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmemblk version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmemlog version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log last lines of log below whole file has lines obj basic integration log pop nil obj basic integration log ctl nil ctx nil cfg string fallocate at create obj basic integration log ctl nil ctx nil source name fallocate at create type arg obj basic integration log nodes name fallocate at create indexes obj basic integration log obj basic integration log path dev layout basic poolsize mode obj basic integration log obj basic integration log setp path dev poolsize minsize minpartsize attr nlanes can have rep obj basic integration log setp path dev poolsize minsize minpartsize pattr nlanes can have rep remote obj basic integration log path dev obj basic integration log setp path dev poolsize minsize obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev size minsize flags obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev filesize create obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log path dev obj basic integration log open sys dev char device align no such file or directory obj basic integration log assertion failure rep part alignment obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log libpmempool version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log obj basic integration log log below obj basic integration log pid program home jenkins greg pmdk src tools pmempool pmempool static debug obj basic integration log librpmem version obj basic integration log src version obj basic integration log compiled with support for valgrind pmemcheck obj basic integration log compiled with support for valgrind helgrind obj basic integration log compiled with support for valgrind memcheck obj basic integration log compiled with support for valgrind drd obj basic integration log compiled with support for shutdown state obj basic integration log compiled with libndctl obj basic integration log obj basic integration log log below obj basic integration log obj basic integration c main obj basic integration start obj basic integration obj basic integration log obj basic integration dev runtests stopping obj basic integration failed test check fs pmem build debug how often bug is revealed always often rare always actual behavior as above expected behavior tests should pass details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low | 1 |
374,403 | 11,089,301,542 | IssuesEvent | 2019-12-14 17:31:30 | bounswe/bounswe2019group4 | https://api.github.com/repos/bounswe/bounswe2019group4 | closed | Backend feature recommendation | Back-End Priority: Medium Type: Development | According to our requirements, system shall recommend new users, articles etc to logged in users according user's actions.
Endpoint of recommendations will be /recommendations , and method type will be GET. | 1.0 | Backend feature recommendation - According to our requirements, system shall recommend new users, articles etc to logged in users according user's actions.
Endpoint of recommendations will be /recommendations , and method type will be GET. | priority | backend feature recommendation according to our requirements system shall recommend new users articles etc to logged in users according user s actions endpoint of recommendations will be recommendations and method type will be get | 1 |
744,347 | 25,939,566,999 | IssuesEvent | 2022-12-16 17:01:38 | asastats/channel | https://api.github.com/repos/asastats/channel | closed | [B2] Wrong NFT purchase prices | bug medium priority addressed | We're going to have a bigger analysys and fixup.
Here are the example problems:
- - - - -
https://discord.com/channels/906917846754418770/908054330265960478/961212380904296459
Wrong last purchase price for FORUM Astro #\2
NFT: last purchase for FORUM Astro #\2 showing 25 ALGO, according to NFTexplorer = 55 ALGO on AlgoGems. Previous sale = 200 ALGO on ALGOxNFT
535631028
- - - - -
https://github.com/asastats/channel/issues/154#issuecomment-1171855303
The time looks good for these two different ones. Although the amount for the second shows 1 ALGO when in fact it was sold for 7.

Here's the group txn for that sale: https://algoexplorer.io/tx/group/%2F0aPItSoO2OM%2F9gf7CUk6qQsfpMnQlviXwJxczlJlh8%3D
- - - - - | 1.0 | [B2] Wrong NFT purchase prices - We're going to have a bigger analysys and fixup.
Here are the example problems:
- - - - -
https://discord.com/channels/906917846754418770/908054330265960478/961212380904296459
Wrong last purchase price for FORUM Astro #\2
NFT: last purchase for FORUM Astro #\2 showing 25 ALGO, according to NFTexplorer = 55 ALGO on AlgoGems. Previous sale = 200 ALGO on ALGOxNFT
535631028
- - - - -
https://github.com/asastats/channel/issues/154#issuecomment-1171855303
The time looks good for these two different ones. Although the amount for the second shows 1 ALGO when in fact it was sold for 7.

Here's the group txn for that sale: https://algoexplorer.io/tx/group/%2F0aPItSoO2OM%2F9gf7CUk6qQsfpMnQlviXwJxczlJlh8%3D
- - - - - | priority | wrong nft purchase prices we re going to have a bigger analysys and fixup here are the example problems wrong last purchase price for forum astro nft last purchase for forum astro showing algo according to nftexplorer algo on algogems previous sale algo on algoxnft the time looks good for these two different ones although the amount for the second shows algo when in fact it was sold for here s the group txn for that sale | 1 |
622,854 | 19,657,893,066 | IssuesEvent | 2022-01-10 14:20:51 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Delete & Select all icons shows up when there is no video | feature-enhancement priority-medium feature-video feature Stale | ### Scope
When even there is no video uploaded by the users it still displays the delete & select all icon on the top under profile->Video or anywhere we display the video tab.
See the Screenshot: https://nimbusweb.me/nimbus-screenshots/5873fd55b7f1d88d819f65dc0fe36755
### Acceptance Criteria
Delete & select all icons should not show when there is no video.
Support Ticket: https://secure.helpscout.net/conversation/1618101603/161750/
**Jira issue** : [PROD-750]
[PROD-750]: https://buddyboss.atlassian.net/browse/PROD-750?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | Delete & Select all icons shows up when there is no video - ### Scope
When even there is no video uploaded by the users it still displays the delete & select all icon on the top under profile->Video or anywhere we display the video tab.
See the Screenshot: https://nimbusweb.me/nimbus-screenshots/5873fd55b7f1d88d819f65dc0fe36755
### Acceptance Criteria
Delete & select all icons should not show when there is no video.
Support Ticket: https://secure.helpscout.net/conversation/1618101603/161750/
**Jira issue** : [PROD-750]
[PROD-750]: https://buddyboss.atlassian.net/browse/PROD-750?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | delete select all icons shows up when there is no video scope when even there is no video uploaded by the users it still displays the delete select all icon on the top under profile video or anywhere we display the video tab see the screenshot acceptance criteria delete select all icons should not show when there is no video support ticket jira issue | 1 |
36,246 | 2,797,417,779 | IssuesEvent | 2015-05-12 13:44:42 | twogee/ant-http | https://api.github.com/repos/twogee/ant-http | closed | [CLOSED] Setup build process with local properties file based on username | auto-migrated Milestone-1.0.1 Priority-Medium Project-ant-http Type-Task | <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday May 11, 2015 at 22:05 GMT_
_Originally opened as https://github.com/twogee/missing-link/issues/3_
----
```
Setup build process with local properties file based on username
* Include a sample properties file with sensitive data not included
* Build project into a single jar
* Build source jar
* Build javadocs
* Build api (javadoc) zip
* Build distribution folder with binary, source and javadocs
* Build distribution zip
```
Original issue reported on code.google.com by `alex.she...@gmail.com` on 13 Mar 2011 at 6:02
| 1.0 | [CLOSED] Setup build process with local properties file based on username - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday May 11, 2015 at 22:05 GMT_
_Originally opened as https://github.com/twogee/missing-link/issues/3_
----
```
Setup build process with local properties file based on username
* Include a sample properties file with sensitive data not included
* Build project into a single jar
* Build source jar
* Build javadocs
* Build api (javadoc) zip
* Build distribution folder with binary, source and javadocs
* Build distribution zip
```
Original issue reported on code.google.com by `alex.she...@gmail.com` on 13 Mar 2011 at 6:02
| priority | setup build process with local properties file based on username issue by monday may at gmt originally opened as setup build process with local properties file based on username include a sample properties file with sensitive data not included build project into a single jar build source jar build javadocs build api javadoc zip build distribution folder with binary source and javadocs build distribution zip original issue reported on code google com by alex she gmail com on mar at | 1 |
823,350 | 30,991,955,238 | IssuesEvent | 2023-08-09 05:46:21 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | opened | [YSQL] Update query which uses UNNEST function that results in Hash join takes 20X time for the same query without UNNEST | area/ysql priority/medium status/awaiting-triage | ### Description
To update a single row of a table(schema details below) using UNNEST it is taking close to ~85ms where as when same query is executed without UNNEST, it results an update with a simple index scan and completes in ~5ms. **The issue is observed when UNNEST results in hash join. UNNEST with nested loop seems to be taking same time when executed without UNNEST.**
This issue was resulting in high latency observed with HammerDB Stored Procedures. For details have a look at this document https://docs.google.com/document/d/1n1kfmNi2FeixmmkIhceSw89-qZsdKFj5rACl9vOQy8k/edit?usp=sharing
**Update query using UNNEST resulting in hash join**
```
yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 FROM UNNEST(ARRAY[1], ARRAY[1]) AS ids(o_id, d_id) WHERE col_bigint_1=ids.o_id AND col_bigint_2=ids.d_id;
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------------------
Update on pkey_rangescan_fulltable_10column_1 (cost=2.50..110.03 rows=2 width=248) (actual time=83.385..83.385 rows=0 loops=1)
-> Hash Join (cost=2.50..110.03 rows=2 width=248) (actual time=7.749..79.554 rows=1 loops=1)
Hash Cond: ((pkey_rangescan_fulltable_10column_1.col_bigint_1 = ids.o_id) AND (pkey_rangescan_fulltable_10column_1.col_bigint_2 = ids.d_id))
-> Seq Scan on pkey_rangescan_fulltable_10column_1 (cost=0.00..100.00 rows=1000 width=216) (actual time=7.714..78.906 rows=10000 loops=1)
-> Hash (cost=1.00..1.00 rows=100 width=40) (actual time=0.023..0.023 rows=1 loops=1)
Buckets: 1024 Batches: 1 Memory Usage: 9kB
-> Function Scan on ids (cost=0.01..1.00 rows=100 width=40) (actual time=0.017..0.017 rows=1 loops=1)
Planning Time: 0.128 ms
Execution Time: 83.454 ms
Peak Memory Usage: 272 kB
(10 rows)
Time: 84.511 ms
```
**Update query without UNNEST**
```
yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 WHERE col_bigint_1=1 AND col_bigint_2=1;
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
--------
Update on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.192..1.192 rows=0 loops=1)
-> Index Scan using pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.174..1.175 rows=1 l
oops=1)
Index Cond: ((col_bigint_1 = 1) AND (col_bigint_2 = 1))
Planning Time: 0.076 ms
Execution Time: 3.109 ms
Peak Memory Usage: 24 kB
(6 rows)
Time: 4.634 ms
```
**Table schema**
```
yugabyte=# \d+ pkey_rangescan_fulltable_10column_1
Table "public.pkey_rangescan_fulltable_10column_1"
Column | Type | Collation | Nullable | Default | Storage | Stats target | Description
-----------------+--------+-----------+----------+---------+---------+--------------+-------------
col_bigint_id_1 | bigint | | not null | | plain | |
col_bigint_1 | bigint | | | | plain | |
col_bigint_2 | bigint | | | | plain | |
col_bigint_3 | bigint | | | | plain | |
col_bigint_4 | bigint | | | | plain | |
col_bigint_5 | bigint | | | | plain | |
col_bigint_6 | bigint | | | | plain | |
col_bigint_7 | bigint | | | | plain | |
col_bigint_8 | bigint | | | | plain | |
col_bigint_9 | bigint | | | | plain | |
Indexes:
"pkey_rangescan_fulltable_10column_1_pkey" PRIMARY KEY, lsm (col_bigint_id_1 ASC)
"pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx" lsm (col_bigint_1 HASH, col_bigint_2 ASC)
```
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information. | 1.0 | [YSQL] Update query which uses UNNEST function that results in Hash join takes 20X time for the same query without UNNEST - ### Description
To update a single row of a table(schema details below) using UNNEST it is taking close to ~85ms where as when same query is executed without UNNEST, it results an update with a simple index scan and completes in ~5ms. **The issue is observed when UNNEST results in hash join. UNNEST with nested loop seems to be taking same time when executed without UNNEST.**
This issue was resulting in high latency observed with HammerDB Stored Procedures. For details have a look at this document https://docs.google.com/document/d/1n1kfmNi2FeixmmkIhceSw89-qZsdKFj5rACl9vOQy8k/edit?usp=sharing
**Update query using UNNEST resulting in hash join**
```
yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 FROM UNNEST(ARRAY[1], ARRAY[1]) AS ids(o_id, d_id) WHERE col_bigint_1=ids.o_id AND col_bigint_2=ids.d_id;
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------------------
Update on pkey_rangescan_fulltable_10column_1 (cost=2.50..110.03 rows=2 width=248) (actual time=83.385..83.385 rows=0 loops=1)
-> Hash Join (cost=2.50..110.03 rows=2 width=248) (actual time=7.749..79.554 rows=1 loops=1)
Hash Cond: ((pkey_rangescan_fulltable_10column_1.col_bigint_1 = ids.o_id) AND (pkey_rangescan_fulltable_10column_1.col_bigint_2 = ids.d_id))
-> Seq Scan on pkey_rangescan_fulltable_10column_1 (cost=0.00..100.00 rows=1000 width=216) (actual time=7.714..78.906 rows=10000 loops=1)
-> Hash (cost=1.00..1.00 rows=100 width=40) (actual time=0.023..0.023 rows=1 loops=1)
Buckets: 1024 Batches: 1 Memory Usage: 9kB
-> Function Scan on ids (cost=0.01..1.00 rows=100 width=40) (actual time=0.017..0.017 rows=1 loops=1)
Planning Time: 0.128 ms
Execution Time: 83.454 ms
Peak Memory Usage: 272 kB
(10 rows)
Time: 84.511 ms
```
**Update query without UNNEST**
```
yugabyte=# explain analyse update pkey_rangescan_fulltable_10column_1 set col_bigint_4=col_bigint_4+1 WHERE col_bigint_1=1 AND col_bigint_2=1;
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
--------
Update on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.192..1.192 rows=0 loops=1)
-> Index Scan using pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx on pkey_rangescan_fulltable_10column_1 (cost=0.00..5.28 rows=10 width=216) (actual time=1.174..1.175 rows=1 l
oops=1)
Index Cond: ((col_bigint_1 = 1) AND (col_bigint_2 = 1))
Planning Time: 0.076 ms
Execution Time: 3.109 ms
Peak Memory Usage: 24 kB
(6 rows)
Time: 4.634 ms
```
**Table schema**
```
yugabyte=# \d+ pkey_rangescan_fulltable_10column_1
Table "public.pkey_rangescan_fulltable_10column_1"
Column | Type | Collation | Nullable | Default | Storage | Stats target | Description
-----------------+--------+-----------+----------+---------+---------+--------------+-------------
col_bigint_id_1 | bigint | | not null | | plain | |
col_bigint_1 | bigint | | | | plain | |
col_bigint_2 | bigint | | | | plain | |
col_bigint_3 | bigint | | | | plain | |
col_bigint_4 | bigint | | | | plain | |
col_bigint_5 | bigint | | | | plain | |
col_bigint_6 | bigint | | | | plain | |
col_bigint_7 | bigint | | | | plain | |
col_bigint_8 | bigint | | | | plain | |
col_bigint_9 | bigint | | | | plain | |
Indexes:
"pkey_rangescan_fulltable_10column_1_pkey" PRIMARY KEY, lsm (col_bigint_id_1 ASC)
"pkey_rangescan_fulltable_10column_col_bigint_1_col_bigint_2_idx" lsm (col_bigint_1 HASH, col_bigint_2 ASC)
```
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information. | priority | update query which uses unnest function that results in hash join takes time for the same query without unnest description to update a single row of a table schema details below using unnest it is taking close to where as when same query is executed without unnest it results an update with a simple index scan and completes in the issue is observed when unnest results in hash join unnest with nested loop seems to be taking same time when executed without unnest this issue was resulting in high latency observed with hammerdb stored procedures for details have a look at this document update query using unnest resulting in hash join yugabyte explain analyse update pkey rangescan fulltable set col bigint col bigint from unnest array array as ids o id d id where col bigint ids o id and col bigint ids d id query plan update on pkey rangescan fulltable cost rows width actual time rows loops hash join cost rows width actual time rows loops hash cond pkey rangescan fulltable col bigint ids o id and pkey rangescan fulltable col bigint ids d id seq scan on pkey rangescan fulltable cost rows width actual time rows loops hash cost rows width actual time rows loops buckets batches memory usage function scan on ids cost rows width actual time rows loops planning time ms execution time ms peak memory usage kb rows time ms update query without unnest yugabyte explain analyse update pkey rangescan fulltable set col bigint col bigint where col bigint and col bigint query plan update on pkey rangescan fulltable cost rows width actual time rows loops index scan using pkey rangescan fulltable col bigint col bigint idx on pkey rangescan fulltable cost rows width actual time rows l oops index cond col bigint and col bigint planning time ms execution time ms peak memory usage kb rows time ms table schema yugabyte d pkey rangescan fulltable table public pkey rangescan fulltable column type collation nullable default storage stats target description col bigint id bigint not null plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain col bigint bigint plain indexes pkey rangescan fulltable pkey primary key lsm col bigint id asc pkey rangescan fulltable col bigint col bigint idx lsm col bigint hash col bigint asc warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information | 1 |
20,580 | 2,622,853,814 | IssuesEvent | 2015-03-04 08:06:40 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | closed | user needs to modify the monitored URL | auto-migrated Priority-Medium | ```
What steps will reproduce the problem? Please include a URL.
Some URLs will be changed automatically and so it can't be monitored correctly.
The link is:
http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district
What is the expected output? What do you see instead?
The link is turned to be:
http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district/district
1_10.html
What version of the Chrome are you using? On what operating system?
30.0.1599.101
Win7 64
```
Original issue reported on code.google.com by `wong....@gmail.com` on 25 Oct 2013 at 10:27
* Merged into: #97 | 1.0 | user needs to modify the monitored URL - ```
What steps will reproduce the problem? Please include a URL.
Some URLs will be changed automatically and so it can't be monitored correctly.
The link is:
http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district
What is the expected output? What do you see instead?
The link is turned to be:
http://www.lcsd.gov.hk/CE/Entertainment/EntertainmentOffice/b5/district/district
1_10.html
What version of the Chrome are you using? On what operating system?
30.0.1599.101
Win7 64
```
Original issue reported on code.google.com by `wong....@gmail.com` on 25 Oct 2013 at 10:27
* Merged into: #97 | priority | user needs to modify the monitored url what steps will reproduce the problem please include a url some urls will be changed automatically and so it can t be monitored correctly the link is what is the expected output what do you see instead the link is turned to be html what version of the chrome are you using on what operating system original issue reported on code google com by wong gmail com on oct at merged into | 1 |
632,846 | 20,236,681,772 | IssuesEvent | 2022-02-14 03:20:48 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | You can't pet the cat/etc | Priority: 1-Urgent Type: Feature Difficulty: 2-Medium | ## Description
If you try to interact with say a caracal and probably other animals, you do not pet them
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
Not sure if this should be only visible to the one petting (and getting pet, I guess), or if it should also be global, though. | 1.0 | You can't pet the cat/etc - ## Description
If you try to interact with say a caracal and probably other animals, you do not pet them
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
Not sure if this should be only visible to the one petting (and getting pet, I guess), or if it should also be global, though. | priority | you can t pet the cat etc description if you try to interact with say a caracal and probably other animals you do not pet them not sure if this should be only visible to the one petting and getting pet i guess or if it should also be global though | 1 |
427,509 | 12,396,155,945 | IssuesEvent | 2020-05-20 19:59:59 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User | Priority: Medium Product: caseflow-queue Team: Echo 🐬 Type: Bug | ## Description
The Assigned To field should display the correct user ID on the search cases page
## Acceptance criteria
- [ ] The assigned to field on the search cases page displays the correct user assigned to the case
## Background/context/resources
During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again
Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots
## Technical notes
Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search.
## Screenshots

| 1.0 | Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User - ## Description
The Assigned To field should display the correct user ID on the search cases page
## Acceptance criteria
- [ ] The assigned to field on the search cases page displays the correct user assigned to the case
## Background/context/resources
During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again
Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots
## Technical notes
Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search.
## Screenshots

| priority | case movement non blocking tasks assigned to field not displaying correct user description the assigned to field should display the correct user id on the search cases page acceptance criteria the assigned to field on the search cases page displays the correct user assigned to the case background context resources during dogfooding assigning a case to an judge legacy reassigning to burnham assigned case to instead of css when searching the case again check the for screenshots technical notes reproduce before digging into hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs this would leave assigned to unchanged in case search screenshots | 1 |
129,740 | 5,102,586,693 | IssuesEvent | 2017-01-04 18:47:12 | TauCetiStation/TauCetiClassic | https://api.github.com/repos/TauCetiStation/TauCetiClassic | closed | Показывает плохое состояние здоровья при 100хп | bug priority: medium | Originally reported by: **Anonymous**
---
Дело было вечером. Админов не было. Находился в районе получаса со сломанной ногой в плохом и крите(в дальнейшем состоянии). Привезли на станцию,все вроде бы залатали,
но теперь первое - никого не слышно,хотя иногда фразы проскакивают
второе - не могу встать с каталки
третье -показывает плохое состояние здоровья при 100 хп
четвертое - меня никто не может поднять
Ps: не гоуст
Pps: под конец уже начали маяться херней,так как онлайн был 10 человек,половина из которых сидела и говорила в LOOC
---
- Bitbucket: https://bitbucket.org/taucetistation/tauceti/issue/154
| 1.0 | Показывает плохое состояние здоровья при 100хп - Originally reported by: **Anonymous**
---
Дело было вечером. Админов не было. Находился в районе получаса со сломанной ногой в плохом и крите(в дальнейшем состоянии). Привезли на станцию,все вроде бы залатали,
но теперь первое - никого не слышно,хотя иногда фразы проскакивают
второе - не могу встать с каталки
третье -показывает плохое состояние здоровья при 100 хп
четвертое - меня никто не может поднять
Ps: не гоуст
Pps: под конец уже начали маяться херней,так как онлайн был 10 человек,половина из которых сидела и говорила в LOOC
---
- Bitbucket: https://bitbucket.org/taucetistation/tauceti/issue/154
| priority | показывает плохое состояние здоровья при originally reported by anonymous дело было вечером админов не было находился в районе получаса со сломанной ногой в плохом и крите в дальнейшем состоянии привезли на станцию все вроде бы залатали но теперь первое никого не слышно хотя иногда фразы проскакивают второе не могу встать с каталки третье показывает плохое состояние здоровья при хп четвертое меня никто не может поднять ps не гоуст pps под конец уже начали маяться херней так как онлайн был человек половина из которых сидела и говорила в looc bitbucket | 1 |
196,324 | 6,926,823,712 | IssuesEvent | 2017-11-30 20:32:48 | AZMAG/map-ATP | https://api.github.com/repos/AZMAG/map-ATP | closed | Remove the date field completely from the main form | Priority: Medium | Removing the validation and just send in the current date. | 1.0 | Remove the date field completely from the main form - Removing the validation and just send in the current date. | priority | remove the date field completely from the main form removing the validation and just send in the current date | 1 |
548,161 | 16,058,980,059 | IssuesEvent | 2021-04-23 09:43:43 | input-output-hk/cardano-node | https://api.github.com/repos/input-output-hk/cardano-node | closed | Add instructions to README on how to run node with testnet and mainnet | byron priority medium | As was fairly [pointed out on Reddit](https://www.reddit.com/r/cardano/comments/f879vw/docker_image_for_cardanonode_160_is_here/): there no instructions on
> ... how to connect the current code-base to either Byron mainnet or testnet, do you have some docs on that or is it going to be more of a trial and error type of thing? | 1.0 | Add instructions to README on how to run node with testnet and mainnet - As was fairly [pointed out on Reddit](https://www.reddit.com/r/cardano/comments/f879vw/docker_image_for_cardanonode_160_is_here/): there no instructions on
> ... how to connect the current code-base to either Byron mainnet or testnet, do you have some docs on that or is it going to be more of a trial and error type of thing? | priority | add instructions to readme on how to run node with testnet and mainnet as was fairly there no instructions on how to connect the current code base to either byron mainnet or testnet do you have some docs on that or is it going to be more of a trial and error type of thing | 1 |
657,841 | 21,869,624,559 | IssuesEvent | 2022-05-19 03:12:27 | pixley/TimelineBuilder | https://api.github.com/repos/pixley/TimelineBuilder | opened | Parallel calendars | type: feature status: to do priority: medium | Allow for dynamic switching between calendar system views on one timeline. For example, a user could switch between Absalom Reckoning and the Imperial Calendar for their Pathfinder timeline, and the dates should change accordingly. | 1.0 | Parallel calendars - Allow for dynamic switching between calendar system views on one timeline. For example, a user could switch between Absalom Reckoning and the Imperial Calendar for their Pathfinder timeline, and the dates should change accordingly. | priority | parallel calendars allow for dynamic switching between calendar system views on one timeline for example a user could switch between absalom reckoning and the imperial calendar for their pathfinder timeline and the dates should change accordingly | 1 |
47,266 | 2,974,605,205 | IssuesEvent | 2015-07-15 02:18:13 | Reimashi/jotai | https://api.github.com/repos/Reimashi/jotai | closed | Laptop battery monitoring support | auto-migrated Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1. Downloaded
2. Started and checked all options available in View/Sensors menu
3. I can not see the fan speed and many others as presented on your web
site. Please see the attached file
What is the expected output? What do you see instead?
I would like to see the meters for all the hardware, as it is presented on
the website
What version of the product are you using? On what operating system?
I am using the latest version on a Windows 7 OS
Please provide any additional information below.
```
Original issue reported on code.google.com by `linasel` on 23 Feb 2010 at 10:12
Attachments:
* [Hardware monitor.jpg](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-4/comment-0/Hardware monitor.jpg)
| 1.0 | Laptop battery monitoring support - ```
What steps will reproduce the problem?
1. Downloaded
2. Started and checked all options available in View/Sensors menu
3. I can not see the fan speed and many others as presented on your web
site. Please see the attached file
What is the expected output? What do you see instead?
I would like to see the meters for all the hardware, as it is presented on
the website
What version of the product are you using? On what operating system?
I am using the latest version on a Windows 7 OS
Please provide any additional information below.
```
Original issue reported on code.google.com by `linasel` on 23 Feb 2010 at 10:12
Attachments:
* [Hardware monitor.jpg](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-4/comment-0/Hardware monitor.jpg)
| priority | laptop battery monitoring support what steps will reproduce the problem downloaded started and checked all options available in view sensors menu i can not see the fan speed and many others as presented on your web site please see the attached file what is the expected output what do you see instead i would like to see the meters for all the hardware as it is presented on the website what version of the product are you using on what operating system i am using the latest version on a windows os please provide any additional information below original issue reported on code google com by linasel on feb at attachments monitor jpg | 1 |
645,903 | 21,032,355,684 | IssuesEvent | 2022-03-31 02:44:24 | hackforla/tdm-calculator | https://api.github.com/repos/hackforla/tdm-calculator | reopened | Add icons for Accordion having external links | role: front-end level: medium priority: MUST HAVE p-Feature - Tool Tips | ### Overview
For any text having links to external websites, we should be adding an icon that indicates that.
### Action Items
- [ ] For external links add the icon that shows its going to an external website (https://designsystem.digital.gov/components/link/ )

- [ ] Add the image for the external links on the AIN/APN row on page 1
- [ ] Add the image for the external links on the Bike Share on page 2
- [ ] Add the image for the external links on the About Us page
### Resources/Instructions
REPLACE THIS TEXT -If there is a website which has documentation that helps with this issue provide the link(s) here.
| 1.0 | Add icons for Accordion having external links - ### Overview
For any text having links to external websites, we should be adding an icon that indicates that.
### Action Items
- [ ] For external links add the icon that shows its going to an external website (https://designsystem.digital.gov/components/link/ )

- [ ] Add the image for the external links on the AIN/APN row on page 1
- [ ] Add the image for the external links on the Bike Share on page 2
- [ ] Add the image for the external links on the About Us page
### Resources/Instructions
REPLACE THIS TEXT -If there is a website which has documentation that helps with this issue provide the link(s) here.
| priority | add icons for accordion having external links overview for any text having links to external websites we should be adding an icon that indicates that action items for external links add the icon that shows its going to an external website add the image for the external links on the ain apn row on page add the image for the external links on the bike share on page add the image for the external links on the about us page resources instructions replace this text if there is a website which has documentation that helps with this issue provide the link s here | 1 |
673,666 | 23,026,459,415 | IssuesEvent | 2022-07-22 09:40:02 | apache/incubator-devlake | https://api.github.com/repos/apache/incubator-devlake | closed | Extract file-level data from git repository | type/feature-request priority/medium need-follow-up | ## User story
As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution.
## Goal
The key is to match `bug` to `component`.
- If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`;
- If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`;
## Proposed Solution
- [ ] Add a `commit_files` table in the domain layer
| field | type | length | description | key |
| ------------- | -------- | -------- | ----------- | -----|
| `commit_sha` | char | 40 | | FK |
| `file_name` | varchar | 255 | | |
| `status` | varchar | 255 | | |
| `additions` | int | | | |
| `deletions` | int | | | |
| `changes` | int | | | |
| `file_component` | varchar | 255 | | |
a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.**
b. how to assign `file_component` value to a commit file? It varies in different programming languages.
- In Go, a component can be found by folder names;
- while in other languages, it might need user input to define the `file_component`.
- [ ] Allow users to define the `file_component` in `.env` by REGEX
## An example of commit files
For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files:

## Belongs to issue
This is a sub-issue of both #949 and #1771
## Additional context
N/A
| 1.0 | Extract file-level data from git repository - ## User story
As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution.
## Goal
The key is to match `bug` to `component`.
- If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`;
- If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`;
## Proposed Solution
- [ ] Add a `commit_files` table in the domain layer
| field | type | length | description | key |
| ------------- | -------- | -------- | ----------- | -----|
| `commit_sha` | char | 40 | | FK |
| `file_name` | varchar | 255 | | |
| `status` | varchar | 255 | | |
| `additions` | int | | | |
| `deletions` | int | | | |
| `changes` | int | | | |
| `file_component` | varchar | 255 | | |
a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.**
b. how to assign `file_component` value to a commit file? It varies in different programming languages.
- In Go, a component can be found by folder names;
- while in other languages, it might need user input to define the `file_component`.
- [ ] Allow users to define the `file_component` in `.env` by REGEX
## An example of commit files
For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files:

## Belongs to issue
This is a sub-issue of both #949 and #1771
## Additional context
N/A
| priority | extract file level data from git repository user story as a tech lead i want to see the bug distribution in low level components from code files as a guide for the future framework evolution goal the key is to match bug to component if we want to find the overall distribution of all bugs we can do issues bug join prs join commits join commit files component if we want to find the overall distribution of bugs fixed in a specific tag we can do ref diff issues bug join prs join commits join commit files component proposed solution add a commit files table in the domain layer field type length description key commit sha char fk file name varchar status varchar additions int deletions int changes int file component varchar a all the columns except file component can be retrieved from note not from or gitlab api b how to assign file component value to a commit file it varies in different programming languages in go a component can be found by folder names while in other languages it might need user input to define the file component allow users to define the file component in env by regex an example of commit files for instance this touches files belongs to issue this is a sub issue of both and additional context n a | 1 |
744,418 | 25,942,244,804 | IssuesEvent | 2022-12-16 19:41:03 | zowe/zowe-cli | https://api.github.com/repos/zowe/zowe-cli | closed | Command "zowe profiles create" prints the password that should be secret | enhancement for-review priority-medium community-upvoted | I am using Zowe CLI in a CI pipeline.
As a part of the pipeline, I call following command:
```
zowe profiles create zosmf-profile zowe \
--host $ZOS_HOST --port $ZOS_ZOSMF_PORT \
--user "$USERID" --pass "$PASSWORD"
```
In the log of the job is the following output:
```
Profile created successfully! Path:
/var/lib/jenkins_agent/workspace/Verify Zowe Prereqs/profiles/zosmf/zowe.yaml
host: myhost
port: 443
user: MYUSER
password: MYPASSWORD
rejectUnauthorized: true
protocol: https
```
I am not happy that my password is printed and all other users can see it. | 1.0 | Command "zowe profiles create" prints the password that should be secret - I am using Zowe CLI in a CI pipeline.
As a part of the pipeline, I call following command:
```
zowe profiles create zosmf-profile zowe \
--host $ZOS_HOST --port $ZOS_ZOSMF_PORT \
--user "$USERID" --pass "$PASSWORD"
```
In the log of the job is the following output:
```
Profile created successfully! Path:
/var/lib/jenkins_agent/workspace/Verify Zowe Prereqs/profiles/zosmf/zowe.yaml
host: myhost
port: 443
user: MYUSER
password: MYPASSWORD
rejectUnauthorized: true
protocol: https
```
I am not happy that my password is printed and all other users can see it. | priority | command zowe profiles create prints the password that should be secret i am using zowe cli in a ci pipeline as a part of the pipeline i call following command zowe profiles create zosmf profile zowe host zos host port zos zosmf port user userid pass password in the log of the job is the following output profile created successfully path var lib jenkins agent workspace verify zowe prereqs profiles zosmf zowe yaml host myhost port user myuser password mypassword rejectunauthorized true protocol https i am not happy that my password is printed and all other users can see it | 1 |
56,835 | 3,081,206,364 | IssuesEvent | 2015-08-22 13:47:43 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Create stats database schema if none is present on server | 015a bug imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on February 10, 2011 16:06:23_
When a server needs a sqlite stats database, if one is not present, we need to create a new one, tables and all.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=57_ | 1.0 | Create stats database schema if none is present on server - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on February 10, 2011 16:06:23_
When a server needs a sqlite stats database, if one is not present, we need to create a new one, tables and all.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=57_ | priority | create stats database schema if none is present on server from on february when a server needs a sqlite stats database if one is not present we need to create a new one tables and all original issue | 1 |
237,759 | 7,763,897,616 | IssuesEvent | 2018-06-01 18:14:30 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | [RFE] [nl] Job Status graph could use 'dd-mm-yyyy' date format (currently uses 'mm/dd') | component:ui priority:medium state:needs_info type:enhancement | ##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- UI
##### SUMMARY
<!-- Briefly describe the problem. -->
In Dutch (Netherlands) google-chrome, Job Status graph could use 'dd-mm-yyyy' date format, as per [1] and [2]
##### ENVIRONMENT
* AWX version: 1.1.0
* AWX install method: docker on linux
* Ansible version: 2.3.1.0
* Operating System: Fedora 25
* Web Browser: google-chrome
### Steps To Reproduce:
<!-- For bugs, please show exactly how to reproduce the problem. For new
features, show how the feature would be used. -->
1. In Dutch google-chrome, do a successful login and login into Tower
2. Post login, observe 'date' format on X axis (Time) of Job Status graph
### Expected Results:
<!-- For bug reports, what did you expect to happen when running the steps
above? -->
dd-mm-yyyy
### Actual Results:
<!-- For bug reports, what actually happened? -->
mm/dd
### Additional Information:
<!-- Include any links to sosreport, database dumps, screenshots or other
information. -->
[1] https://en.wikipedia.org/wiki/Date_format_by_country
[2] https://en.wikipedia.org/wiki/Date_and_time_notation_in_the_Netherlands
[3] Screenshot of the Job Status graph currently showing date in 'mm/dd' format on X axis (Time):

| 1.0 | [RFE] [nl] Job Status graph could use 'dd-mm-yyyy' date format (currently uses 'mm/dd') - ##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- UI
##### SUMMARY
<!-- Briefly describe the problem. -->
In Dutch (Netherlands) google-chrome, Job Status graph could use 'dd-mm-yyyy' date format, as per [1] and [2]
##### ENVIRONMENT
* AWX version: 1.1.0
* AWX install method: docker on linux
* Ansible version: 2.3.1.0
* Operating System: Fedora 25
* Web Browser: google-chrome
### Steps To Reproduce:
<!-- For bugs, please show exactly how to reproduce the problem. For new
features, show how the feature would be used. -->
1. In Dutch google-chrome, do a successful login and login into Tower
2. Post login, observe 'date' format on X axis (Time) of Job Status graph
### Expected Results:
<!-- For bug reports, what did you expect to happen when running the steps
above? -->
dd-mm-yyyy
### Actual Results:
<!-- For bug reports, what actually happened? -->
mm/dd
### Additional Information:
<!-- Include any links to sosreport, database dumps, screenshots or other
information. -->
[1] https://en.wikipedia.org/wiki/Date_format_by_country
[2] https://en.wikipedia.org/wiki/Date_and_time_notation_in_the_Netherlands
[3] Screenshot of the Job Status graph currently showing date in 'mm/dd' format on X axis (Time):

| priority | job status graph could use dd mm yyyy date format currently uses mm dd issue type feature idea component name ui summary in dutch netherlands google chrome job status graph could use dd mm yyyy date format as per and environment awx version awx install method docker on linux ansible version operating system fedora web browser google chrome steps to reproduce for bugs please show exactly how to reproduce the problem for new features show how the feature would be used in dutch google chrome do a successful login and login into tower post login observe date format on x axis time of job status graph expected results for bug reports what did you expect to happen when running the steps above dd mm yyyy actual results mm dd additional information include any links to sosreport database dumps screenshots or other information screenshot of the job status graph currently showing date in mm dd format on x axis time | 1 |
495,632 | 14,285,435,322 | IssuesEvent | 2020-11-23 13:55:09 | geocollections/sarv-edit | https://api.github.com/repos/geocollections/sarv-edit | closed | Drillcore box alternative list views | Difficulty: Medium Priority: Low Source: App Status: Available Type: Feature Request | Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared).
A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):
 

| 1.0 | Drillcore box alternative list views - Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared).
A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):
 

| priority | drillcore box alternative list views drillcore box filter list view should have separate view with box images one for each box if no image placeholder should be shown similar to file list view where images are shown within table when clicking on box image large image should be shown in separate popup window that will remain open so that two or more box images can be compared a different view is needed for printing box images into pdf two landscape images per portrait sheet   | 1 |
236,049 | 7,745,634,737 | IssuesEvent | 2018-05-29 18:58:35 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Have to press E twice to re-open a world object ui you just closed | Medium Priority | This seems to be an error in PlayerHeldObject line ~255, something about the previous interaction object being the same. | 1.0 | Have to press E twice to re-open a world object ui you just closed - This seems to be an error in PlayerHeldObject line ~255, something about the previous interaction object being the same. | priority | have to press e twice to re open a world object ui you just closed this seems to be an error in playerheldobject line something about the previous interaction object being the same | 1 |
26,545 | 2,684,862,783 | IssuesEvent | 2015-03-29 13:08:53 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | opened | Crashes on binary data. | 1 star bug imported Priority-Medium | _From [gynvael](https://code.google.com/u/gynvael/) on May 12, 2013 03:29:56_
Required information! OS version: Win7 SP1 x64 ConEmu version: 130427 [64]
Far version (if you are using Far Manager): n/a *Bug description* To check the stability of ConEmu I've played a little with outputting binary data and invalid terminal control sequences. In some cases some crashes do appear - usually the crash is in the console application (and not ConEmu itself), but it's due to the ConEmuHk hook. Sometimes ConEmu itself crashes (though tbh I have not managed to reproduce it on the newest ConEmu version; this problem was mostly in older versions).
(I've attached the test app - ansifuzz.cpp)
Additionally, sometimes a simple: int main(){for(;;)putchar(rand());} freezez ConEmu after pressing ctrl-c to break the output.
Note: I have enabled the following In-console options (I think they are relevant):
* Inject ConEmuHk * ANSI X3.64 / xterm 256 colors *Steps to reproduction* 1. Download the attached file and run: type last_test.bin
2. It should go for a while and then step. Press ctrl+c - it will output some more text and crash.
I didn't analyze the bugs further.
**Attachment:** [ansifuzz.cpp last_test.bin](http://code.google.com/p/conemu-maximus5/issues/detail?id=1064)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1064_ | 1.0 | Crashes on binary data. - _From [gynvael](https://code.google.com/u/gynvael/) on May 12, 2013 03:29:56_
Required information! OS version: Win7 SP1 x64 ConEmu version: 130427 [64]
Far version (if you are using Far Manager): n/a *Bug description* To check the stability of ConEmu I've played a little with outputting binary data and invalid terminal control sequences. In some cases some crashes do appear - usually the crash is in the console application (and not ConEmu itself), but it's due to the ConEmuHk hook. Sometimes ConEmu itself crashes (though tbh I have not managed to reproduce it on the newest ConEmu version; this problem was mostly in older versions).
(I've attached the test app - ansifuzz.cpp)
Additionally, sometimes a simple: int main(){for(;;)putchar(rand());} freezez ConEmu after pressing ctrl-c to break the output.
Note: I have enabled the following In-console options (I think they are relevant):
* Inject ConEmuHk * ANSI X3.64 / xterm 256 colors *Steps to reproduction* 1. Download the attached file and run: type last_test.bin
2. It should go for a while and then step. Press ctrl+c - it will output some more text and crash.
I didn't analyze the bugs further.
**Attachment:** [ansifuzz.cpp last_test.bin](http://code.google.com/p/conemu-maximus5/issues/detail?id=1064)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1064_ | priority | crashes on binary data from on may required information os version conemu version far version if you are using far manager n a bug description to check the stability of conemu i ve played a little with outputting binary data and invalid terminal control sequences in some cases some crashes do appear usually the crash is in the console application and not conemu itself but it s due to the conemuhk hook sometimes conemu itself crashes though tbh i have not managed to reproduce it on the newest conemu version this problem was mostly in older versions i ve attached the test app ansifuzz cpp additionally sometimes a simple int main for putchar rand freezez conemu after pressing ctrl c to break the output note i have enabled the following in console options i think they are relevant inject conemuhk ansi xterm colors steps to reproduction download the attached file and run type last test bin it should go for a while and then step press ctrl c it will output some more text and crash i didn t analyze the bugs further attachment original issue | 1 |
817,561 | 30,644,288,791 | IssuesEvent | 2023-07-25 02:22:59 | diba-io/bitmask-core | https://api.github.com/repos/diba-io/bitmask-core | closed | Return Genesis Timestamp in RGB Struct | priority-medium | I'm pretty sure RGB embeds a timestamp in contract genesis, or at least, it used to. We need that. | 1.0 | Return Genesis Timestamp in RGB Struct - I'm pretty sure RGB embeds a timestamp in contract genesis, or at least, it used to. We need that. | priority | return genesis timestamp in rgb struct i m pretty sure rgb embeds a timestamp in contract genesis or at least it used to we need that | 1 |
585,894 | 17,537,624,676 | IssuesEvent | 2021-08-12 08:21:39 | sjeong12/My-Editor | https://api.github.com/repos/sjeong12/My-Editor | closed | 개발도구 조사 및 선정 | doc priority: medium | ## 개요
다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정
- 유저 인터랙션에 대한 반응속도 100ms 미만을 목표로
- SPA로 구성
- 다양한 상태관리 필요
## 작업내용
- [x] 요구사항에 맞는 개발 도구 조사
- [x] 각 개발도구 성능 비교 및 선정
| 1.0 | 개발도구 조사 및 선정 - ## 개요
다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정
- 유저 인터랙션에 대한 반응속도 100ms 미만을 목표로
- SPA로 구성
- 다양한 상태관리 필요
## 작업내용
- [x] 요구사항에 맞는 개발 도구 조사
- [x] 각 개발도구 성능 비교 및 선정
| priority | 개발도구 조사 및 선정 개요 다음 에디터의 특성과 목표에 따라 적절한 개발도구를 선정 유저 인터랙션에 대한 반응속도 미만을 목표로 spa로 구성 다양한 상태관리 필요 작업내용 요구사항에 맞는 개발 도구 조사 각 개발도구 성능 비교 및 선정 | 1 |
206,046 | 7,108,221,723 | IssuesEvent | 2018-01-16 22:59:19 | qlicker/qlicker | https://api.github.com/repos/qlicker/qlicker | closed | course/:id/results route for students | Medium priority enhancement | This should show the same view as for an instructor, except with only the row for the current student. It means modifying the ClassParticipationList element to subscribe to the correct data.
Currently, **have set the route for a student to redirect to their login route**, but should update this
Presumably, this is originally meant to be handled by the route /results/:studentId/:courseId, but it currently doesn't work (assume that StudentResultPage needs to be updated. **I've commented this route out**, which is called in sessionClickHandler of student/courses.jsx - I don't think that an overview of the course results is what we want anyway, when clicking on a session item.
| 1.0 | course/:id/results route for students - This should show the same view as for an instructor, except with only the row for the current student. It means modifying the ClassParticipationList element to subscribe to the correct data.
Currently, **have set the route for a student to redirect to their login route**, but should update this
Presumably, this is originally meant to be handled by the route /results/:studentId/:courseId, but it currently doesn't work (assume that StudentResultPage needs to be updated. **I've commented this route out**, which is called in sessionClickHandler of student/courses.jsx - I don't think that an overview of the course results is what we want anyway, when clicking on a session item.
| priority | course id results route for students this should show the same view as for an instructor except with only the row for the current student it means modifying the classparticipationlist element to subscribe to the correct data currently have set the route for a student to redirect to their login route but should update this presumably this is originally meant to be handled by the route results studentid courseid but it currently doesn t work assume that studentresultpage needs to be updated i ve commented this route out which is called in sessionclickhandler of student courses jsx i don t think that an overview of the course results is what we want anyway when clicking on a session item | 1 |
105,001 | 4,227,399,840 | IssuesEvent | 2016-07-03 05:43:57 | OperationCode/operationcode | https://api.github.com/repos/OperationCode/operationcode | closed | [Code Schools] Add The Guild of Software Architects | Priority: Medium Status: In Progress Type: Maintenance | /codeschools
The Guild of Software Architects
6170 Research Rd, Suite 107
Frisco,TX 75033 | 1.0 | [Code Schools] Add The Guild of Software Architects - /codeschools
The Guild of Software Architects
6170 Research Rd, Suite 107
Frisco,TX 75033 | priority | add the guild of software architects codeschools the guild of software architects research rd suite frisco tx | 1 |
296,721 | 9,125,640,227 | IssuesEvent | 2019-02-24 15:25:47 | pixijs/pixi.js | https://api.github.com/repos/pixijs/pixi.js | closed | Double RenderTexture destroy | Plugin: cacheAsBitmap Priority: Medium Resolution: Won't Fix Type: Bug | This demo hacks the destroyTexture method and adds counter to it. If you place a breakpoint in it, you see that every renderTexture is destroyed two times: one time when the BaseTexture is destroyed and TextureManager disposes the texture, second time when the FrameBuffer is destroyed.
https://exponenta.github.io/gstatsjs/exampless/pixi/ | 1.0 | Double RenderTexture destroy - This demo hacks the destroyTexture method and adds counter to it. If you place a breakpoint in it, you see that every renderTexture is destroyed two times: one time when the BaseTexture is destroyed and TextureManager disposes the texture, second time when the FrameBuffer is destroyed.
https://exponenta.github.io/gstatsjs/exampless/pixi/ | priority | double rendertexture destroy this demo hacks the destroytexture method and adds counter to it if you place a breakpoint in it you see that every rendertexture is destroyed two times one time when the basetexture is destroyed and texturemanager disposes the texture second time when the framebuffer is destroyed | 1 |
7,677 | 2,603,742,773 | IssuesEvent | 2015-02-24 17:41:39 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | proposal on sneaky units | auto-migrated Priority-Medium Type-Enhancement | ```
You may have noticed, when there is a cloacked enemy unit, it's not quite
invisible, but there is a slight background distortion in place. People
easily localize cloacked units when they are moving, but of course can not
destroy them until tecnically "detected".
I suggest a new state Unit::isSneaking() which is always true when the
cloacked unit is moving and turns false as soon as the unit stops or
uncloakes. When isSneaking()==true you can access the unit's type and all
movement/order related properties like position, but cannot access any of
the other values like player, and ofcourse, trying to attack it would
result in an error.
Some callback will help greatly, like onUnitSneaking.
When the unit stops onUnitCloak will be called again.
```
-----
Original issue reported on code.google.com by `goo...@teabix.com` on 18 Nov 2009 at 2:17 | 1.0 | proposal on sneaky units - ```
You may have noticed, when there is a cloacked enemy unit, it's not quite
invisible, but there is a slight background distortion in place. People
easily localize cloacked units when they are moving, but of course can not
destroy them until tecnically "detected".
I suggest a new state Unit::isSneaking() which is always true when the
cloacked unit is moving and turns false as soon as the unit stops or
uncloakes. When isSneaking()==true you can access the unit's type and all
movement/order related properties like position, but cannot access any of
the other values like player, and ofcourse, trying to attack it would
result in an error.
Some callback will help greatly, like onUnitSneaking.
When the unit stops onUnitCloak will be called again.
```
-----
Original issue reported on code.google.com by `goo...@teabix.com` on 18 Nov 2009 at 2:17 | priority | proposal on sneaky units you may have noticed when there is a cloacked enemy unit it s not quite invisible but there is a slight background distortion in place people easily localize cloacked units when they are moving but of course can not destroy them until tecnically detected i suggest a new state unit issneaking which is always true when the cloacked unit is moving and turns false as soon as the unit stops or uncloakes when issneaking true you can access the unit s type and all movement order related properties like position but cannot access any of the other values like player and ofcourse trying to attack it would result in an error some callback will help greatly like onunitsneaking when the unit stops onunitcloak will be called again original issue reported on code google com by goo teabix com on nov at | 1 |
676,068 | 23,115,327,470 | IssuesEvent | 2022-07-27 16:09:05 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | refactor platform.findPr to return the newest PRs first | priority-3-medium type:refactor status:ready | ### Describe the proposed change(s).
refactor according to the discussion
https://github.com/renovatebot/renovate/discussions/16803 | 1.0 | refactor platform.findPr to return the newest PRs first - ### Describe the proposed change(s).
refactor according to the discussion
https://github.com/renovatebot/renovate/discussions/16803 | priority | refactor platform findpr to return the newest prs first describe the proposed change s refactor according to the discussion | 1 |
246,314 | 7,894,564,684 | IssuesEvent | 2018-06-28 22:00:12 | enforcer574/smashclub | https://api.github.com/repos/enforcer574/smashclub | opened | Smash Ultimate UI Refresh | Complexity: Medium Priority: 3 - Medium Type: User Request | Considerations for UI update:
- Change header fonts to something more "exciting"
- Replace character images with SmashUltimate images | 1.0 | Smash Ultimate UI Refresh - Considerations for UI update:
- Change header fonts to something more "exciting"
- Replace character images with SmashUltimate images | priority | smash ultimate ui refresh considerations for ui update change header fonts to something more exciting replace character images with smashultimate images | 1 |
639,636 | 20,760,691,595 | IssuesEvent | 2022-03-15 15:54:53 | AY2122S2-CS2103T-T12-4/tp | https://api.github.com/repos/AY2122S2-CS2103T-T12-4/tp | opened | Show tasks by date | priority.Medium | As a user, I can see my tasks grouped by date so that I can better plan my schedule. | 1.0 | Show tasks by date - As a user, I can see my tasks grouped by date so that I can better plan my schedule. | priority | show tasks by date as a user i can see my tasks grouped by date so that i can better plan my schedule | 1 |
217,642 | 7,325,571,507 | IssuesEvent | 2018-03-03 10:43:14 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | opened | pfstats pointer and connection errors seen in logs | Priority: Medium Type: Bug | > Mar 3 05:42:15 packetfence pfstats: panic: runtime error: invalid memory address or nil pointer dereference
> Mar 3 05:42:15 packetfence pfstats: [signal SIGSEGV: segmentation violation code=0x1 addr=0x30 pc=0x61f597]
> Mar 3 05:42:15 packetfence pfstats: goroutine 34 [running]:
> Mar 3 05:42:15 packetfence pfstats: github.com/inverse-inc/packetfence/go/log.LoggerWContext(0x0, 0x0, 0x0, 0x0, 0x0, 0xc4200a4140, 0xc42020f93b)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/log/log.go:128 +0x37
> Mar 3 05:42:15 packetfence pfstats: main.gauge.Send(0xc42020f900, 0x43, 0x696b00, 0x8fb040)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:53 +0x103
> Mar 3 05:42:15 packetfence pfstats: main.forward(0x79d0e0, 0xc42024a000)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:218 +0x5c2
> Mar 3 05:42:15 packetfence pfstats: created by main.main
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:356 +0x789
> Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=eror msg="write udp 127.0.0.1:51946->127.0.0.1:8125: write: connection refused" pid=21267
> Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=info msg="Starting stats server" pid=21267
> | 1.0 | pfstats pointer and connection errors seen in logs - > Mar 3 05:42:15 packetfence pfstats: panic: runtime error: invalid memory address or nil pointer dereference
> Mar 3 05:42:15 packetfence pfstats: [signal SIGSEGV: segmentation violation code=0x1 addr=0x30 pc=0x61f597]
> Mar 3 05:42:15 packetfence pfstats: goroutine 34 [running]:
> Mar 3 05:42:15 packetfence pfstats: github.com/inverse-inc/packetfence/go/log.LoggerWContext(0x0, 0x0, 0x0, 0x0, 0x0, 0xc4200a4140, 0xc42020f93b)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/log/log.go:128 +0x37
> Mar 3 05:42:15 packetfence pfstats: main.gauge.Send(0xc42020f900, 0x43, 0x696b00, 0x8fb040)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:53 +0x103
> Mar 3 05:42:15 packetfence pfstats: main.forward(0x79d0e0, 0xc42024a000)
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:218 +0x5c2
> Mar 3 05:42:15 packetfence pfstats: created by main.main
> Mar 3 05:42:15 packetfence pfstats: /root/gospace/src/github.com/inverse-inc/packetfence/go/stats/main.go:356 +0x789
> Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=eror msg="write udp 127.0.0.1:51946->127.0.0.1:8125: write: connection refused" pid=21267
> Mar 3 05:42:15 packetfence /usr/local/pf/bin/pfstats[21267]: t=2018-03-03T05:42:15-0500 lvl=info msg="Starting stats server" pid=21267
> | priority | pfstats pointer and connection errors seen in logs mar packetfence pfstats panic runtime error invalid memory address or nil pointer dereference mar packetfence pfstats mar packetfence pfstats goroutine mar packetfence pfstats github com inverse inc packetfence go log loggerwcontext mar packetfence pfstats root gospace src github com inverse inc packetfence go log log go mar packetfence pfstats main gauge send mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence pfstats main forward mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence pfstats created by main main mar packetfence pfstats root gospace src github com inverse inc packetfence go stats main go mar packetfence usr local pf bin pfstats t lvl eror msg write udp write connection refused pid mar packetfence usr local pf bin pfstats t lvl info msg starting stats server pid | 1 |
552,075 | 16,194,264,574 | IssuesEvent | 2021-05-04 12:49:35 | EvanQuan/Chubberino | https://api.github.com/repos/EvanQuan/Chubberino | opened | Refund heisters on quitting mid-heists | enhancement high priority medium effort | Currently if the program ends in the middle of any heists, all heisters lose all cheese they wagered.
Should return all cheese on quitting, preferrable with a message. | 1.0 | Refund heisters on quitting mid-heists - Currently if the program ends in the middle of any heists, all heisters lose all cheese they wagered.
Should return all cheese on quitting, preferrable with a message. | priority | refund heisters on quitting mid heists currently if the program ends in the middle of any heists all heisters lose all cheese they wagered should return all cheese on quitting preferrable with a message | 1 |
388,670 | 11,490,579,134 | IssuesEvent | 2020-02-11 17:19:55 | radical-cybertools/radical.pilot | https://api.github.com/repos/radical-cybertools/radical.pilot | closed | Netiface requirement for 0.10.4 | layer:rct priority:medium topic:deployment type:bug | When installing RP from pip on MacOS 10.14 I see the following:
```
radical-pilot 0.50.22 has requirement netifaces==0.10.4, but you'll have netifaces 0.10.9 which is incompatible.
```
Local tests with EnTK perform as expected but I am not sure whether this would be a problem with more "realistic" runs. Is requirement for `==0.10.4` a hard requirement? | 1.0 | Netiface requirement for 0.10.4 - When installing RP from pip on MacOS 10.14 I see the following:
```
radical-pilot 0.50.22 has requirement netifaces==0.10.4, but you'll have netifaces 0.10.9 which is incompatible.
```
Local tests with EnTK perform as expected but I am not sure whether this would be a problem with more "realistic" runs. Is requirement for `==0.10.4` a hard requirement? | priority | netiface requirement for when installing rp from pip on macos i see the following radical pilot has requirement netifaces but you ll have netifaces which is incompatible local tests with entk perform as expected but i am not sure whether this would be a problem with more realistic runs is requirement for a hard requirement | 1 |
604,225 | 18,679,852,017 | IssuesEvent | 2021-11-01 03:08:56 | AY2122S1-CS2113T-T12-3/tp | https://api.github.com/repos/AY2122S1-CS2113T-T12-3/tp | closed | [PE-D] When the excel files has some problems, the app just wipes all the excel entries. | type.Task priority.Medium severity.Medium | It is better to ask the user to backup the file or make a copy (eg. StonksXD_Entries.csv.temp) of that excel entry instead of just wiping the corrupted file.


<!--session: 1635497124781-460394b5-e63f-406f-b23c-56ce5c43e5d1-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.FunctionalityBug` `severity.Medium`
original: YEOWEIHNGWHYELAB/ped#3 | 1.0 | [PE-D] When the excel files has some problems, the app just wipes all the excel entries. - It is better to ask the user to backup the file or make a copy (eg. StonksXD_Entries.csv.temp) of that excel entry instead of just wiping the corrupted file.


<!--session: 1635497124781-460394b5-e63f-406f-b23c-56ce5c43e5d1-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.FunctionalityBug` `severity.Medium`
original: YEOWEIHNGWHYELAB/ped#3 | priority | when the excel files has some problems the app just wipes all the excel entries it is better to ask the user to backup the file or make a copy eg stonksxd entries csv temp of that excel entry instead of just wiping the corrupted file labels type functionalitybug severity medium original yeoweihngwhyelab ped | 1 |
362,200 | 10,724,196,528 | IssuesEvent | 2019-10-28 00:08:45 | qlcchain/go-qlc | https://api.github.com/repos/qlcchain/go-qlc | closed | badger may lock if gqlc exit abnormally | Priority: Medium Stale Status: Review Needed Type: Bug | ### Description of the issue
badger may lock if gqlc exit abnormally
### Issue-Type
- [x] bug report
- [ ] feature request
- [ ] Documentation improvement
### Checklist
- [ ] Running latest version of code.
- [ ] This issue has not been reported earlier.
### Your environment
* OS
* Go version
* Release tag/commit of the code
### Expected behavior
What should happen?
### Actual behavior
What is actually happening?
### Steps to reproduce
1. Provide detailed steps
2. how to reproduce this issue
### Any extra info ( for eg. code snippet to reproduce, logs, etc. )
If necessary, provide some extra information like code-snippets or error-logs. | 1.0 | badger may lock if gqlc exit abnormally - ### Description of the issue
badger may lock if gqlc exit abnormally
### Issue-Type
- [x] bug report
- [ ] feature request
- [ ] Documentation improvement
### Checklist
- [ ] Running latest version of code.
- [ ] This issue has not been reported earlier.
### Your environment
* OS
* Go version
* Release tag/commit of the code
### Expected behavior
What should happen?
### Actual behavior
What is actually happening?
### Steps to reproduce
1. Provide detailed steps
2. how to reproduce this issue
### Any extra info ( for eg. code snippet to reproduce, logs, etc. )
If necessary, provide some extra information like code-snippets or error-logs. | priority | badger may lock if gqlc exit abnormally description of the issue badger may lock if gqlc exit abnormally issue type bug report feature request documentation improvement checklist running latest version of code this issue has not been reported earlier your environment os go version release tag commit of the code expected behavior what should happen actual behavior what is actually happening steps to reproduce provide detailed steps how to reproduce this issue any extra info for eg code snippet to reproduce logs etc if necessary provide some extra information like code snippets or error logs | 1 |
690,676 | 23,669,124,892 | IssuesEvent | 2022-08-27 04:12:38 | ecotiya/ecotiya-portfolio-site | https://api.github.com/repos/ecotiya/ecotiya-portfolio-site | opened | Github Actionsを利用して、AWSにDeployできる仕組みの構築。 | enhancement Priority: Medium | <!-- 要望のテンプレート -->
## 概要
Github Actionsを利用して、AWSにDeployできる仕組みの構築。
ECRをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。
Terraformでインフラをコード化する。
## 目的
基本的にGithub Actionsを利用して、AWSにDeployできる仕組みの構築ができれば良い。
その際にはテストコードを走らせ、自動的にAWSにデプロイしてほしい。勉強用。
## タスク
- [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を踏まえて、疑問・調査事項の解決を実施する。
- [ ] 何をすべきかをタスク化。
## 疑問・調査事項
- [ ] Github Actionsとは何をするのかを調査。
- [ ] Terraformとは何かを調査。現行は[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)を使用してコード化しているため、必要ないのではないか。
- [ ] letest管理からの脱却。
## 補足
・一般的に、TerraformやCloudFormationのソースコードってgithubに上げるべきなのか確認。 | 1.0 | Github Actionsを利用して、AWSにDeployできる仕組みの構築。 - <!-- 要望のテンプレート -->
## 概要
Github Actionsを利用して、AWSにDeployできる仕組みの構築。
ECRをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。
Terraformでインフラをコード化する。
## 目的
基本的にGithub Actionsを利用して、AWSにDeployできる仕組みの構築ができれば良い。
その際にはテストコードを走らせ、自動的にAWSにデプロイしてほしい。勉強用。
## タスク
- [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を踏まえて、疑問・調査事項の解決を実施する。
- [ ] 何をすべきかをタスク化。
## 疑問・調査事項
- [ ] Github Actionsとは何をするのかを調査。
- [ ] Terraformとは何かを調査。現行は[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)を使用してコード化しているため、必要ないのではないか。
- [ ] letest管理からの脱却。
## 補足
・一般的に、TerraformやCloudFormationのソースコードってgithubに上げるべきなのか確認。 | priority | github actionsを利用して、awsにdeployできる仕組みの構築。 概要 github actionsを利用して、awsにdeployできる仕組みの構築。 ecrをバージョン管理する。バージョン毎に本番環境へ配信できるように対応を実施する。 terraformでインフラをコード化する。 目的 基本的にgithub actionsを利用して、awsにdeployできる仕組みの構築ができれば良い。 その際にはテストコードを走らせ、自動的にawsにデプロイしてほしい。勉強用。 タスク 何をすべきかをタスク化。 疑問・調査事項 github actionsとは何をするのかを調査。 terraformとは何かを調査。現行は letest管理からの脱却。 補足 ・一般的に、terraformやcloudformationのソースコードってgithubに上げるべきなのか確認。 | 1 |
57,205 | 3,081,248,330 | IssuesEvent | 2015-08-22 14:39:21 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | settime command lost ability to do unlimited time | bug imported Priority-Medium | _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on May 02, 2014 20:33:55_
in a game, do: /settime 0
It will end the level.
I don't remember, was this intended behavior?
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=428_ | 1.0 | settime command lost ability to do unlimited time - _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on May 02, 2014 20:33:55_
in a game, do: /settime 0
It will end the level.
I don't remember, was this intended behavior?
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=428_ | priority | settime command lost ability to do unlimited time from on may in a game do settime it will end the level i don t remember was this intended behavior original issue | 1 |
201,417 | 7,031,175,406 | IssuesEvent | 2017-12-26 15:38:25 | xcodeswift/sake | https://api.github.com/repos/xcodeswift/sake | closed | Fix issue with extra line breaks | difficulty:easy good first issue priority:medium status:ready-development | ## Context 🕵️♀️
There seems to be an issue when printing the output from the shell commands. Sometimes it adds unnecessary extra lines.
## Proposal 🎉
Fix it! | 1.0 | Fix issue with extra line breaks - ## Context 🕵️♀️
There seems to be an issue when printing the output from the shell commands. Sometimes it adds unnecessary extra lines.
## Proposal 🎉
Fix it! | priority | fix issue with extra line breaks context 🕵️♀️ there seems to be an issue when printing the output from the shell commands sometimes it adds unnecessary extra lines proposal 🎉 fix it | 1 |
169,156 | 6,395,853,074 | IssuesEvent | 2017-08-04 14:14:19 | vigetlabs/GroundRules | https://api.github.com/repos/vigetlabs/GroundRules | reopened | [QA] On Mobile, Text Size for Device Problems Too Big/Long | fed medium-priority mobile | Browser: Safari, Mobile
URL: GroundRules.co/devices
_Current:_
*Device problem text runs into/past the arrow at the end of the line
*Device problem text is not vertically centered
_Expected:_
*Device problem text should end before the arrow at the end of the line
*Device problem text should be vertically centered with arrow | 1.0 | [QA] On Mobile, Text Size for Device Problems Too Big/Long - Browser: Safari, Mobile
URL: GroundRules.co/devices
_Current:_
*Device problem text runs into/past the arrow at the end of the line
*Device problem text is not vertically centered
_Expected:_
*Device problem text should end before the arrow at the end of the line
*Device problem text should be vertically centered with arrow | priority | on mobile text size for device problems too big long browser safari mobile url groundrules co devices current device problem text runs into past the arrow at the end of the line device problem text is not vertically centered expected device problem text should end before the arrow at the end of the line device problem text should be vertically centered with arrow | 1 |
168,396 | 6,370,773,632 | IssuesEvent | 2017-08-01 14:49:00 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Add support for conflating a randomly calculated tile using the HG workflow | Category: Core Priority: Medium Status: Ready For Review Type: Feature | Create new bug issues doing single tile conflation as they are found. | 1.0 | Add support for conflating a randomly calculated tile using the HG workflow - Create new bug issues doing single tile conflation as they are found. | priority | add support for conflating a randomly calculated tile using the hg workflow create new bug issues doing single tile conflation as they are found | 1 |
243,875 | 7,868,156,147 | IssuesEvent | 2018-06-23 17:51:23 | cdgco/VestaWebInterface | https://api.github.com/repos/cdgco/VestaWebInterface | closed | Edit Backup Exclusions | Backend Priority: Medium Status: Accepted Type: Enhancement | Support for editing of backup exclusions. Data must be uploaded from frontend to backend, added to temp file, then directory linked to user.
Frontend framework in place. | 1.0 | Edit Backup Exclusions - Support for editing of backup exclusions. Data must be uploaded from frontend to backend, added to temp file, then directory linked to user.
Frontend framework in place. | priority | edit backup exclusions support for editing of backup exclusions data must be uploaded from frontend to backend added to temp file then directory linked to user frontend framework in place | 1 |
723,768 | 24,907,429,710 | IssuesEvent | 2022-10-29 12:41:11 | KryptonMC/Krypton | https://api.github.com/repos/KryptonMC/Krypton | closed | Implement a backup system | priority: medium area: server | ### Description
Introduce a backup system that's built-in to the server so users can guarantee the security of their worlds if something goes wrong.
### Goals
The following are goals of this system:
- Provide a high-performance and efficient automatic backup systems
- Provide the ability to automatically backup your worlds on a timer, and also manually queue backups
- Provide hooks in to common backup systems, such as AWS S3, Google Drive, or Microsoft OneDrive.
- Provide an API for listening for backups and creating new upload sources.
- Possibly using a custom data storage format to more efficiently store data than Anvil, and supporting custom writing formats in the API.
### Non goals
The following, however, are not goals of this system:
- Replace any of the standard loading or saving logic
- Using a custom packing format, such as a replacement for ZIP or TAR.
### Motivation
Worlds are very important, they are where most of the game takes place, and securing their usability should be vital. However, sometimes things go wrong, and sometimes we can't guarantee this security. In Krypton, this is especially true, as it is quite unstable, and in my time testing it, I have had to delete and repopulate the region folder of the world I use to test so many times that I've actually forgotten most of them.
For vanilla Minecraft, and especially Bukkit, there are numerous solutions that attempt to implement backups in to Minecraft. However, I have had mixed results using them in production environments for servers I have worked for in the past, them being slow, inefficient, or consuming huge amounts of memory.
This system is intended to replace any possible third-party solutions. | 1.0 | Implement a backup system - ### Description
Introduce a backup system that's built-in to the server so users can guarantee the security of their worlds if something goes wrong.
### Goals
The following are goals of this system:
- Provide a high-performance and efficient automatic backup systems
- Provide the ability to automatically backup your worlds on a timer, and also manually queue backups
- Provide hooks in to common backup systems, such as AWS S3, Google Drive, or Microsoft OneDrive.
- Provide an API for listening for backups and creating new upload sources.
- Possibly using a custom data storage format to more efficiently store data than Anvil, and supporting custom writing formats in the API.
### Non goals
The following, however, are not goals of this system:
- Replace any of the standard loading or saving logic
- Using a custom packing format, such as a replacement for ZIP or TAR.
### Motivation
Worlds are very important, they are where most of the game takes place, and securing their usability should be vital. However, sometimes things go wrong, and sometimes we can't guarantee this security. In Krypton, this is especially true, as it is quite unstable, and in my time testing it, I have had to delete and repopulate the region folder of the world I use to test so many times that I've actually forgotten most of them.
For vanilla Minecraft, and especially Bukkit, there are numerous solutions that attempt to implement backups in to Minecraft. However, I have had mixed results using them in production environments for servers I have worked for in the past, them being slow, inefficient, or consuming huge amounts of memory.
This system is intended to replace any possible third-party solutions. | priority | implement a backup system description introduce a backup system that s built in to the server so users can guarantee the security of their worlds if something goes wrong goals the following are goals of this system provide a high performance and efficient automatic backup systems provide the ability to automatically backup your worlds on a timer and also manually queue backups provide hooks in to common backup systems such as aws google drive or microsoft onedrive provide an api for listening for backups and creating new upload sources possibly using a custom data storage format to more efficiently store data than anvil and supporting custom writing formats in the api non goals the following however are not goals of this system replace any of the standard loading or saving logic using a custom packing format such as a replacement for zip or tar motivation worlds are very important they are where most of the game takes place and securing their usability should be vital however sometimes things go wrong and sometimes we can t guarantee this security in krypton this is especially true as it is quite unstable and in my time testing it i have had to delete and repopulate the region folder of the world i use to test so many times that i ve actually forgotten most of them for vanilla minecraft and especially bukkit there are numerous solutions that attempt to implement backups in to minecraft however i have had mixed results using them in production environments for servers i have worked for in the past them being slow inefficient or consuming huge amounts of memory this system is intended to replace any possible third party solutions | 1 |
484,832 | 13,957,956,679 | IssuesEvent | 2020-10-24 09:32:15 | sButtons/sbuttons | https://api.github.com/repos/sButtons/sbuttons | closed | Removing border from splash button | Priority: Medium buttons enhancement | **Is your feature request related to a problem? Please describe.**
Splash button is having borders of very odd colors , which don't go good with the gradient inside of them.
**Describe the solution you'd like**
After removing the borders, it will have a good sleek n elegant look.
**How the button currently looks \ How it'll look after removing wierd borders :-**
The gradient is animated in this button so it looks good with no borders.

**Additional notes**
I'll make a pull request as soon as possible if assigned. | 1.0 | Removing border from splash button - **Is your feature request related to a problem? Please describe.**
Splash button is having borders of very odd colors , which don't go good with the gradient inside of them.
**Describe the solution you'd like**
After removing the borders, it will have a good sleek n elegant look.
**How the button currently looks \ How it'll look after removing wierd borders :-**
The gradient is animated in this button so it looks good with no borders.

**Additional notes**
I'll make a pull request as soon as possible if assigned. | priority | removing border from splash button is your feature request related to a problem please describe splash button is having borders of very odd colors which don t go good with the gradient inside of them describe the solution you d like after removing the borders it will have a good sleek n elegant look how the button currently looks how it ll look after removing wierd borders the gradient is animated in this button so it looks good with no borders additional notes i ll make a pull request as soon as possible if assigned | 1 |
731,994 | 25,240,318,297 | IssuesEvent | 2022-11-15 06:43:45 | OpenMined/PySyft | https://api.github.com/repos/OpenMined/PySyft | closed | Login with an URL doesn't work without specifying port | Type: Improvement :chart_with_upwards_trend: Priority: 3 - Medium :unamused: | ## Description
When specifying an URL, we need to add the port to both the URL and port parameters. This is not a big deal as the error output helps with that but it might be a better UX to remove the one or the other.
## How to Reproduce
Start a domain with `hagrid launch --tag=latest` and run
```
import numpy as np
import syft as sy
client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0:8081")
client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0")
```
## Expected Behavior
If we already specify the port as a parameter we shouldn't specify it in the URL as well.
## Screenshots
This is the output when running the code:

## Additional Context
This is relevant when connecting to a domain over the Internet.
| 1.0 | Login with an URL doesn't work without specifying port - ## Description
When specifying an URL, we need to add the port to both the URL and port parameters. This is not a big deal as the error output helps with that but it might be a better UX to remove the one or the other.
## How to Reproduce
Start a domain with `hagrid launch --tag=latest` and run
```
import numpy as np
import syft as sy
client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0:8081")
client = sy.login(email="info@openmined.org", password="changethis", port=8081,url="http://0.0.0.0")
```
## Expected Behavior
If we already specify the port as a parameter we shouldn't specify it in the URL as well.
## Screenshots
This is the output when running the code:

## Additional Context
This is relevant when connecting to a domain over the Internet.
| priority | login with an url doesn t work without specifying port description when specifying an url we need to add the port to both the url and port parameters this is not a big deal as the error output helps with that but it might be a better ux to remove the one or the other how to reproduce start a domain with hagrid launch tag latest and run import numpy as np import syft as sy client sy login email info openmined org password changethis port url client sy login email info openmined org password changethis port url expected behavior if we already specify the port as a parameter we shouldn t specify it in the url as well screenshots this is the output when running the code additional context this is relevant when connecting to a domain over the internet | 1 |
384,637 | 11,395,727,632 | IssuesEvent | 2020-01-30 12:04:42 | luna/enso | https://api.github.com/repos/luna/enso | closed | File Manager — Rust Client Library | Category: Libraries Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement | ### Summary
Provide a Rust library that can connect to the file manager server (#395) and exposes its API.
### Value
Being able to use file manager service is necessary for many components in GUI, like the filesystem tree view.
### Specification
* connects to the websocket service
* implements the file manager websocket protocol
* wraps in in the methods that are easily callable from Rust
* can target wasm
### Acceptance Criteria & Test Cases
Rust code can access filesystem through the running File Manager server. | 1.0 | File Manager — Rust Client Library - ### Summary
Provide a Rust library that can connect to the file manager server (#395) and exposes its API.
### Value
Being able to use file manager service is necessary for many components in GUI, like the filesystem tree view.
### Specification
* connects to the websocket service
* implements the file manager websocket protocol
* wraps in in the methods that are easily callable from Rust
* can target wasm
### Acceptance Criteria & Test Cases
Rust code can access filesystem through the running File Manager server. | priority | file manager — rust client library summary provide a rust library that can connect to the file manager server and exposes its api value being able to use file manager service is necessary for many components in gui like the filesystem tree view specification connects to the websocket service implements the file manager websocket protocol wraps in in the methods that are easily callable from rust can target wasm acceptance criteria test cases rust code can access filesystem through the running file manager server | 1 |
255,079 | 8,108,257,694 | IssuesEvent | 2018-08-14 00:35:10 | MARKETProtocol/service-listener | https://api.github.com/repos/MARKETProtocol/service-listener | closed | Implement events for contract removal | Priority: Medium To Be Funded Type: Enhancement | <!--
Hello!
Please use the template below for issue ideas or bugs found within MARKET Protocol.
If it is general support you need, reach out to us at
https://marketprotocol.io/discord
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
<!--
## Before you `start work`
Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties)
If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible.
Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline.
Please also note that in order for work to be accepted, all code must be accompanied by test cases as well.
-->
### Description
Listen for event `AddressRemovedFromWhiteList` and create a new record in the database for the created contract using `MARKET.js` and the `MARKET` api.
### Definition of Done
[comment]: # (Any other information that would be useful, bullets are helpful.)
- [ ] should soft delete from database using `market-api`(sets is_whitelisted to false)
- [ ] unit tests
### Additional Information
[comment]: # (Any other information that would be useful, content, screenshots, etc.)
point of contact @eswarasai or @travisdmathis | 1.0 | Implement events for contract removal - <!--
Hello!
Please use the template below for issue ideas or bugs found within MARKET Protocol.
If it is general support you need, reach out to us at
https://marketprotocol.io/discord
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
<!--
## Before you `start work`
Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties)
If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible.
Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline.
Please also note that in order for work to be accepted, all code must be accompanied by test cases as well.
-->
### Description
Listen for event `AddressRemovedFromWhiteList` and create a new record in the database for the created contract using `MARKET.js` and the `MARKET` api.
### Definition of Done
[comment]: # (Any other information that would be useful, bullets are helpful.)
- [ ] should soft delete from database using `market-api`(sets is_whitelisted to false)
- [ ] unit tests
### Additional Information
[comment]: # (Any other information that would be useful, content, screenshots, etc.)
point of contact @eswarasai or @travisdmathis | priority | implement events for contract removal hello please use the template below for issue ideas or bugs found within market protocol if it is general support you need reach out to us at provide a general summary of the issue in the title above and use relevant fields below to define the problem before you start work please read our contribution and if there is a bounty involved please also see if you have ongoing work from other bounties with us where funding has not been released please do not pick up a new issue we would like to involve as many contributors as possible and parallelize the work flow as much as possible please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline please also note that in order for work to be accepted all code must be accompanied by test cases as well description listen for event addressremovedfromwhitelist and create a new record in the database for the created contract using market js and the market api definition of done any other information that would be useful bullets are helpful should soft delete from database using market api sets is whitelisted to false unit tests additional information any other information that would be useful content screenshots etc point of contact eswarasai or travisdmathis | 1 |
508,980 | 14,709,968,746 | IssuesEvent | 2021-01-05 03:50:10 | syrus-bot/syrus-bot | https://api.github.com/repos/syrus-bot/syrus-bot | opened | feat(commands:core): finish guild settings commands | feat: core priority: medium | **Is your feature request related to a problem? Please describe.**
Currently, modification of guild settings is not handled properly. You cannot set multiple array values, you cannot easily change object values, and things are improperly documented.
**Describe the solution you'd like**
- [ ] Properly document settings modification
- [ ] (?) Add smart recursive lookup nesting
- [ ] Extend `GuildSchema` on `@data/MongoDB` to include log channel for #52
**Describe alternatives you've considered**
N/A.
**Additional context**
N/A. | 1.0 | feat(commands:core): finish guild settings commands - **Is your feature request related to a problem? Please describe.**
Currently, modification of guild settings is not handled properly. You cannot set multiple array values, you cannot easily change object values, and things are improperly documented.
**Describe the solution you'd like**
- [ ] Properly document settings modification
- [ ] (?) Add smart recursive lookup nesting
- [ ] Extend `GuildSchema` on `@data/MongoDB` to include log channel for #52
**Describe alternatives you've considered**
N/A.
**Additional context**
N/A. | priority | feat commands core finish guild settings commands is your feature request related to a problem please describe currently modification of guild settings is not handled properly you cannot set multiple array values you cannot easily change object values and things are improperly documented describe the solution you d like properly document settings modification add smart recursive lookup nesting extend guildschema on data mongodb to include log channel for describe alternatives you ve considered n a additional context n a | 1 |
185,350 | 6,721,150,594 | IssuesEvent | 2017-10-16 10:30:56 | CS2103AUG2017-T17-B1/main | https://api.github.com/repos/CS2103AUG2017-T17-B1/main | opened | As a debt collector I want to have the ability to perform validity checks on client's personal infomation(HP, email, address) | priority.medium | so that I can verify that their data is genuine | 1.0 | As a debt collector I want to have the ability to perform validity checks on client's personal infomation(HP, email, address) - so that I can verify that their data is genuine | priority | as a debt collector i want to have the ability to perform validity checks on client s personal infomation hp email address so that i can verify that their data is genuine | 1 |
593,256 | 17,953,572,222 | IssuesEvent | 2021-09-13 03:03:34 | francheska-vicente/cssweng | https://api.github.com/repos/francheska-vicente/cssweng | closed | Validation and design changes for mobile number is not present in Check-in form | bug priority: medium severity: medium issue: validation | ### Summary:
Previously implemented validation checks for mobile number (length, starting number) is absent from this form; the change from contact number to mobile number and the placeholder text is also missing
### Steps to Reproduce:
- Proceed to Check-in form
- Enter 123 for Mobile Number
- Click Check-in
### Visual Proof:

### Expected Results:
- Error Message
### Actual Results:
- Invalid input is accepted
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 | | 1.0 | Validation and design changes for mobile number is not present in Check-in form - ### Summary:
Previously implemented validation checks for mobile number (length, starting number) is absent from this form; the change from contact number to mobile number and the placeholder text is also missing
### Steps to Reproduce:
- Proceed to Check-in form
- Enter 123 for Mobile Number
- Click Check-in
### Visual Proof:

### Expected Results:
- Error Message
### Actual Results:
- Invalid input is accepted
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 | | priority | validation and design changes for mobile number is not present in check in form summary previously implemented validation checks for mobile number length starting number is absent from this form the change from contact number to mobile number and the placeholder text is also missing steps to reproduce proceed to check in form enter for mobile number click check in visual proof expected results error message actual results invalid input is accepted additional information platform engine google operating system windows | 1 |
632,031 | 20,168,894,381 | IssuesEvent | 2022-02-10 08:33:37 | TencentBlueKing/bk-iam-saas | https://api.github.com/repos/TencentBlueKing/bk-iam-saas | closed | [日志] 批量优化日志打印 | Type: Enhancement Layer: SaaS Priority: Medium Size: M | 背景: 目前SaaS的logger需要重新处理, 打的信息过少, 导致信息缺失毫无意义; 要么去掉, 要么打完整; 操作人和被操作对象, 时间, 操作, 来源入口等等
1. Django 日志, info 日志目前只有部分信息,缺乏更全面的信息
例如, 谁 从 哪个group 删除了 谁
2. 什么地方该打日志, 什么地方不应该打? 需要从现有线上日志分类确定下, 去掉高频无用的日志, 优化关键路径的日志
3. 增加layer字段快速定位, 用于定位哪一层打的日志 | 1.0 | [日志] 批量优化日志打印 - 背景: 目前SaaS的logger需要重新处理, 打的信息过少, 导致信息缺失毫无意义; 要么去掉, 要么打完整; 操作人和被操作对象, 时间, 操作, 来源入口等等
1. Django 日志, info 日志目前只有部分信息,缺乏更全面的信息
例如, 谁 从 哪个group 删除了 谁
2. 什么地方该打日志, 什么地方不应该打? 需要从现有线上日志分类确定下, 去掉高频无用的日志, 优化关键路径的日志
3. 增加layer字段快速定位, 用于定位哪一层打的日志 | priority | 批量优化日志打印 背景 目前saas的logger需要重新处理 打的信息过少 导致信息缺失毫无意义 要么去掉 要么打完整 操作人和被操作对象 时间 操作 来源入口等等 django 日志 info 日志目前只有部分信息 缺乏更全面的信息 例如 谁 从 哪个group 删除了 谁 什么地方该打日志 什么地方不应该打 需要从现有线上日志分类确定下 去掉高频无用的日志 优化关键路径的日志 增加layer字段快速定位 用于定位哪一层打的日志 | 1 |
29,494 | 2,716,237,964 | IssuesEvent | 2015-04-10 17:51:34 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | The RemoteStreamDataSource is duplicating the last record | bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.2 | _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 13:48:42_
The RemoteStreamingDataSource is duplicating the last record on a page when the number of records = PageSize - 1:
Example:
If a page can hold five records, but there is only four records, the fourth record will be duplicated to fill the page.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=416_ | 1.0 | The RemoteStreamDataSource is duplicating the last record - _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on July 02, 2014 13:48:42_
The RemoteStreamingDataSource is duplicating the last record on a page when the number of records = PageSize - 1:
Example:
If a page can hold five records, but there is only four records, the fourth record will be duplicated to fill the page.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=416_ | priority | the remotestreamdatasource is duplicating the last record from on july the remotestreamingdatasource is duplicating the last record on a page when the number of records pagesize example if a page can hold five records but there is only four records the fourth record will be duplicated to fill the page original issue | 1 |
390,966 | 11,566,600,838 | IssuesEvent | 2020-02-20 12:49:50 | robotology/human-dynamics-estimation | https://api.github.com/repos/robotology/human-dynamics-estimation | closed | Add option to express net external wrench estimates of dummy source (hands) with orientation of world frame | complexity:medium component:HumanDynamicsEstimation component:HumanWrenchProvider priority:high type:enhancement type:task | Currently, the force-torque measurements from the ftShoes are expressed (both origin and orientation) with respect the human foot frames (`LeftFoot` and `RightFoot`). So, on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117), the net external wrench estimates on `LeftFoot` and `RightFoot` links are correctly obtained in the body frame, and if the covariances are correctly set for the MAP estimator, the measurements and the estimates on `LeftFoot` and `RightFoot` links match closely. So, there is no need to modify the function [extractLinkNetExternalWrenchesFromDynamicVariables()](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) for `LeftFoot` and `RightFoot` links.
Now, coming to the case of the links `LeftHand` and `RightHand`, they are considered to be dummy sources of force-torques measurements (set to **0**). The net external wrench estimates for `LeftHand` and `RightHand` obtained on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) are expressed (both origin and orientation) in their body frames. To highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands, it is useful to express them at the origin of the links `LeftHand` and `RightHand` but with the orientation of the world frame.
As pointed out by @traversaro this code is best suited on the front end of HDE rather than in the back end of Berdy in iDynTree.
One of the problems in achieving this is to know inside `HumanDynamicsEstimator` device which link has a dummy wrench source attached. Currently, this information is present in `HumanWrenchProvider` device https://github.com/robotology/human-dynamics-estimation/blob/feature/visualize-berdy-estimated-wrench/conf/xml/Human.xml#L261 but it is not propagated to `HumnaDynamicsEstimator` device.
This issue will track the details related to updating HDE code for expressing the estimated net external wrench of dummy sources with the orientation of world frame.
@lrapetti @claudia-lat @traversaro | 1.0 | Add option to express net external wrench estimates of dummy source (hands) with orientation of world frame - Currently, the force-torque measurements from the ftShoes are expressed (both origin and orientation) with respect the human foot frames (`LeftFoot` and `RightFoot`). So, on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117), the net external wrench estimates on `LeftFoot` and `RightFoot` links are correctly obtained in the body frame, and if the covariances are correctly set for the MAP estimator, the measurements and the estimates on `LeftFoot` and `RightFoot` links match closely. So, there is no need to modify the function [extractLinkNetExternalWrenchesFromDynamicVariables()](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) for `LeftFoot` and `RightFoot` links.
Now, coming to the case of the links `LeftHand` and `RightHand`, they are considered to be dummy sources of force-torques measurements (set to **0**). The net external wrench estimates for `LeftHand` and `RightHand` obtained on calling [extractLinkNetExternalWrenchesFromDynamicVariables(const VectorDynSize& d, LinkNetExternalWrenches& netExtWrenches, const bool task1)](https://github.com/robotology/idyntree/blob/master/src/estimation/src/BerdyHelper.cpp#L2117) are expressed (both origin and orientation) in their body frames. To highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands, it is useful to express them at the origin of the links `LeftHand` and `RightHand` but with the orientation of the world frame.
As pointed out by @traversaro this code is best suited on the front end of HDE rather than in the back end of Berdy in iDynTree.
One of the problems in achieving this is to know inside `HumanDynamicsEstimator` device which link has a dummy wrench source attached. Currently, this information is present in `HumanWrenchProvider` device https://github.com/robotology/human-dynamics-estimation/blob/feature/visualize-berdy-estimated-wrench/conf/xml/Human.xml#L261 but it is not propagated to `HumnaDynamicsEstimator` device.
This issue will track the details related to updating HDE code for expressing the estimated net external wrench of dummy sources with the orientation of world frame.
@lrapetti @claudia-lat @traversaro | priority | add option to express net external wrench estimates of dummy source hands with orientation of world frame currently the force torque measurements from the ftshoes are expressed both origin and orientation with respect the human foot frames leftfoot and rightfoot so on calling the net external wrench estimates on leftfoot and rightfoot links are correctly obtained in the body frame and if the covariances are correctly set for the map estimator the measurements and the estimates on leftfoot and rightfoot links match closely so there is no need to modify the function for leftfoot and rightfoot links now coming to the case of the links lefthand and righthand they are considered to be dummy sources of force torques measurements set to the net external wrench estimates for lefthand and righthand obtained on calling are expressed both origin and orientation in their body frames to highlight the fact that these estimates at the hands are a reflection of the estimates of the object weight at hands it is useful to express them at the origin of the links lefthand and righthand but with the orientation of the world frame as pointed out by traversaro this code is best suited on the front end of hde rather than in the back end of berdy in idyntree one of the problems in achieving this is to know inside humandynamicsestimator device which link has a dummy wrench source attached currently this information is present in humanwrenchprovider device but it is not propagated to humnadynamicsestimator device this issue will track the details related to updating hde code for expressing the estimated net external wrench of dummy sources with the orientation of world frame lrapetti claudia lat traversaro | 1 |
122,649 | 4,838,540,124 | IssuesEvent | 2016-11-09 04:05:12 | smirkspace/smirkspace | https://api.github.com/repos/smirkspace/smirkspace | opened | Add Analytics | backend enhancement medium priority | Add Analytics such as how many times a button is pressed, how much time spent on a page, etc. Possibly use MixPanel for this. | 1.0 | Add Analytics - Add Analytics such as how many times a button is pressed, how much time spent on a page, etc. Possibly use MixPanel for this. | priority | add analytics add analytics such as how many times a button is pressed how much time spent on a page etc possibly use mixpanel for this | 1 |
56,247 | 3,078,627,002 | IssuesEvent | 2015-08-21 11:38:18 | nfprojects/nfengine | https://api.github.com/repos/nfprojects/nfengine | closed | Remove all useless #if ... #else ... #endif sections | bug high priority medium | Some parts of engine have code hidden by #if ... #else ... #endif sequence. Search for all of them and either remove them, or provide different way to determine which section to use (avoid preprocessor macros, we want the engine to be entirely compiled). | 1.0 | Remove all useless #if ... #else ... #endif sections - Some parts of engine have code hidden by #if ... #else ... #endif sequence. Search for all of them and either remove them, or provide different way to determine which section to use (avoid preprocessor macros, we want the engine to be entirely compiled). | priority | remove all useless if else endif sections some parts of engine have code hidden by if else endif sequence search for all of them and either remove them or provide different way to determine which section to use avoid preprocessor macros we want the engine to be entirely compiled | 1 |
376,160 | 11,139,228,124 | IssuesEvent | 2019-12-21 03:06:46 | rstudio/gt | https://api.github.com/repos/rstudio/gt | closed | Error in values[index] (only 0's may be mixed with negative subscripts) when one variable in df contains only zeroes | Difficulty: ② Intermediate Effort: ② Medium Priority: ♨︎ Critical Type: ☹︎ Bug | First of all, let me say congrats for this awseome package gt you are building! :-)
Issue detected: If you have a dataframe more than one variable, and one of them contains only empty values, you end up hitting this error:
`Error in values[index] : only 0's may be mixed with negative subscripts`
(using R 3.6.x under GNU/Linux 64 bits, in case it matters)
Reproducible example:
```r
data.frame(x=rnorm(10), y=rep(0,10)) %>%
gt() %>%
fmt_number(
columns = vars(x,
y),
decimals = 2,
suffixing = TRUE
)
```
FYI, If you attempt to make the similar reprex but only with one column (y in the previous example), then no error is triggered and the expected result is produced.
```r
data.frame(y=rep(0,10)) %>%
gt() %>%
fmt_number(
columns = vars(y),
decimals = 2,
suffixing = TRUE
)
``` | 1.0 | Error in values[index] (only 0's may be mixed with negative subscripts) when one variable in df contains only zeroes - First of all, let me say congrats for this awseome package gt you are building! :-)
Issue detected: If you have a dataframe more than one variable, and one of them contains only empty values, you end up hitting this error:
`Error in values[index] : only 0's may be mixed with negative subscripts`
(using R 3.6.x under GNU/Linux 64 bits, in case it matters)
Reproducible example:
```r
data.frame(x=rnorm(10), y=rep(0,10)) %>%
gt() %>%
fmt_number(
columns = vars(x,
y),
decimals = 2,
suffixing = TRUE
)
```
FYI, If you attempt to make the similar reprex but only with one column (y in the previous example), then no error is triggered and the expected result is produced.
```r
data.frame(y=rep(0,10)) %>%
gt() %>%
fmt_number(
columns = vars(y),
decimals = 2,
suffixing = TRUE
)
``` | priority | error in values only s may be mixed with negative subscripts when one variable in df contains only zeroes first of all let me say congrats for this awseome package gt you are building issue detected if you have a dataframe more than one variable and one of them contains only empty values you end up hitting this error error in values only s may be mixed with negative subscripts using r x under gnu linux bits in case it matters reproducible example r data frame x rnorm y rep gt fmt number columns vars x y decimals suffixing true fyi if you attempt to make the similar reprex but only with one column y in the previous example then no error is triggered and the expected result is produced r data frame y rep gt fmt number columns vars y decimals suffixing true | 1 |
44,318 | 2,903,519,828 | IssuesEvent | 2015-06-18 13:51:23 | CenterForOpenScience/osf.io | https://api.github.com/repos/CenterForOpenScience/osf.io | closed | [Staging & Production] Dataverse: Delete fails when trying to delete a draft file | 2 - ready bug: production priority - medium | ## Steps
1. Go to a public project with Dataverse authorized (link: https://osf.io/9jtkg/ )
2. Go to the files tab
3. Switch the Dataverse folder from Released to Draft
4. Wait for the folder to re-load with draft files
5. Click on the red 'x' next to a file to delete it
6. Hit 'OK' to confirm
## Expected
That a dataverse draft file can be deleted from the files tab
## Actual
There is an error in the file grid stating that 'Delete failed'. Also the console reports a 500 error.

| 1.0 | [Staging & Production] Dataverse: Delete fails when trying to delete a draft file - ## Steps
1. Go to a public project with Dataverse authorized (link: https://osf.io/9jtkg/ )
2. Go to the files tab
3. Switch the Dataverse folder from Released to Draft
4. Wait for the folder to re-load with draft files
5. Click on the red 'x' next to a file to delete it
6. Hit 'OK' to confirm
## Expected
That a dataverse draft file can be deleted from the files tab
## Actual
There is an error in the file grid stating that 'Delete failed'. Also the console reports a 500 error.

| priority | dataverse delete fails when trying to delete a draft file steps go to a public project with dataverse authorized link go to the files tab switch the dataverse folder from released to draft wait for the folder to re load with draft files click on the red x next to a file to delete it hit ok to confirm expected that a dataverse draft file can be deleted from the files tab actual there is an error in the file grid stating that delete failed also the console reports a error | 1 |
26,330 | 2,684,319,085 | IssuesEvent | 2015-03-28 21:28:10 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | Crashes when I press Win+Right or Win+Left | 1 star bug duplicate imported Priority-Medium | _From [CoreyJJo...@gmail.com](https://code.google.com/u/117850621500591170194/) on March 19, 2013 07:08:08_
Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* ... Crashes when I press Win+Right or Win+Left to move the window on Win7 Pro 64-bit *Steps to reproduction* 1. 2. 3.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=993_ | 1.0 | Crashes when I press Win+Right or Win+Left - _From [CoreyJJo...@gmail.com](https://code.google.com/u/117850621500591170194/) on March 19, 2013 07:08:08_
Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* ... Crashes when I press Win+Right or Win+Left to move the window on Win7 Pro 64-bit *Steps to reproduction* 1. 2. 3.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=993_ | priority | crashes when i press win right or win left from on march required information os version winxp vista sp conemu version far version if you are using far manager bug description crashes when i press win right or win left to move the window on pro bit steps to reproduction original issue | 1 |
223,118 | 7,446,984,076 | IssuesEvent | 2018-03-28 10:55:35 | vanilla-framework/vanilla-framework | https://api.github.com/repos/vanilla-framework/vanilla-framework | closed | Margin top on br in Firefox | Priority: Medium Type: Bug | Go to: https://snapcraft.io/spotify
There is unnecessary spacing on bullet points, because Vanilla adds top margin to `br`.
Related to: #1626
Also reported in snapcraft: https://github.com/canonical-websites/snapcraft.io/issues/405
What you see in Chromium: | What you see in Firefox:
-|-
 | 
What you should see: The description in Firefox should not have all the extra line spacing. | 1.0 | Margin top on br in Firefox - Go to: https://snapcraft.io/spotify
There is unnecessary spacing on bullet points, because Vanilla adds top margin to `br`.
Related to: #1626
Also reported in snapcraft: https://github.com/canonical-websites/snapcraft.io/issues/405
What you see in Chromium: | What you see in Firefox:
-|-
 | 
What you should see: The description in Firefox should not have all the extra line spacing. | priority | margin top on br in firefox go to there is unnecessary spacing on bullet points because vanilla adds top margin to br related to also reported in snapcraft what you see in chromium what you see in firefox what you should see the description in firefox should not have all the extra line spacing | 1 |
759,903 | 26,617,638,635 | IssuesEvent | 2023-01-24 08:45:23 | apache/airflow | https://api.github.com/repos/apache/airflow | closed | graph disappears during run time when using branch_task and a dynamic classic operator | kind:bug good first issue priority:medium area:UI | ### Apache Airflow version
2.5.1
### What happened
when using a dynamically generated task that gets the expand data from xcom after a branch_task the graph doesn't render.
It reappears once the dag run is finished.
tried with BashOperator and a KubernetesPodOperator.
the developer console in the browser shows the error:
`Uncaught TypeError: Cannot read properties of undefined (reading 'length')
at z (graph.1c0596dfced26c638bfe.js:2:17499)
at graph.1c0596dfced26c638bfe.js:2:17654
at Array.map (<anonymous>)
at z (graph.1c0596dfced26c638bfe.js:2:17646)
at graph.1c0596dfced26c638bfe.js:2:26602
at graph.1c0596dfced26c638bfe.js:2:26655
at graph.1c0596dfced26c638bfe.js:2:26661
at graph.1c0596dfced26c638bfe.js:2:222
at graph.1c0596dfced26c638bfe.js:2:227
z @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
z @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
`
grid view renders fine.
### What you think should happen instead
graph should be rendered.
### How to reproduce
```@dag('branch_dynamic', schedule_interval=None, default_args=default_args, catchup=False)
def branch_dynamic_flow():
@branch_task
def choose_path():
return 'b'
@task
def a():
print('a')
@task
def get_args():
return ['echo 1', 'echo 2']
b = BashOperator.partial(task_id="b").expand(bash_command=get_args())
path = choose_path()
path >> a()
path >> b
```
### Operating System
red hat
### Versions of Apache Airflow Providers
apache-airflow-providers-cncf-kubernetes | 5.1.1 | Kubernetes
### Deployment
Official Apache Airflow Helm Chart
### Deployment details
_No response_
### Anything else
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
| 1.0 | graph disappears during run time when using branch_task and a dynamic classic operator - ### Apache Airflow version
2.5.1
### What happened
when using a dynamically generated task that gets the expand data from xcom after a branch_task the graph doesn't render.
It reappears once the dag run is finished.
tried with BashOperator and a KubernetesPodOperator.
the developer console in the browser shows the error:
`Uncaught TypeError: Cannot read properties of undefined (reading 'length')
at z (graph.1c0596dfced26c638bfe.js:2:17499)
at graph.1c0596dfced26c638bfe.js:2:17654
at Array.map (<anonymous>)
at z (graph.1c0596dfced26c638bfe.js:2:17646)
at graph.1c0596dfced26c638bfe.js:2:26602
at graph.1c0596dfced26c638bfe.js:2:26655
at graph.1c0596dfced26c638bfe.js:2:26661
at graph.1c0596dfced26c638bfe.js:2:222
at graph.1c0596dfced26c638bfe.js:2:227
z @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
z @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
(anonymous) @ graph.1c0596dfced26c638bfe.js:2
`
grid view renders fine.
### What you think should happen instead
graph should be rendered.
### How to reproduce
```@dag('branch_dynamic', schedule_interval=None, default_args=default_args, catchup=False)
def branch_dynamic_flow():
@branch_task
def choose_path():
return 'b'
@task
def a():
print('a')
@task
def get_args():
return ['echo 1', 'echo 2']
b = BashOperator.partial(task_id="b").expand(bash_command=get_args())
path = choose_path()
path >> a()
path >> b
```
### Operating System
red hat
### Versions of Apache Airflow Providers
apache-airflow-providers-cncf-kubernetes | 5.1.1 | Kubernetes
### Deployment
Official Apache Airflow Helm Chart
### Deployment details
_No response_
### Anything else
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
| priority | graph disappears during run time when using branch task and a dynamic classic operator apache airflow version what happened when using a dynamically generated task that gets the expand data from xcom after a branch task the graph doesn t render it reappears once the dag run is finished tried with bashoperator and a kubernetespodoperator the developer console in the browser shows the error uncaught typeerror cannot read properties of undefined reading length at z graph js at graph js at array map at z graph js at graph js at graph js at graph js at graph js at graph js z graph js anonymous graph js z graph js anonymous graph js anonymous graph js anonymous graph js anonymous graph js anonymous graph js grid view renders fine what you think should happen instead graph should be rendered how to reproduce dag branch dynamic schedule interval none default args default args catchup false def branch dynamic flow branch task def choose path return b task def a print a task def get args return b bashoperator partial task id b expand bash command get args path choose path path a path b operating system red hat versions of apache airflow providers apache airflow providers cncf kubernetes kubernetes deployment official apache airflow helm chart deployment details no response anything else no response are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s | 1 |
438,847 | 12,652,489,481 | IssuesEvent | 2020-06-17 03:45:29 | Open-Source-Medical-Supplies/react-projects-gallery | https://api.github.com/repos/Open-Source-Medical-Supplies/react-projects-gallery | closed | Full card: semantics per Kimly | correction medium priority | 1. 'Reviewed By', text should be, "Medical Status:"
2. the text from the tags 'Reviewed By' does need to be shown, ex 'Affiliated with: UW Health'
3. Text of Medical Supply Category should, be bold or somehow different from the Title Header Style, and link back to category supply page
4. In Materials List, underneath Material Name, should be hyper formatted text from 'Detail', base 'Bill of Materials' | 1.0 | Full card: semantics per Kimly - 1. 'Reviewed By', text should be, "Medical Status:"
2. the text from the tags 'Reviewed By' does need to be shown, ex 'Affiliated with: UW Health'
3. Text of Medical Supply Category should, be bold or somehow different from the Title Header Style, and link back to category supply page
4. In Materials List, underneath Material Name, should be hyper formatted text from 'Detail', base 'Bill of Materials' | priority | full card semantics per kimly reviewed by text should be medical status the text from the tags reviewed by does need to be shown ex affiliated with uw health text of medical supply category should be bold or somehow different from the title header style and link back to category supply page in materials list underneath material name should be hyper formatted text from detail base bill of materials | 1 |
670,180 | 22,678,837,531 | IssuesEvent | 2022-07-04 08:03:54 | input-output-hk/cardano-graphql | https://api.github.com/repos/input-output-hk/cardano-graphql | closed | Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes. | BUG SEVERITY:LOW PRIORITY:MEDIUM | ### Environment
Testnet https://gimbalabs.com/dandelion/endpoints/graphql-api
### Steps to reproduce the bug
Send well formatted transaction with:
```graphql
mutation submitTransaction(
$transaction: String!
) {
submitTransaction(transaction: $transaction) {
hash
}
}
```
I got the error: `Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes.`
But the transaction is actually send to network: https://explorer.cardano-testnet.iohkdev.io/en/transaction?id=6f5762b2e52cd8f81380e197daae2ab6bded8db34e62346c3434085c0b6b27cd
### What is the expected behavior?
No false negative error. | 1.0 | Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes. - ### Environment
Testnet https://gimbalabs.com/dandelion/endpoints/graphql-api
### Steps to reproduce the bug
Send well formatted transaction with:
```graphql
mutation submitTransaction(
$transaction: String!
) {
submitTransaction(transaction: $transaction) {
hash
}
}
```
I got the error: `Deserialization failed in Transaction because: Invalid cbor: not enough bytes, expect 0 bytes but received 0 bytes.`
But the transaction is actually send to network: https://explorer.cardano-testnet.iohkdev.io/en/transaction?id=6f5762b2e52cd8f81380e197daae2ab6bded8db34e62346c3434085c0b6b27cd
### What is the expected behavior?
No false negative error. | priority | deserialization failed in transaction because invalid cbor not enough bytes expect bytes but received bytes environment testnet steps to reproduce the bug send well formatted transaction with graphql mutation submittransaction transaction string submittransaction transaction transaction hash i got the error deserialization failed in transaction because invalid cbor not enough bytes expect bytes but received bytes but the transaction is actually send to network what is the expected behavior no false negative error | 1 |
638,830 | 20,739,650,985 | IssuesEvent | 2022-03-14 16:31:03 | CDCgov/prime-reportstream | https://api.github.com/repos/CDCgov/prime-reportstream | closed | MD - Incorrect observation status of "X" in SimpleReport | onboarding-ops receiver data-issue Medium Priority | MD is also reporting receiving "X" in OBR-25 under observation status instead of P, F or C
Sample messages
Complete Care Severna Park 2/9/22 5:36 AM ```3a12a0df-d1ab-4556-a477-54ce808c188f```
sanchez pediatrics 2/10/22 11:26 AM ```ae25a69f-5726-48f3-b48d-1192d743e022```
ResultsLab,LLC 2/14/22 5:30 AM ```7a9e18e3-63c0-435a-9da5-a44d3fa90833```
Brightview White Marsh 2/8/22 2:16 PM ```ae25a69f-5726-48f3-b48d-1192d743e022``` | 1.0 | MD - Incorrect observation status of "X" in SimpleReport - MD is also reporting receiving "X" in OBR-25 under observation status instead of P, F or C
Sample messages
Complete Care Severna Park 2/9/22 5:36 AM ```3a12a0df-d1ab-4556-a477-54ce808c188f```
sanchez pediatrics 2/10/22 11:26 AM ```ae25a69f-5726-48f3-b48d-1192d743e022```
ResultsLab,LLC 2/14/22 5:30 AM ```7a9e18e3-63c0-435a-9da5-a44d3fa90833```
Brightview White Marsh 2/8/22 2:16 PM ```ae25a69f-5726-48f3-b48d-1192d743e022``` | priority | md incorrect observation status of x in simplereport md is also reporting receiving x in obr under observation status instead of p f or c sample messages complete care severna park am sanchez pediatrics am resultslab llc am brightview white marsh pm | 1 |
302,676 | 9,285,331,388 | IssuesEvent | 2019-03-21 06:37:18 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | tests/kernel/threads/no-multithreading/testcase.yam: Not Booting | area: NIOS2 bug priority: medium | **Describe the bug**
Failed to boot up on altera_max10 board
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=altera_max10
3. make
4. make run
**Expected behavior**
Should able to boot zephyr and show log on console output.
**Screenshots or console output**
no console output.
OS: fedora29
Toolchain: zephyr
Commit SHA or Version used: 6368ed4 | 1.0 | tests/kernel/threads/no-multithreading/testcase.yam: Not Booting - **Describe the bug**
Failed to boot up on altera_max10 board
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=altera_max10
3. make
4. make run
**Expected behavior**
Should able to boot zephyr and show log on console output.
**Screenshots or console output**
no console output.
OS: fedora29
Toolchain: zephyr
Commit SHA or Version used: 6368ed4 | priority | tests kernel threads no multithreading testcase yam not booting describe the bug failed to boot up on altera board to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard altera make make run expected behavior should able to boot zephyr and show log on console output screenshots or console output no console output os toolchain zephyr commit sha or version used | 1 |
228,719 | 7,567,005,138 | IssuesEvent | 2018-04-22 04:03:34 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] UsersAdminComponent has a templateUrl, include it. Cannot read template for UsersAdminComponent | 1. Type: Bug Priority: Medium Status: Accepted Status: Waiting feedback Time: ~1 hour | ##### **Overview of the issue**
I am unable to run it for templates
the application compiles and runs perfectly. the template is in the same folder as component
**I get the following error:**
UserPreferencesComponent has a templateUrl, include it
[15:13:37] UserProfileComponent has a templateUrl, include it
[15:13:37] UserRegistrationComponent has a templateUrl, include it
[15:13:37] UserSigninComponent has a templateUrl, include it
[15:13:37] UserVerificationComponent has a templateUrl, include it
[15:13:37] UsersAdminComponent has a templateUrl, include it
[15:13:37] Cannot read template for UsersAdminComponent
##### **Operating System, Node.js, npm, compodoc version(s)**
angular/cli: 1.4.6
node: 6.11.3
os: darwin x64
angular/animations: 5.1.2
angular/cdk: 5.0.2
angular/common: 5.1.2
angular/compiler: 5.1.2
angular/core: 5.1.2
angular/forms: 5.1.2
angular/http: 5.1.2
angular/material: 5.0.2
angular/platform-browser: 5.1.2
angular/platform-browser-dynamic: 5.1.2
angular/router: 5.1.2
angular/cli: 1.4.6
angular/compiler-cli: 5.1.2
angular/language-service: 5.1.2
typescript: 2.5.3
##### **Angular configuration, a `package.json` file in the root folder**
**users-admin.component.ts**
```
@Component({
selector: 'app-users-admin',
templateUrl: './users-admin.component.html',
styleUrls: ['./users-admin.component.scss']
})
export class UsersAdminComponent implements OnInit {
```
**users-admin.component.html**
```
<p>
<button mat-raised-button color="primary" [routerLink]="['/users/atsadmin']">Back to Dashboard</button>
</p>
<p>
users-admin works!
</p>
<div class="master-detail-container">
<div class="master-detail-header">
</div>
<div class="master-detail">
<div class="master">
<mat-card>
<mat-card-title-group>
<mat-card-title>Users</mat-card-title>
<mat-card-subtitle>Subtitle</mat-card-subtitle>
<img matCardImage src="Side image">
</mat-card-title-group>
<img matCardImage src="Card wide image source">
<mat-card-content>
<mat-list dense>
<mat-list-item *ngFor="let user of users; let i=index" (click)="switchUser(i)">
{{user.firstName}} {{user.lastName}}
</mat-list-item>
</mat-list>
</mat-card-content>
<mat-card-actions align="start">
<button mat-button matTooltip="Add User" (click)="onAction(event)">Add User</button>
<button mat-button matTooltip="Delete User" (click)="onAction(event)">Delete Users</button>
</mat-card-actions>
<mat-card-footer>
Footer
</mat-card-footer>
</mat-card>
</div>
<div class="detail">
<mat-card>
<mat-card-title-group>
<mat-card-title>User Details</mat-card-title>
<mat-card-subtitle>details</mat-card-subtitle>
<img matCardImage src="Side image">
</mat-card-title-group>
<img matCardImage src="Card wide image source">
<mat-card-content>
<form *ngIf="currentUser" [formGroup]="userDetailsFormGroup">
<mat-form-field>
<input matInput formControlName="firstName" type="text" placeholder="First Name" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['firstName'].valid &&
userDetailsFormGroup.controls['firstName'].touched &&
userDetailsFormGroup.controls['firstName'].errors.required">
First Name is required
</mat-error>
</mat-form-field>
<mat-form-field>
<input matInput formControlName="lastName" type="text" placeholder="Last Name" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['lastName'].valid &&
userDetailsFormGroup.controls['lastName'].touched &&
userDetailsFormGroup.controls['lastName'].errors.required">
Last Name is required
</mat-error>
</mat-form-field>
<mat-form-field>
<mat-select formControlName="roles" placeholder="Roles" multiple required>
<mat-option *ngFor="let role of rolesList" [value]=role.key>{{ role.displayedValue }}</mat-option>
</mat-select>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['roles'].valid &&
userDetailsFormGroup.controls['roles'].touched &&
userDetailsFormGroup.controls['roles'].errors.required">
Role is required
</mat-error>
</mat-form-field>
<mat-form-field>
<input matInput formControlName="email" type="email" placeholder="Email" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid &&
userDetailsFormGroup.controls['email'].touched &&
userDetailsFormGroup.controls['email'].errors.required">
Email is required
</mat-error>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid &&
userDetailsFormGroup.controls['email'].touched &&
userDetailsFormGroup.controls['email'].errors.email">
Email is not valid
</mat-error>
</mat-form-field>
<button mat-raised-button matTooltip="Add Course to this user" (click)="openSelectItemDialog()">Add Course</button>
</form>
</mat-card-content>
<mat-card-actions align="start">
<button mat-button matTooltip="Save User" (click)="saveUser(event)">Save User</button>
<button mat-button matTooltip="Delete User" (click)="deleteUser(event)">Delete User</button>
</mat-card-actions>
<mat-card-footer>
Footer
</mat-card-footer>
</mat-card>
</div>
<!-- end detail -->
</div>
<!-- end master detail -->
</div>
<!-- end master detail container-->
```
<!-- This is mandatory for bug reports. This will help us to replicate the scenario. -->
##### **Compodoc installed globally or locally ?**
Locally
##### **Motivation for or Use Case**
<Unable to resolve templates
##### **Reproduce the error**
<!-- an unambiguous set of steps to reproduce the error. -->
##### **Related issues**
<!-- has a similar issue been reported before? -->
##### **Suggest a Fix**
no idea
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | 1.0 | [BUG] UsersAdminComponent has a templateUrl, include it. Cannot read template for UsersAdminComponent - ##### **Overview of the issue**
I am unable to run it for templates
the application compiles and runs perfectly. the template is in the same folder as component
**I get the following error:**
UserPreferencesComponent has a templateUrl, include it
[15:13:37] UserProfileComponent has a templateUrl, include it
[15:13:37] UserRegistrationComponent has a templateUrl, include it
[15:13:37] UserSigninComponent has a templateUrl, include it
[15:13:37] UserVerificationComponent has a templateUrl, include it
[15:13:37] UsersAdminComponent has a templateUrl, include it
[15:13:37] Cannot read template for UsersAdminComponent
##### **Operating System, Node.js, npm, compodoc version(s)**
angular/cli: 1.4.6
node: 6.11.3
os: darwin x64
angular/animations: 5.1.2
angular/cdk: 5.0.2
angular/common: 5.1.2
angular/compiler: 5.1.2
angular/core: 5.1.2
angular/forms: 5.1.2
angular/http: 5.1.2
angular/material: 5.0.2
angular/platform-browser: 5.1.2
angular/platform-browser-dynamic: 5.1.2
angular/router: 5.1.2
angular/cli: 1.4.6
angular/compiler-cli: 5.1.2
angular/language-service: 5.1.2
typescript: 2.5.3
##### **Angular configuration, a `package.json` file in the root folder**
**users-admin.component.ts**
```
@Component({
selector: 'app-users-admin',
templateUrl: './users-admin.component.html',
styleUrls: ['./users-admin.component.scss']
})
export class UsersAdminComponent implements OnInit {
```
**users-admin.component.html**
```
<p>
<button mat-raised-button color="primary" [routerLink]="['/users/atsadmin']">Back to Dashboard</button>
</p>
<p>
users-admin works!
</p>
<div class="master-detail-container">
<div class="master-detail-header">
</div>
<div class="master-detail">
<div class="master">
<mat-card>
<mat-card-title-group>
<mat-card-title>Users</mat-card-title>
<mat-card-subtitle>Subtitle</mat-card-subtitle>
<img matCardImage src="Side image">
</mat-card-title-group>
<img matCardImage src="Card wide image source">
<mat-card-content>
<mat-list dense>
<mat-list-item *ngFor="let user of users; let i=index" (click)="switchUser(i)">
{{user.firstName}} {{user.lastName}}
</mat-list-item>
</mat-list>
</mat-card-content>
<mat-card-actions align="start">
<button mat-button matTooltip="Add User" (click)="onAction(event)">Add User</button>
<button mat-button matTooltip="Delete User" (click)="onAction(event)">Delete Users</button>
</mat-card-actions>
<mat-card-footer>
Footer
</mat-card-footer>
</mat-card>
</div>
<div class="detail">
<mat-card>
<mat-card-title-group>
<mat-card-title>User Details</mat-card-title>
<mat-card-subtitle>details</mat-card-subtitle>
<img matCardImage src="Side image">
</mat-card-title-group>
<img matCardImage src="Card wide image source">
<mat-card-content>
<form *ngIf="currentUser" [formGroup]="userDetailsFormGroup">
<mat-form-field>
<input matInput formControlName="firstName" type="text" placeholder="First Name" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['firstName'].valid &&
userDetailsFormGroup.controls['firstName'].touched &&
userDetailsFormGroup.controls['firstName'].errors.required">
First Name is required
</mat-error>
</mat-form-field>
<mat-form-field>
<input matInput formControlName="lastName" type="text" placeholder="Last Name" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['lastName'].valid &&
userDetailsFormGroup.controls['lastName'].touched &&
userDetailsFormGroup.controls['lastName'].errors.required">
Last Name is required
</mat-error>
</mat-form-field>
<mat-form-field>
<mat-select formControlName="roles" placeholder="Roles" multiple required>
<mat-option *ngFor="let role of rolesList" [value]=role.key>{{ role.displayedValue }}</mat-option>
</mat-select>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['roles'].valid &&
userDetailsFormGroup.controls['roles'].touched &&
userDetailsFormGroup.controls['roles'].errors.required">
Role is required
</mat-error>
</mat-form-field>
<mat-form-field>
<input matInput formControlName="email" type="email" placeholder="Email" required>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid &&
userDetailsFormGroup.controls['email'].touched &&
userDetailsFormGroup.controls['email'].errors.required">
Email is required
</mat-error>
<mat-error align="start" *ngIf="!userDetailsFormGroup.controls['email'].valid &&
userDetailsFormGroup.controls['email'].touched &&
userDetailsFormGroup.controls['email'].errors.email">
Email is not valid
</mat-error>
</mat-form-field>
<button mat-raised-button matTooltip="Add Course to this user" (click)="openSelectItemDialog()">Add Course</button>
</form>
</mat-card-content>
<mat-card-actions align="start">
<button mat-button matTooltip="Save User" (click)="saveUser(event)">Save User</button>
<button mat-button matTooltip="Delete User" (click)="deleteUser(event)">Delete User</button>
</mat-card-actions>
<mat-card-footer>
Footer
</mat-card-footer>
</mat-card>
</div>
<!-- end detail -->
</div>
<!-- end master detail -->
</div>
<!-- end master detail container-->
```
<!-- This is mandatory for bug reports. This will help us to replicate the scenario. -->
##### **Compodoc installed globally or locally ?**
Locally
##### **Motivation for or Use Case**
<Unable to resolve templates
##### **Reproduce the error**
<!-- an unambiguous set of steps to reproduce the error. -->
##### **Related issues**
<!-- has a similar issue been reported before? -->
##### **Suggest a Fix**
no idea
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | priority | usersadmincomponent has a templateurl include it cannot read template for usersadmincomponent overview of the issue i am unable to run it for templates the application compiles and runs perfectly the template is in the same folder as component i get the following error userpreferencescomponent has a templateurl include it userprofilecomponent has a templateurl include it userregistrationcomponent has a templateurl include it usersignincomponent has a templateurl include it userverificationcomponent has a templateurl include it usersadmincomponent has a templateurl include it cannot read template for usersadmincomponent operating system node js npm compodoc version s angular cli node os darwin angular animations angular cdk angular common angular compiler angular core angular forms angular http angular material angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli angular language service typescript angular configuration a package json file in the root folder users admin component ts component selector app users admin templateurl users admin component html styleurls export class usersadmincomponent implements oninit users admin component html back to dashboard users admin works users subtitle user firstname user lastname add user delete users footer user details details mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required first name is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required last name is required role displayedvalue mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required role is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors required email is required mat error align start ngif userdetailsformgroup controls valid userdetailsformgroup controls touched userdetailsformgroup controls errors email email is not valid add course save user delete user footer compodoc installed globally or locally locally motivation for or use case unable to resolve templates reproduce the error related issues suggest a fix no idea love compodoc please consider supporting our collective 👉 | 1 |
112,676 | 4,535,786,650 | IssuesEvent | 2016-09-08 18:25:03 | newamericafoundation/newamerica-data-projects | https://api.github.com/repos/newamericafoundation/newamerica-data-projects | closed | Title panel heading css | medium-priority | Font seemed to have gotten smaller and padding was added
`
.title-panel__title { font-size: 3rem; } .title-panel__title { text-align: center; font-weight: 700; font-size: 2rem; line-height: 1.2; padding: 0px 0; }` | 1.0 | Title panel heading css - Font seemed to have gotten smaller and padding was added
`
.title-panel__title { font-size: 3rem; } .title-panel__title { text-align: center; font-weight: 700; font-size: 2rem; line-height: 1.2; padding: 0px 0; }` | priority | title panel heading css font seemed to have gotten smaller and padding was added title panel title font size title panel title text align center font weight font size line height padding | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.