Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
957
| labels
stringlengths 4
1.11k
| body
stringlengths 1
261k
| index
stringclasses 11
values | text_combine
stringlengths 95
261k
| label
stringclasses 2
values | text
stringlengths 96
250k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
122,791
| 16,330,079,915
|
IssuesEvent
|
2021-05-12 08:08:48
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
Naming of naming styles should reflect their casing
|
Area-IDE Feature Request Need Design Review
|
**Version Used**:
Microsoft Visual Studio Enterprise 2019 Int Preview
Version 16.1.0 Preview 3.0 [28823.117.d16.1]
VisualStudio.16.IntPreview/16.1.0-pre.3.0+28823.117.d16.1
Microsoft .NET Framework
Version 4.8.03752
**Steps to Reproduce**:
1. Ctrl Q
2. Type 'C# naming'
3. Enter
4. Open the drop down "Required Style" for any entry
**Expected Behavior**:
Finding the following entries:
* `PascalCase`
* `camelCase`
**Actual Behavior**:
These entries:
* `Pascal Case`
* `Camel Case`
[Was raised here](https://twitter.com/gotheap/status/1121118929097637888).
|
1.0
|
Naming of naming styles should reflect their casing - **Version Used**:
Microsoft Visual Studio Enterprise 2019 Int Preview
Version 16.1.0 Preview 3.0 [28823.117.d16.1]
VisualStudio.16.IntPreview/16.1.0-pre.3.0+28823.117.d16.1
Microsoft .NET Framework
Version 4.8.03752
**Steps to Reproduce**:
1. Ctrl Q
2. Type 'C# naming'
3. Enter
4. Open the drop down "Required Style" for any entry
**Expected Behavior**:
Finding the following entries:
* `PascalCase`
* `camelCase`
**Actual Behavior**:
These entries:
* `Pascal Case`
* `Camel Case`
[Was raised here](https://twitter.com/gotheap/status/1121118929097637888).
|
design
|
naming of naming styles should reflect their casing version used microsoft visual studio enterprise int preview version preview visualstudio intpreview pre microsoft net framework version steps to reproduce ctrl q type c naming enter open the drop down required style for any entry expected behavior finding the following entries pascalcase camelcase actual behavior these entries pascal case camel case
| 1
|
113,414
| 14,434,456,750
|
IssuesEvent
|
2020-12-07 07:06:51
|
h-yoshikawa0724/ooui-memo
|
https://api.github.com/repos/h-yoshikawa0724/ooui-memo
|
opened
|
[新規]設計書作成
|
basic design
|
## 目的
アプリケーションを作るにあたっての、設計書をおこす。
## 作業内容
OOUI や UI プロトタイプの情報をもとに、設計書にする
- ER 図
- 機能一覧
- 構成図
|
1.0
|
[新規]設計書作成 - ## 目的
アプリケーションを作るにあたっての、設計書をおこす。
## 作業内容
OOUI や UI プロトタイプの情報をもとに、設計書にする
- ER 図
- 機能一覧
- 構成図
|
design
|
設計書作成 目的 アプリケーションを作るにあたっての、設計書をおこす。 作業内容 ooui や ui プロトタイプの情報をもとに、設計書にする er 図 機能一覧 構成図
| 1
|
58,535
| 7,160,619,982
|
IssuesEvent
|
2018-01-28 03:06:24
|
CCBlueX/LiquidBounce1.8-Issues
|
https://api.github.com/repos/CCBlueX/LiquidBounce1.8-Issues
|
closed
|
Keybinds GUI
|
GUI (Design etc) Request
|
Ich wäre dafür, dass es in der GUI eine Art "Button" gibt, welcher einen direkt die Keybinds setzen lässt.
Oder eventuell mit der mittleren Maustaste auf die Modules drücken und dann erscheint eine virtuelle Tastatur, wo man nur noch die Tasten anklicken muss. (Wie in Null).
|
1.0
|
Keybinds GUI - Ich wäre dafür, dass es in der GUI eine Art "Button" gibt, welcher einen direkt die Keybinds setzen lässt.
Oder eventuell mit der mittleren Maustaste auf die Modules drücken und dann erscheint eine virtuelle Tastatur, wo man nur noch die Tasten anklicken muss. (Wie in Null).
|
design
|
keybinds gui ich wäre dafür dass es in der gui eine art button gibt welcher einen direkt die keybinds setzen lässt oder eventuell mit der mittleren maustaste auf die modules drücken und dann erscheint eine virtuelle tastatur wo man nur noch die tasten anklicken muss wie in null
| 1
|
20,031
| 13,633,717,237
|
IssuesEvent
|
2020-09-24 22:00:59
|
dotnet/aspnetcore
|
https://api.github.com/repos/dotnet/aspnetcore
|
closed
|
cannot build using restore.cmd
|
area-infrastructure
|
### If you believe you have an issue that affects the security of the platform please do NOT create an issue and instead email your issue details to secure@microsoft.com. Your report may be eligible for our [bug bounty](https://technet.microsoft.com/en-us/mt764065.aspx) but ONLY if it is reported through email.
### Describe the bug
tried to build asp.net core and followed all steps - got following error on restore.cmd
PS C:\dotnetcore\AspNetCore> .\restore.cmd
Detected JDK in C:\dotnetcore\AspNetCore\.tools\jdk\win-x64\ (via local repo convention)
C:\Users\jlee\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19369.2\tools\Tools.proj : error : Unable to locate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the installed version.
C:\Users\jlee\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19369.2\tools\Tools.proj : error MSB4236: The SDK 'Microsoft.NET.Sdk' specified could not be found.
Build failed.
### To Reproduce
Steps to reproduce the behavior:
1. Using this version of ASP.NET Core '...'
2. Run this code '....'
3. With these arguments '....'
4. See error
### Expected behavior
A clear and concise description of what you expected to happen.
### Screenshots
If applicable, add screenshots to help explain your problem.
### Additional context
Add any other context about the problem here.
Include the output of `dotnet --info`
|
1.0
|
cannot build using restore.cmd - ### If you believe you have an issue that affects the security of the platform please do NOT create an issue and instead email your issue details to secure@microsoft.com. Your report may be eligible for our [bug bounty](https://technet.microsoft.com/en-us/mt764065.aspx) but ONLY if it is reported through email.
### Describe the bug
tried to build asp.net core and followed all steps - got following error on restore.cmd
PS C:\dotnetcore\AspNetCore> .\restore.cmd
Detected JDK in C:\dotnetcore\AspNetCore\.tools\jdk\win-x64\ (via local repo convention)
C:\Users\jlee\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19369.2\tools\Tools.proj : error : Unable to locate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the installed version.
C:\Users\jlee\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19369.2\tools\Tools.proj : error MSB4236: The SDK 'Microsoft.NET.Sdk' specified could not be found.
Build failed.
### To Reproduce
Steps to reproduce the behavior:
1. Using this version of ASP.NET Core '...'
2. Run this code '....'
3. With these arguments '....'
4. See error
### Expected behavior
A clear and concise description of what you expected to happen.
### Screenshots
If applicable, add screenshots to help explain your problem.
### Additional context
Add any other context about the problem here.
Include the output of `dotnet --info`
|
non_design
|
cannot build using restore cmd if you believe you have an issue that affects the security of the platform please do not create an issue and instead email your issue details to secure microsoft com your report may be eligible for our but only if it is reported through email describe the bug tried to build asp net core and followed all steps got following error on restore cmd ps c dotnetcore aspnetcore restore cmd detected jdk in c dotnetcore aspnetcore tools jdk win via local repo convention c users jlee nuget packages microsoft dotnet arcade sdk beta tools tools proj error unable to locate the net core sdk check that it is installed and that the version specified in global json if any matches the installed version c users jlee nuget packages microsoft dotnet arcade sdk beta tools tools proj error the sdk microsoft net sdk specified could not be found build failed to reproduce steps to reproduce the behavior using this version of asp net core run this code with these arguments see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here include the output of dotnet info
| 0
|
119,614
| 15,586,378,565
|
IssuesEvent
|
2021-03-18 01:49:27
|
UOA-SE701-Group3-2021/3Lancers
|
https://api.github.com/repos/UOA-SE701-Group3-2021/3Lancers
|
opened
|
Hi-fi design of Habit Tracker User Input Menu
|
design
|
**Is your feature request related to a problem? Please describe.**
As a frontend developer, I want a detailed prototype of the Habit Tracker User Input Menu I'm implementing, so that I can finalise the styling and logic.
**Describe the solution you'd like**
A hi-fi prototype of the Habit Tracker User Input Menu
**Describe alternatives you've considered**
Start coding right away - would result in conflict further down the road.
Come up with own styling - would stand out from the UI.
**Additional context**
This is the menu that the user can interact with to add a new habit or edit an existing habit.
|
1.0
|
Hi-fi design of Habit Tracker User Input Menu - **Is your feature request related to a problem? Please describe.**
As a frontend developer, I want a detailed prototype of the Habit Tracker User Input Menu I'm implementing, so that I can finalise the styling and logic.
**Describe the solution you'd like**
A hi-fi prototype of the Habit Tracker User Input Menu
**Describe alternatives you've considered**
Start coding right away - would result in conflict further down the road.
Come up with own styling - would stand out from the UI.
**Additional context**
This is the menu that the user can interact with to add a new habit or edit an existing habit.
|
design
|
hi fi design of habit tracker user input menu is your feature request related to a problem please describe as a frontend developer i want a detailed prototype of the habit tracker user input menu i m implementing so that i can finalise the styling and logic describe the solution you d like a hi fi prototype of the habit tracker user input menu describe alternatives you ve considered start coding right away would result in conflict further down the road come up with own styling would stand out from the ui additional context this is the menu that the user can interact with to add a new habit or edit an existing habit
| 1
|
41,587
| 5,344,632,965
|
IssuesEvent
|
2017-02-17 15:02:01
|
JDTeamAcetabulum/qna
|
https://api.github.com/repos/JDTeamAcetabulum/qna
|
closed
|
As a user, I want the product to have a consistent brand so that I can easily identify it
|
design question
|
## Story/task details
I @strburst can get started on the logo/color scheme after we pick a name if no one feels strongly.
- [x] Choose a name for the product (`qna`)
- [ ] Create a logo (an `svg`?)
- [ ] Choose a decent color scheme
## Acceptance scenarios <!-- Under what conditions is this story applicable? -->
### Given <!-- Things that should have happened before the feature comes into play -->
Nothing needs to happen beforehand.
### When <!-- Specific events that trigger the feature -->
Visiting the site should trigger this feature, since color, name, and logo should all be clearly visible.
### Then <!-- What this feature will do -->
We will follow consistent design guidelines when implementing the product.
## Definition of done <!-- This feature will be ready to ship to the customer when... -->
This feature is ready when the product has all the essential branding elements chosen.
|
1.0
|
As a user, I want the product to have a consistent brand so that I can easily identify it - ## Story/task details
I @strburst can get started on the logo/color scheme after we pick a name if no one feels strongly.
- [x] Choose a name for the product (`qna`)
- [ ] Create a logo (an `svg`?)
- [ ] Choose a decent color scheme
## Acceptance scenarios <!-- Under what conditions is this story applicable? -->
### Given <!-- Things that should have happened before the feature comes into play -->
Nothing needs to happen beforehand.
### When <!-- Specific events that trigger the feature -->
Visiting the site should trigger this feature, since color, name, and logo should all be clearly visible.
### Then <!-- What this feature will do -->
We will follow consistent design guidelines when implementing the product.
## Definition of done <!-- This feature will be ready to ship to the customer when... -->
This feature is ready when the product has all the essential branding elements chosen.
|
design
|
as a user i want the product to have a consistent brand so that i can easily identify it story task details i strburst can get started on the logo color scheme after we pick a name if no one feels strongly choose a name for the product qna create a logo an svg choose a decent color scheme acceptance scenarios given nothing needs to happen beforehand when visiting the site should trigger this feature since color name and logo should all be clearly visible then we will follow consistent design guidelines when implementing the product definition of done this feature is ready when the product has all the essential branding elements chosen
| 1
|
443,560
| 30,923,656,412
|
IssuesEvent
|
2023-08-06 08:01:23
|
kubecub/go-project-layout
|
https://api.github.com/repos/kubecub/go-project-layout
|
closed
|
Bug reports for links in kubecub docs
|
kind/documentation triage/unresolved report lifecycle/stale
|
## Summary
| Status | Count |
|---------------|-------|
| 🔍 Total | 171 |
| ✅ Successful | 164 |
| ⏳ Timeouts | 1 |
| 🔀 Redirected | 0 |
| 👻 Excluded | 0 |
| ❓ Unknown | 0 |
| 🚫 Errors | 6 |
## Errors per input
### Errors in CONTRIBUTING.md
* [TIMEOUT] [https://twitter.com/xxw3293172751](https://twitter.com/xxw3293172751) | Timeout
### Errors in .github/CODE_OF_CONDUCT.md
* [404] [https://github.com/kubecub/community/blob/main/DEVELOPGUIDE.md](https://github.com/kubecub/community/blob/main/DEVELOPGUIDE.md) | Failed: Network error: Not Found
* [404] [https://github.com/kubecub/kubecub/tree/main/.github/ISSUE_TEMPLATE](https://github.com/kubecub/kubecub/tree/main/.github/ISSUE_TEMPLATE) | Failed: Network error: Not Found
* [ERR] [file:///home/runner/work/go-project-layout/go-project-layout/.github/google.com/search](file:///home/runner/work/go-project-layout/go-project-layout/.github/google.com/search) | Failed: Cannot find file
* [ERR] [file:///home/runner/work/go-project-layout/go-project-layout/.github/nsddd.top](file:///home/runner/work/go-project-layout/go-project-layout/.github/nsddd.top) | Failed: Cannot find file
### Errors in README.md
* [400] [https://github.com/issues?q=org%kubecub+is%3Aissue+label%3A%22good+first+issue%22+no%3Aassignee](https://github.com/issues?q=org%kubecub+is%3Aissue+label%3A%22good+first+issue%22+no%3Aassignee) | Failed: Network error: Bad Request
* [404] [https://github.com/kubecub/go-project-layout/generate](https://github.com/kubecub/go-project-layout/generate) | Failed: Network error: Not Found
[Full Github Actions output](https://github.com/kubecub/go-project-layout/actions/runs/5636734830?check_suite_focus=true)
|
1.0
|
Bug reports for links in kubecub docs - ## Summary
| Status | Count |
|---------------|-------|
| 🔍 Total | 171 |
| ✅ Successful | 164 |
| ⏳ Timeouts | 1 |
| 🔀 Redirected | 0 |
| 👻 Excluded | 0 |
| ❓ Unknown | 0 |
| 🚫 Errors | 6 |
## Errors per input
### Errors in CONTRIBUTING.md
* [TIMEOUT] [https://twitter.com/xxw3293172751](https://twitter.com/xxw3293172751) | Timeout
### Errors in .github/CODE_OF_CONDUCT.md
* [404] [https://github.com/kubecub/community/blob/main/DEVELOPGUIDE.md](https://github.com/kubecub/community/blob/main/DEVELOPGUIDE.md) | Failed: Network error: Not Found
* [404] [https://github.com/kubecub/kubecub/tree/main/.github/ISSUE_TEMPLATE](https://github.com/kubecub/kubecub/tree/main/.github/ISSUE_TEMPLATE) | Failed: Network error: Not Found
* [ERR] [file:///home/runner/work/go-project-layout/go-project-layout/.github/google.com/search](file:///home/runner/work/go-project-layout/go-project-layout/.github/google.com/search) | Failed: Cannot find file
* [ERR] [file:///home/runner/work/go-project-layout/go-project-layout/.github/nsddd.top](file:///home/runner/work/go-project-layout/go-project-layout/.github/nsddd.top) | Failed: Cannot find file
### Errors in README.md
* [400] [https://github.com/issues?q=org%kubecub+is%3Aissue+label%3A%22good+first+issue%22+no%3Aassignee](https://github.com/issues?q=org%kubecub+is%3Aissue+label%3A%22good+first+issue%22+no%3Aassignee) | Failed: Network error: Bad Request
* [404] [https://github.com/kubecub/go-project-layout/generate](https://github.com/kubecub/go-project-layout/generate) | Failed: Network error: Not Found
[Full Github Actions output](https://github.com/kubecub/go-project-layout/actions/runs/5636734830?check_suite_focus=true)
|
non_design
|
bug reports for links in kubecub docs summary status count 🔍 total ✅ successful ⏳ timeouts 🔀 redirected 👻 excluded ❓ unknown 🚫 errors errors per input errors in contributing md timeout errors in github code of conduct md failed network error not found failed network error not found file home runner work go project layout go project layout github google com search failed cannot find file file home runner work go project layout go project layout github nsddd top failed cannot find file errors in readme md failed network error bad request failed network error not found
| 0
|
18,583
| 13,055,589,413
|
IssuesEvent
|
2020-07-30 02:08:45
|
jstanden/cerb
|
https://api.github.com/repos/jstanden/cerb
|
closed
|
Nested square brackets in outgoing HTML email cause problems in plain-text copy
|
bug usability
|
Using nested square brackets in an outgoing HTML email causes that line to be missing from the plaintext copy.
Source is:

received HTML is:

Received raw plain-text is:

|
True
|
Nested square brackets in outgoing HTML email cause problems in plain-text copy - Using nested square brackets in an outgoing HTML email causes that line to be missing from the plaintext copy.
Source is:

received HTML is:

Received raw plain-text is:

|
non_design
|
nested square brackets in outgoing html email cause problems in plain text copy using nested square brackets in an outgoing html email causes that line to be missing from the plaintext copy source is received html is received raw plain text is
| 0
|
23,668
| 6,469,771,645
|
IssuesEvent
|
2017-08-17 07:11:33
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
[RyuJIT/armel] HFA tests fail
|
arch-arm32 area-CodeGen bug
|
With #13284, most regressions caused by #13023 were fixed. But still there is HFA test regression.
All of these run successfully, but the fails with Unexpected Results. (wrong result value)
- JIT/jit64/hfa/main/testB/hfa_sf2B_d/hfa_sf2B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sf0B_r/hfa_sf0B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sf2B_r/hfa_sf2B_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd0G_r/hfa_sd0G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd1G_r/hfa_sd1G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd2G_r/hfa_sd2G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sf2G_r/hfa_sf2G_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd0A_r/hfa_sd0A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd1A_r/hfa_sd1A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd2A_d/hfa_sd2A_d.exe
- JIT/jit64/hfa/main/testA/hfa_sd2A_r/hfa_sd2A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf0A_r/hfa_sf0A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf1A_r/hfa_sf1A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf2A_d/hfa_sf2A_d.exe
- JIT/jit64/hfa/main/testA/hfa_sf2A_r/hfa_sf2A_r.exe
- JIT/jit64/hfa/main/testB/hfa_sd0B_d/hfa_sd0B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sd0B_r/hfa_sd0B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sd2B_d/hfa_sd2B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sd2B_r/hfa_sd2B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sf0B_d/hfa_sf0B_d.exe
|
1.0
|
[RyuJIT/armel] HFA tests fail - With #13284, most regressions caused by #13023 were fixed. But still there is HFA test regression.
All of these run successfully, but the fails with Unexpected Results. (wrong result value)
- JIT/jit64/hfa/main/testB/hfa_sf2B_d/hfa_sf2B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sf0B_r/hfa_sf0B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sf2B_r/hfa_sf2B_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd0G_r/hfa_sd0G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd1G_r/hfa_sd1G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sd2G_r/hfa_sd2G_r.exe
- JIT/jit64/hfa/main/testG/hfa_sf2G_r/hfa_sf2G_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd0A_r/hfa_sd0A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd1A_r/hfa_sd1A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sd2A_d/hfa_sd2A_d.exe
- JIT/jit64/hfa/main/testA/hfa_sd2A_r/hfa_sd2A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf0A_r/hfa_sf0A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf1A_r/hfa_sf1A_r.exe
- JIT/jit64/hfa/main/testA/hfa_sf2A_d/hfa_sf2A_d.exe
- JIT/jit64/hfa/main/testA/hfa_sf2A_r/hfa_sf2A_r.exe
- JIT/jit64/hfa/main/testB/hfa_sd0B_d/hfa_sd0B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sd0B_r/hfa_sd0B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sd2B_d/hfa_sd2B_d.exe
- JIT/jit64/hfa/main/testB/hfa_sd2B_r/hfa_sd2B_r.exe
- JIT/jit64/hfa/main/testB/hfa_sf0B_d/hfa_sf0B_d.exe
|
non_design
|
hfa tests fail with most regressions caused by were fixed but still there is hfa test regression all of these run successfully but the fails with unexpected results wrong result value jit hfa main testb hfa d hfa d exe jit hfa main testb hfa r hfa r exe jit hfa main testb hfa r hfa r exe jit hfa main testg hfa r hfa r exe jit hfa main testg hfa r hfa r exe jit hfa main testg hfa r hfa r exe jit hfa main testg hfa r hfa r exe jit hfa main testa hfa r hfa r exe jit hfa main testa hfa r hfa r exe jit hfa main testa hfa d hfa d exe jit hfa main testa hfa r hfa r exe jit hfa main testa hfa r hfa r exe jit hfa main testa hfa r hfa r exe jit hfa main testa hfa d hfa d exe jit hfa main testa hfa r hfa r exe jit hfa main testb hfa d hfa d exe jit hfa main testb hfa r hfa r exe jit hfa main testb hfa d hfa d exe jit hfa main testb hfa r hfa r exe jit hfa main testb hfa d hfa d exe
| 0
|
77,615
| 9,602,792,189
|
IssuesEvent
|
2019-05-10 15:22:27
|
AugurProject/augur
|
https://api.github.com/repos/AugurProject/augur
|
closed
|
Post cutoff: Global banner across the app with link to post
|
Design Review
|
- Display a global banner across the app with link to post
https://www.figma.com/file/csVtbWhGCAYe3nVe2LI1UEPM/v1-updates-to-the-UI?node-id=301%3A2551
(scroll right for more)
- have this show up only if the date is after the cutoff date so it is triggered by the time
- change forking banner to new color
- move forking banner to top of the page and the global banner shows up under it
|
1.0
|
Post cutoff: Global banner across the app with link to post - - Display a global banner across the app with link to post
https://www.figma.com/file/csVtbWhGCAYe3nVe2LI1UEPM/v1-updates-to-the-UI?node-id=301%3A2551
(scroll right for more)
- have this show up only if the date is after the cutoff date so it is triggered by the time
- change forking banner to new color
- move forking banner to top of the page and the global banner shows up under it
|
design
|
post cutoff global banner across the app with link to post display a global banner across the app with link to post scroll right for more have this show up only if the date is after the cutoff date so it is triggered by the time change forking banner to new color move forking banner to top of the page and the global banner shows up under it
| 1
|
211,542
| 23,833,143,205
|
IssuesEvent
|
2022-09-06 01:07:26
|
RG4421/java-slack-sdk
|
https://api.github.com/repos/RG4421/java-slack-sdk
|
opened
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.25.jar, snakeyaml-1.24.jar
|
security vulnerability
|
## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.25.jar</b>, <b>snakeyaml-1.24.jar</b></p></summary>
<p>
<details><summary><b>snakeyaml-1.25.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /bolt-spring-boot-examples/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.7.RELEASE.jar (Root Library)
- spring-boot-starter-2.2.7.RELEASE.jar
- :x: **snakeyaml-1.25.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.24.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /bolt-helidon/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.24/snakeyaml-1.24.jar</p>
<p>
Dependency Hierarchy:
- helidon-config-yaml-1.4.4.jar (Root Library)
- :x: **snakeyaml-1.24.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/RG4421/java-slack-sdk/commit/247198cd1cfbaa95a065923a5268b83aa877f12a">247198cd1cfbaa95a065923a5268b83aa877f12a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
True
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.25.jar, snakeyaml-1.24.jar - ## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.25.jar</b>, <b>snakeyaml-1.24.jar</b></p></summary>
<p>
<details><summary><b>snakeyaml-1.25.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /bolt-spring-boot-examples/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.25/snakeyaml-1.25.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.7.RELEASE.jar (Root Library)
- spring-boot-starter-2.2.7.RELEASE.jar
- :x: **snakeyaml-1.25.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.24.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /bolt-helidon/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.24/snakeyaml-1.24.jar</p>
<p>
Dependency Hierarchy:
- helidon-config-yaml-1.4.4.jar (Root Library)
- :x: **snakeyaml-1.24.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/RG4421/java-slack-sdk/commit/247198cd1cfbaa95a065923a5268b83aa877f12a">247198cd1cfbaa95a065923a5268b83aa877f12a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
non_design
|
cve medium detected in snakeyaml jar snakeyaml jar cve medium severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file bolt spring boot examples pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter web release jar root library spring boot starter release jar x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file bolt helidon pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy helidon config yaml jar root library x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
| 0
|
174,189
| 27,591,237,883
|
IssuesEvent
|
2023-03-09 00:39:16
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[REMOTO] [TAMBÉM PCD] [PHP] [SÊNIOR] Pessoa Desenvolvedora PHP - Sênior | Conta PF na [PICPAY]
|
HOME OFFICE PHP MYSQL MONGODB LARAVEL SENIOR GIT DOCKER REMOTO KAFKA RABBITMQ LUMEN DESIGN PATTERNS RESTFUL METODOLOGIAS ÁGEIS HELP WANTED VAGA PARA PCD TAMBÉM OWASP TOP 10 Stale
|
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- O PicPay quer mudar a maneira de como lidamos com o dinheiro, tanto pessoal quanto profissionalmente. Somos uma plataforma de pagamentos criada para quebrar barreiras e eliminar burocracias. Sendo assim, nós existimos para melhorar a vida das pessoas e só conseguimos fazer isso porque temos um time fantástico aqui dentro!
- Contamos com uma equipe de mais de 1.800 PicPay Lovers - engenheiros, desenvolvedores, profissionais de marketing, designers, criadores, nerds loucos e mais um punhado de profissionais sonhadores que dedicam suas horas desafiando as regras de tempo x propósito.
- Buscamos pessoas que sejam apaixonadas por construir aplicações distribuídas e de alta performance, escaláveis para milhões de novos usuários em todo o mundo e que suportem a melhor experiência de pagamentos. Somos um dos melhores lugares para se trabalhar no Brasil (pelo menos as nossas pessoas acham isso)!
- Manda bem nos requisitos abaixo? Então essa vaga é pra você! (:
- Junte-se ao lado verde da força!
**Responsabilidades e atribuições:**
- Criar funcionalidades de grande escala pensando na performance e segurança;
- Criar e manter documentações dos serviços;
- Avaliar e discutir requisitos;
- Implementar testes para melhor manutenção.
## Local
- Home Office
## Benefícios
- Assistências médica (você e +1 dependente ficam por nossa conta!);
- Assistência odontológica;
- Seguro de vida;
- Vale Transporte e/ou Auxílio combustível;
- Vale Refeição e/ou Vale Alimentação;
- Gympass;
- PicPay Acolhe - Programa que cuida da gente e dos nossos familiares, oferecendo apoio jurídico, social, psicológico e financeiro;
- PPR - Participação nos Resultados do PicPay;
- Horário flexível e possibilidade de home office.
## Requisitos
**Obrigatórios:**
- PHP 7.*;
- Frameworks Laravel ou Lumen, com ênfase em aplicações RESTful;
- Armazenamento de dados em MySQL (ou outros bancos relacionais) ou MongoDB (ou outros bancos não relacionais);
- Ambiente de desenvolvimento com Docker;
- Atenção para qualidade de código e design patterns;
- Experiência sólida com desenvolvimento orientado a testes;
- Atenção para trabalhar com criação de documentação;
- Versionamento de códigos utilizando git;
- Conhecimento sobre serviços síncronos e assíncronos;
- Conhecimento dos riscos comuns de segurança (OWASP).
**Diferenciais:**
- Metodologias ágeis;
- Design e desenvolvimento de micro serviços horizontalmente escaláveis;
- Jobs de alto desempenho e comunicação entre serviços utilizando soluções de fila como Beanstalk, RabbitMQ, Kafka, etc.
## Contratação
- Efetivo
## Nossa empresa
- Fazer todos os seus pagamentos de forma simples e rápida. Na nossa plataforma PicPay dá para enviar e receber dinheiro, fazer um Pix, pagar boletos, estabelecimentos, comprar crédito para o celular, para games online e muito mais. Tudo isso direto do celular.
- Nos últimos anos, o PicPay conquistou milhões de usuários e busca diariamente tornar a vida das pessoas mais fácil.
- Contamos com um time de pessoas fantásticas e que amam o que fazem, somos #PicPayLovers.
## Como se candidatar
- [Clique aqui para se candidatar](https://picpay.gupy.io/jobs/605055?jobBoardSource=gupy_public_page)
|
1.0
|
[REMOTO] [TAMBÉM PCD] [PHP] [SÊNIOR] Pessoa Desenvolvedora PHP - Sênior | Conta PF na [PICPAY] - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- O PicPay quer mudar a maneira de como lidamos com o dinheiro, tanto pessoal quanto profissionalmente. Somos uma plataforma de pagamentos criada para quebrar barreiras e eliminar burocracias. Sendo assim, nós existimos para melhorar a vida das pessoas e só conseguimos fazer isso porque temos um time fantástico aqui dentro!
- Contamos com uma equipe de mais de 1.800 PicPay Lovers - engenheiros, desenvolvedores, profissionais de marketing, designers, criadores, nerds loucos e mais um punhado de profissionais sonhadores que dedicam suas horas desafiando as regras de tempo x propósito.
- Buscamos pessoas que sejam apaixonadas por construir aplicações distribuídas e de alta performance, escaláveis para milhões de novos usuários em todo o mundo e que suportem a melhor experiência de pagamentos. Somos um dos melhores lugares para se trabalhar no Brasil (pelo menos as nossas pessoas acham isso)!
- Manda bem nos requisitos abaixo? Então essa vaga é pra você! (:
- Junte-se ao lado verde da força!
**Responsabilidades e atribuições:**
- Criar funcionalidades de grande escala pensando na performance e segurança;
- Criar e manter documentações dos serviços;
- Avaliar e discutir requisitos;
- Implementar testes para melhor manutenção.
## Local
- Home Office
## Benefícios
- Assistências médica (você e +1 dependente ficam por nossa conta!);
- Assistência odontológica;
- Seguro de vida;
- Vale Transporte e/ou Auxílio combustível;
- Vale Refeição e/ou Vale Alimentação;
- Gympass;
- PicPay Acolhe - Programa que cuida da gente e dos nossos familiares, oferecendo apoio jurídico, social, psicológico e financeiro;
- PPR - Participação nos Resultados do PicPay;
- Horário flexível e possibilidade de home office.
## Requisitos
**Obrigatórios:**
- PHP 7.*;
- Frameworks Laravel ou Lumen, com ênfase em aplicações RESTful;
- Armazenamento de dados em MySQL (ou outros bancos relacionais) ou MongoDB (ou outros bancos não relacionais);
- Ambiente de desenvolvimento com Docker;
- Atenção para qualidade de código e design patterns;
- Experiência sólida com desenvolvimento orientado a testes;
- Atenção para trabalhar com criação de documentação;
- Versionamento de códigos utilizando git;
- Conhecimento sobre serviços síncronos e assíncronos;
- Conhecimento dos riscos comuns de segurança (OWASP).
**Diferenciais:**
- Metodologias ágeis;
- Design e desenvolvimento de micro serviços horizontalmente escaláveis;
- Jobs de alto desempenho e comunicação entre serviços utilizando soluções de fila como Beanstalk, RabbitMQ, Kafka, etc.
## Contratação
- Efetivo
## Nossa empresa
- Fazer todos os seus pagamentos de forma simples e rápida. Na nossa plataforma PicPay dá para enviar e receber dinheiro, fazer um Pix, pagar boletos, estabelecimentos, comprar crédito para o celular, para games online e muito mais. Tudo isso direto do celular.
- Nos últimos anos, o PicPay conquistou milhões de usuários e busca diariamente tornar a vida das pessoas mais fácil.
- Contamos com um time de pessoas fantásticas e que amam o que fazem, somos #PicPayLovers.
## Como se candidatar
- [Clique aqui para se candidatar](https://picpay.gupy.io/jobs/605055?jobBoardSource=gupy_public_page)
|
design
|
pessoa desenvolvedora php sênior conta pf na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga o picpay quer mudar a maneira de como lidamos com o dinheiro tanto pessoal quanto profissionalmente somos uma plataforma de pagamentos criada para quebrar barreiras e eliminar burocracias sendo assim nós existimos para melhorar a vida das pessoas e só conseguimos fazer isso porque temos um time fantástico aqui dentro contamos com uma equipe de mais de picpay lovers engenheiros desenvolvedores profissionais de marketing designers criadores nerds loucos e mais um punhado de profissionais sonhadores que dedicam suas horas desafiando as regras de tempo x propósito buscamos pessoas que sejam apaixonadas por construir aplicações distribuídas e de alta performance escaláveis para milhões de novos usuários em todo o mundo e que suportem a melhor experiência de pagamentos somos um dos melhores lugares para se trabalhar no brasil pelo menos as nossas pessoas acham isso manda bem nos requisitos abaixo então essa vaga é pra você junte se ao lado verde da força responsabilidades e atribuições criar funcionalidades de grande escala pensando na performance e segurança criar e manter documentações dos serviços avaliar e discutir requisitos implementar testes para melhor manutenção local home office benefícios assistências médica você e dependente ficam por nossa conta assistência odontológica seguro de vida vale transporte e ou auxílio combustível vale refeição e ou vale alimentação gympass picpay acolhe programa que cuida da gente e dos nossos familiares oferecendo apoio jurídico social psicológico e financeiro ppr participação nos resultados do picpay horário flexível e possibilidade de home office requisitos obrigatórios php frameworks laravel ou lumen com ênfase em aplicações restful armazenamento de dados em mysql ou outros bancos relacionais ou mongodb ou outros bancos não relacionais ambiente de desenvolvimento com docker atenção para qualidade de código e design patterns experiência sólida com desenvolvimento orientado a testes atenção para trabalhar com criação de documentação versionamento de códigos utilizando git conhecimento sobre serviços síncronos e assíncronos conhecimento dos riscos comuns de segurança owasp diferenciais metodologias ágeis design e desenvolvimento de micro serviços horizontalmente escaláveis jobs de alto desempenho e comunicação entre serviços utilizando soluções de fila como beanstalk rabbitmq kafka etc contratação efetivo nossa empresa fazer todos os seus pagamentos de forma simples e rápida na nossa plataforma picpay dá para enviar e receber dinheiro fazer um pix pagar boletos estabelecimentos comprar crédito para o celular para games online e muito mais tudo isso direto do celular nos últimos anos o picpay conquistou milhões de usuários e busca diariamente tornar a vida das pessoas mais fácil contamos com um time de pessoas fantásticas e que amam o que fazem somos picpaylovers como se candidatar
| 1
|
138,058
| 20,322,203,854
|
IssuesEvent
|
2022-02-18 00:13:06
|
microsoft/pyright
|
https://api.github.com/repos/microsoft/pyright
|
closed
|
Pyright generates error "unpacking not allowed" in context of __get_item__calls for numpy, tensorflow tensors
|
as designed
|
Note: if you are reporting a wrong signature of a function or a class in the standard library, then the typeshed tracker is better suited for this report: https://github.com/python/typeshed/issues.
**Describe the bug**
"unpack operation is not allowed in this context" is generated. However, in python 3.8 this executes just fine.
**To Reproduce**
>>>
import numpy as np
a = np.array([[1,2]])
b = a[ (*[np.newaxis]*3), ...] # <<-- generates pyright error
b
array([[[[[1, 2]]]]])
**Expected behavior**
no pyright error
**Screenshots or Code**
**Additional context**
python 3.8, current versions of numpy and/or tensorflow, VSCode pylance plugin
|
1.0
|
Pyright generates error "unpacking not allowed" in context of __get_item__calls for numpy, tensorflow tensors - Note: if you are reporting a wrong signature of a function or a class in the standard library, then the typeshed tracker is better suited for this report: https://github.com/python/typeshed/issues.
**Describe the bug**
"unpack operation is not allowed in this context" is generated. However, in python 3.8 this executes just fine.
**To Reproduce**
>>>
import numpy as np
a = np.array([[1,2]])
b = a[ (*[np.newaxis]*3), ...] # <<-- generates pyright error
b
array([[[[[1, 2]]]]])
**Expected behavior**
no pyright error
**Screenshots or Code**
**Additional context**
python 3.8, current versions of numpy and/or tensorflow, VSCode pylance plugin
|
design
|
pyright generates error unpacking not allowed in context of get item calls for numpy tensorflow tensors note if you are reporting a wrong signature of a function or a class in the standard library then the typeshed tracker is better suited for this report describe the bug unpack operation is not allowed in this context is generated however in python this executes just fine to reproduce import numpy as np a np array b a generates pyright error b array expected behavior no pyright error screenshots or code additional context python current versions of numpy and or tensorflow vscode pylance plugin
| 1
|
51,917
| 3,015,572,706
|
IssuesEvent
|
2015-07-29 20:19:56
|
WikiEducationFoundation/WikiEduDashboard
|
https://api.github.com/repos/WikiEducationFoundation/WikiEduDashboard
|
closed
|
Course announcements are being made multiple times for the same course
|
bug top priority
|
Looks like this is still an issue: https://en.wikipedia.org/wiki/User:Cassell04
|
1.0
|
Course announcements are being made multiple times for the same course - Looks like this is still an issue: https://en.wikipedia.org/wiki/User:Cassell04
|
non_design
|
course announcements are being made multiple times for the same course looks like this is still an issue
| 0
|
77,004
| 21,646,226,166
|
IssuesEvent
|
2022-05-06 02:28:18
|
PyAV-Org/PyAV
|
https://api.github.com/repos/PyAV-Org/PyAV
|
opened
|
Install fails on Ubuntu 18
|
build
|
**IMPORTANT:** Be sure to replace all template sections {{ like this }} or your issue may be discarded.
## Overview
The bug is the following error when trying to run `pip install av` on ubuntu 18:
```
Defaulting to user installation because normal site-packages is not writeable
Collecting av
Using cached av-9.2.0.tar.gz (2.4 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python /usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py get_requires_for_build_wheel /tmp/tmpY4CaeY
cwd: /tmp/pip-install-uMRD3m/av
Complete output (18 lines):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 280, in <module>
main()
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 263, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 114, in get_requires_for_build_wheel
return hook(config_settings)
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 146, in get_requires_for_build_wheel
return self._get_build_requires(config_settings, requirements=['wheel'])
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 127, in _get_build_requires
self.run_setup()
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 243, in run_setup
self).run_setup(setup_script=setup_script)
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 142, in run_setup
exec(compile(code, __file__, 'exec'), locals())
File "setup.py", line 3, in <module>
import pathlib
ImportError: No module named pathlib
----------------------------------------
ERROR: Command errored out with exit status 1: /usr/bin/python /usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py get_requires_for_build_wheel /tmp/tmpY4CaeY Check the logs for full command output.
```
## Expected behavior
I expected the package to install with no errors.
## Actual behavior
The package failed to install due to errors.
Build report: I'm not actually trying to build from source so this is meaningless:
```
erinline@erinline-MacBookPro:~/test_tello$ python setup.py build
python: can't open file 'setup.py': [Errno 2] No such file or directory
```
## Investigation
I ran `pip install pathlib` as well as `pip3 install pathlib`, `python -m pip install pathlib`, and all similar with `av` as well. Pathlib is definitely already installed for both Python2 and Python3, but I get the same ImportError about pathlib for all python versions and all install commands.
## Reproduction
Run `pip install av`.
## Versions
- OS: Ubuntu 18.04
- PyAV runtime: N/A
```
{{ Complete output of `python -m av --version` if you can run it. }}
```
- PyAV build:
```
{{ Complete output of `python setup.py config --verbose`. }}
```
- FFmpeg:
```
ffmpeg version 3.4.8-0ubuntu0.2 Copyright (c) 2000-2020 the FFmpeg developers
built with gcc 7 (Ubuntu 7.5.0-3ubuntu1~18.04)
configuration: --prefix=/usr --extra-version=0ubuntu0.2 --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --enable-gpl --disable-stripping --enable-avresample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librubberband --enable-librsvg --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libopencv --enable-libx264 --enable-shared
libavutil 55. 78.100 / 55. 78.100
libavcodec 57.107.100 / 57.107.100
libavformat 57. 83.100 / 57. 83.100
libavdevice 57. 10.100 / 57. 10.100
libavfilter 6.107.100 / 6.107.100
libavresample 3. 7. 0 / 3. 7. 0
libswscale 4. 8.100 / 4. 8.100
libswresample 2. 9.100 / 2. 9.100
libpostproc 54. 7.100 / 54. 7.100
```
## Research
I have done the following:
- [ ] Checked the [PyAV documentation](https://pyav.org/docs)
- [ ] Searched on [Google](https://www.google.com/search?q=pyav+how+do+I+foo)
- [ ] Searched on [Stack Overflow](https://stackoverflow.com/search?q=pyav)
- [ ] Looked through [old GitHub issues](https://github.com/PyAV-Org/PyAV/issues?&q=is%3Aissue)
- [ ] Asked on [PyAV Gitter](https://gitter.im/PyAV-Org)
- [ ] ... and waited 72 hours for a response.
## Additional context
No other context.
|
1.0
|
Install fails on Ubuntu 18 - **IMPORTANT:** Be sure to replace all template sections {{ like this }} or your issue may be discarded.
## Overview
The bug is the following error when trying to run `pip install av` on ubuntu 18:
```
Defaulting to user installation because normal site-packages is not writeable
Collecting av
Using cached av-9.2.0.tar.gz (2.4 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python /usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py get_requires_for_build_wheel /tmp/tmpY4CaeY
cwd: /tmp/pip-install-uMRD3m/av
Complete output (18 lines):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 280, in <module>
main()
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 263, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py", line 114, in get_requires_for_build_wheel
return hook(config_settings)
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 146, in get_requires_for_build_wheel
return self._get_build_requires(config_settings, requirements=['wheel'])
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 127, in _get_build_requires
self.run_setup()
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 243, in run_setup
self).run_setup(setup_script=setup_script)
File "/tmp/pip-build-env-Fa6ICz/overlay/lib/python2.7/site-packages/setuptools/build_meta.py", line 142, in run_setup
exec(compile(code, __file__, 'exec'), locals())
File "setup.py", line 3, in <module>
import pathlib
ImportError: No module named pathlib
----------------------------------------
ERROR: Command errored out with exit status 1: /usr/bin/python /usr/local/lib/python2.7/dist-packages/pip/_vendor/pep517/_in_process.py get_requires_for_build_wheel /tmp/tmpY4CaeY Check the logs for full command output.
```
## Expected behavior
I expected the package to install with no errors.
## Actual behavior
The package failed to install due to errors.
Build report: I'm not actually trying to build from source so this is meaningless:
```
erinline@erinline-MacBookPro:~/test_tello$ python setup.py build
python: can't open file 'setup.py': [Errno 2] No such file or directory
```
## Investigation
I ran `pip install pathlib` as well as `pip3 install pathlib`, `python -m pip install pathlib`, and all similar with `av` as well. Pathlib is definitely already installed for both Python2 and Python3, but I get the same ImportError about pathlib for all python versions and all install commands.
## Reproduction
Run `pip install av`.
## Versions
- OS: Ubuntu 18.04
- PyAV runtime: N/A
```
{{ Complete output of `python -m av --version` if you can run it. }}
```
- PyAV build:
```
{{ Complete output of `python setup.py config --verbose`. }}
```
- FFmpeg:
```
ffmpeg version 3.4.8-0ubuntu0.2 Copyright (c) 2000-2020 the FFmpeg developers
built with gcc 7 (Ubuntu 7.5.0-3ubuntu1~18.04)
configuration: --prefix=/usr --extra-version=0ubuntu0.2 --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --enable-gpl --disable-stripping --enable-avresample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librubberband --enable-librsvg --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libopencv --enable-libx264 --enable-shared
libavutil 55. 78.100 / 55. 78.100
libavcodec 57.107.100 / 57.107.100
libavformat 57. 83.100 / 57. 83.100
libavdevice 57. 10.100 / 57. 10.100
libavfilter 6.107.100 / 6.107.100
libavresample 3. 7. 0 / 3. 7. 0
libswscale 4. 8.100 / 4. 8.100
libswresample 2. 9.100 / 2. 9.100
libpostproc 54. 7.100 / 54. 7.100
```
## Research
I have done the following:
- [ ] Checked the [PyAV documentation](https://pyav.org/docs)
- [ ] Searched on [Google](https://www.google.com/search?q=pyav+how+do+I+foo)
- [ ] Searched on [Stack Overflow](https://stackoverflow.com/search?q=pyav)
- [ ] Looked through [old GitHub issues](https://github.com/PyAV-Org/PyAV/issues?&q=is%3Aissue)
- [ ] Asked on [PyAV Gitter](https://gitter.im/PyAV-Org)
- [ ] ... and waited 72 hours for a response.
## Additional context
No other context.
|
non_design
|
install fails on ubuntu important be sure to replace all template sections like this or your issue may be discarded overview the bug is the following error when trying to run pip install av on ubuntu defaulting to user installation because normal site packages is not writeable collecting av using cached av tar gz mb installing build dependencies done getting requirements to build wheel error error command errored out with exit status command usr bin python usr local lib dist packages pip vendor in process py get requires for build wheel tmp cwd tmp pip install av complete output lines traceback most recent call last file usr local lib dist packages pip vendor in process py line in main file usr local lib dist packages pip vendor in process py line in main json out hook hook input file usr local lib dist packages pip vendor in process py line in get requires for build wheel return hook config settings file tmp pip build env overlay lib site packages setuptools build meta py line in get requires for build wheel return self get build requires config settings requirements file tmp pip build env overlay lib site packages setuptools build meta py line in get build requires self run setup file tmp pip build env overlay lib site packages setuptools build meta py line in run setup self run setup setup script setup script file tmp pip build env overlay lib site packages setuptools build meta py line in run setup exec compile code file exec locals file setup py line in import pathlib importerror no module named pathlib error command errored out with exit status usr bin python usr local lib dist packages pip vendor in process py get requires for build wheel tmp check the logs for full command output expected behavior i expected the package to install with no errors actual behavior the package failed to install due to errors build report i m not actually trying to build from source so this is meaningless erinline erinline macbookpro test tello python setup py build python can t open file setup py no such file or directory investigation i ran pip install pathlib as well as install pathlib python m pip install pathlib and all similar with av as well pathlib is definitely already installed for both and but i get the same importerror about pathlib for all python versions and all install commands reproduction run pip install av versions os ubuntu pyav runtime n a complete output of python m av version if you can run it pyav build complete output of python setup py config verbose ffmpeg ffmpeg version copyright c the ffmpeg developers built with gcc ubuntu configuration prefix usr extra version toolchain hardened libdir usr lib linux gnu incdir usr include linux gnu enable gpl disable stripping enable avresample enable avisynth enable gnutls enable ladspa enable libass enable libbluray enable enable libcaca enable libcdio enable libflite enable libfontconfig enable libfreetype enable libfribidi enable libgme enable libgsm enable enable libmysofa enable libopenjpeg enable libopenmpt enable libopus enable libpulse enable librubberband enable librsvg enable libshine enable libsnappy enable libsoxr enable libspeex enable libssh enable libtheora enable libtwolame enable libvorbis enable libvpx enable libwavpack enable libwebp enable enable enable libxvid enable libzmq enable libzvbi enable omx enable openal enable opengl enable enable enable libdrm enable enable chromaprint enable enable libopencv enable enable shared libavutil libavcodec libavformat libavdevice libavfilter libavresample libswscale libswresample libpostproc research i have done the following checked the searched on searched on looked through asked on and waited hours for a response additional context no other context
| 0
|
50,688
| 10,546,987,413
|
IssuesEvent
|
2019-10-02 23:12:36
|
MicrosoftDocs/visualstudio-docs
|
https://api.github.com/repos/MicrosoftDocs/visualstudio-docs
|
closed
|
unclear documentation
|
Pri2 area - C++ doc-bug visual-studio-windows/prod vs-ide-code-analysis/tech
|
Please define what slicing is.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: be6a0e19-d61f-fefb-4a83-afa87219e168
* Version Independent ID: 206fffd7-3651-5783-49e0-c586a9b8b992
* Content: [C26437 - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/code-quality/c26437?f1url=https%3A%2F%2Fmsdn.microsoft.com%2Fquery%2Fdev16.query%3FappId%3DDev16IDEF1%26l%3DEN-US%26k%3Dk(C26437)%26rd%3Dtrue&view=vs-2019#feedback)
* Content Source: [docs/code-quality/C26437.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/code-quality/C26437.md)
* Product: **visual-studio-windows**
* Technology: **vs-ide-code-analysis**
* GitHub Login: @mikeblome
* Microsoft Alias: **mblome**
|
1.0
|
unclear documentation - Please define what slicing is.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: be6a0e19-d61f-fefb-4a83-afa87219e168
* Version Independent ID: 206fffd7-3651-5783-49e0-c586a9b8b992
* Content: [C26437 - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/code-quality/c26437?f1url=https%3A%2F%2Fmsdn.microsoft.com%2Fquery%2Fdev16.query%3FappId%3DDev16IDEF1%26l%3DEN-US%26k%3Dk(C26437)%26rd%3Dtrue&view=vs-2019#feedback)
* Content Source: [docs/code-quality/C26437.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/code-quality/C26437.md)
* Product: **visual-studio-windows**
* Technology: **vs-ide-code-analysis**
* GitHub Login: @mikeblome
* Microsoft Alias: **mblome**
|
non_design
|
unclear documentation please define what slicing is document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id fefb version independent id content content source product visual studio windows technology vs ide code analysis github login mikeblome microsoft alias mblome
| 0
|
568,055
| 16,945,845,927
|
IssuesEvent
|
2021-06-28 06:38:22
|
PlaceOS/drivers
|
https://api.github.com/repos/PlaceOS/drivers
|
closed
|
Gallagher Crystal Driver
|
bhq priority: high type: driver type: migrate driver
|
Convert/Rewrite existing Ruby Gallagher Driver to Crystal for Suncorp
|
1.0
|
Gallagher Crystal Driver - Convert/Rewrite existing Ruby Gallagher Driver to Crystal for Suncorp
|
non_design
|
gallagher crystal driver convert rewrite existing ruby gallagher driver to crystal for suncorp
| 0
|
128,237
| 17,466,121,817
|
IssuesEvent
|
2021-08-06 17:05:58
|
cagov/ui-claim-tracker
|
https://api.github.com/repos/cagov/ui-claim-tracker
|
closed
|
Run rapid usability testing
|
Design
|
### Description
Facilitate content testing for:
- Test copy for scenario 1 (depends on #154)
- Test copy for baseline state for alpha release (depends on #158)
### Acceptance Criteria
- [x] Lead interviews
- [ ] #273
- [ ] #356
<!--
_Note_ When you create this issue, remember to add:
- an assignee
- the project, so that it will show up in our kanban view
- a label for story points estimate (or comment at the assignee to request that they add an estimate)
- a label for priority
-->
|
1.0
|
Run rapid usability testing - ### Description
Facilitate content testing for:
- Test copy for scenario 1 (depends on #154)
- Test copy for baseline state for alpha release (depends on #158)
### Acceptance Criteria
- [x] Lead interviews
- [ ] #273
- [ ] #356
<!--
_Note_ When you create this issue, remember to add:
- an assignee
- the project, so that it will show up in our kanban view
- a label for story points estimate (or comment at the assignee to request that they add an estimate)
- a label for priority
-->
|
design
|
run rapid usability testing description facilitate content testing for test copy for scenario depends on test copy for baseline state for alpha release depends on acceptance criteria lead interviews note when you create this issue remember to add an assignee the project so that it will show up in our kanban view a label for story points estimate or comment at the assignee to request that they add an estimate a label for priority
| 1
|
20,042
| 26,529,381,488
|
IssuesEvent
|
2023-01-19 11:15:17
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Introspection of MySQL views
|
process/candidate topic: introspection topic: re-introspection tech/engines/introspection engine team/schema topic: view kind/subtask
|
We should fit this beauty to the SQL describer:
```sql
select col.table_schema as database_name,
col.table_name as view_name,
col.ordinal_position,
col.column_name,
col.data_type,
case when col.character_maximum_length is not null
then col.character_maximum_length
else col.numeric_precision end as max_length,
col.is_nullable
from information_schema.columns col
join information_schema.views vie on vie.table_schema = col.table_schema
and vie.table_name = col.table_name
where col.table_schema not in ('sys','information_schema',
'mysql', 'performance_schema')
-- and vie.table_schema = 'database_name' -- put your database name here
order by col.table_schema,
col.table_name,
col.ordinal_position;
```
There can be relations between views, or models and views. These are not based on foreign keys. They should be kept in re-introspection, for now.
Part of: https://github.com/prisma/prisma/issues/17412
|
1.0
|
Introspection of MySQL views - We should fit this beauty to the SQL describer:
```sql
select col.table_schema as database_name,
col.table_name as view_name,
col.ordinal_position,
col.column_name,
col.data_type,
case when col.character_maximum_length is not null
then col.character_maximum_length
else col.numeric_precision end as max_length,
col.is_nullable
from information_schema.columns col
join information_schema.views vie on vie.table_schema = col.table_schema
and vie.table_name = col.table_name
where col.table_schema not in ('sys','information_schema',
'mysql', 'performance_schema')
-- and vie.table_schema = 'database_name' -- put your database name here
order by col.table_schema,
col.table_name,
col.ordinal_position;
```
There can be relations between views, or models and views. These are not based on foreign keys. They should be kept in re-introspection, for now.
Part of: https://github.com/prisma/prisma/issues/17412
|
non_design
|
introspection of mysql views we should fit this beauty to the sql describer sql select col table schema as database name col table name as view name col ordinal position col column name col data type case when col character maximum length is not null then col character maximum length else col numeric precision end as max length col is nullable from information schema columns col join information schema views vie on vie table schema col table schema and vie table name col table name where col table schema not in sys information schema mysql performance schema and vie table schema database name put your database name here order by col table schema col table name col ordinal position there can be relations between views or models and views these are not based on foreign keys they should be kept in re introspection for now part of
| 0
|
64,215
| 6,896,120,404
|
IssuesEvent
|
2017-11-23 16:20:33
|
medic/medic-webapp
|
https://api.github.com/repos/medic/medic-webapp
|
closed
|
Migration extract-person-contacts broken
|
Priority: 1 - High Status: 4 - Acceptance testing Type: Bug
|
The server details are shared on slack if further investigation is required.
Migration from Version 0.4 to Version 2.13
Error:
```
2017-11-10T12:18:21.578Z - error: Failed to restore contact on facility 42fbf87e-9857-8ad1-071fc2f9e15ac6b9, contact: {"name":"Redacted ","phone":"+251920---","type":"person","reported_date":1510316301222}
Migration extract-person-contacts failed
Fatal error initialising medic-api
Error: Failed to update contact on facility 42fbf87e-9857-8ad1-071fc2f9e15ac6b9: {
"code": 400,
"message": "Place 42fbf87e-9857-8ad1-071fc2f9e15ac6b9 is missing a \"parent\" property."
}
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/migrations/extract-person-contacts.js:109:20
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/migrations/extract-person-contacts.js:123:7
at async.series.err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:276:7)
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:3838:9
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:421:16
at iterateeCallback (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:924:17)
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:906:16
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:3835:13
at self._validatePlace.err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:263:18)
at err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:99:12)
```
However, the document when requested from the original server is:
```
curl https://user:pass@amref-ethiopia.app.medicmobile.org/medic/42fbf87e-9857-8ad1-071fc2f9e15ac6b9 | jq
{
"_id": "42fbf87e-9857-8ad1-071fc2f9e15ac6b9",
"_rev": "433-068bcc2c2bbd7babd166f410959f3b2f",
"type": "clinic",
"name": "Redacted ",
"contact": {
"name": "Redacted ",
"phone": "+25192----"
},
"parent": {
"_id": "42fbf87e-9857-8ad1-071fc2f9e15a5e26",
"_rev": "15-0f2d0e53d04251e7171a51b76c881dbb",
"type": "health_center",
"name": "Redacted",
"contact": {
"name": "Redacted",
"phone": "+25192----"
},
"parent": {
"_id": "eec1f64b-b093-0a25-f0bd72d7f6d68010",
"_rev": "3-c2fd98be72a3ff7ececf452c67f68f5a",
"type": "district_hospital",
"name": "Redacted",
"parent": {},
"contact": {
"name": "Redacted",
"phone": "+25198-----"
}
}
},
"sent_forms": {
"ANCR": "2015-12-12T11:00:18.886Z",
"F": "2015-08-06T11:08:45.000Z"
}
}
```
Let's request the parent object to confirm that it's an actual document in the database.
```
henok$ curl https://user:pass@amref-ethiopia.app.medicmobile.org/medic/eec1f64b-b093-0a25-f0bd72d7f6d68010 | jq
{
"_id": "eec1f64b-b093-0a25-f0bd72d7f6d68010",
"_rev": "3-c2fd98be72a3ff7ececf452c67f68f5a",
"type": "district_hospital",
"name": "Redacted",
"parent": {},
"contact": {
"name": "Redacted",
"phone": "+25198-----"
}
}
```
|
1.0
|
Migration extract-person-contacts broken - The server details are shared on slack if further investigation is required.
Migration from Version 0.4 to Version 2.13
Error:
```
2017-11-10T12:18:21.578Z - error: Failed to restore contact on facility 42fbf87e-9857-8ad1-071fc2f9e15ac6b9, contact: {"name":"Redacted ","phone":"+251920---","type":"person","reported_date":1510316301222}
Migration extract-person-contacts failed
Fatal error initialising medic-api
Error: Failed to update contact on facility 42fbf87e-9857-8ad1-071fc2f9e15ac6b9: {
"code": 400,
"message": "Place 42fbf87e-9857-8ad1-071fc2f9e15ac6b9 is missing a \"parent\" property."
}
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/migrations/extract-person-contacts.js:109:20
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/migrations/extract-person-contacts.js:123:7
at async.series.err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:276:7)
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:3838:9
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:421:16
at iterateeCallback (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:924:17)
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:906:16
at /srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/node_modules/async/dist/async.js:3835:13
at self._validatePlace.err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:263:18)
at err (/srv/storage/gardener/data/working_dir/aHR0cDovL2xvY2FsaG9zdDo1OTg0L21lZGljL19kZXNpZ24vbWVkaWMvbWVkaWMtYXBp/node_modules/medic-api/controllers/places.js:99:12)
```
However, the document when requested from the original server is:
```
curl https://user:pass@amref-ethiopia.app.medicmobile.org/medic/42fbf87e-9857-8ad1-071fc2f9e15ac6b9 | jq
{
"_id": "42fbf87e-9857-8ad1-071fc2f9e15ac6b9",
"_rev": "433-068bcc2c2bbd7babd166f410959f3b2f",
"type": "clinic",
"name": "Redacted ",
"contact": {
"name": "Redacted ",
"phone": "+25192----"
},
"parent": {
"_id": "42fbf87e-9857-8ad1-071fc2f9e15a5e26",
"_rev": "15-0f2d0e53d04251e7171a51b76c881dbb",
"type": "health_center",
"name": "Redacted",
"contact": {
"name": "Redacted",
"phone": "+25192----"
},
"parent": {
"_id": "eec1f64b-b093-0a25-f0bd72d7f6d68010",
"_rev": "3-c2fd98be72a3ff7ececf452c67f68f5a",
"type": "district_hospital",
"name": "Redacted",
"parent": {},
"contact": {
"name": "Redacted",
"phone": "+25198-----"
}
}
},
"sent_forms": {
"ANCR": "2015-12-12T11:00:18.886Z",
"F": "2015-08-06T11:08:45.000Z"
}
}
```
Let's request the parent object to confirm that it's an actual document in the database.
```
henok$ curl https://user:pass@amref-ethiopia.app.medicmobile.org/medic/eec1f64b-b093-0a25-f0bd72d7f6d68010 | jq
{
"_id": "eec1f64b-b093-0a25-f0bd72d7f6d68010",
"_rev": "3-c2fd98be72a3ff7ececf452c67f68f5a",
"type": "district_hospital",
"name": "Redacted",
"parent": {},
"contact": {
"name": "Redacted",
"phone": "+25198-----"
}
}
```
|
non_design
|
migration extract person contacts broken the server details are shared on slack if further investigation is required migration from version to version error error failed to restore contact on facility contact name redacted phone type person reported date migration extract person contacts failed fatal error initialising medic api error failed to update contact on facility code message place is missing a parent property at srv storage gardener data working dir node modules medic api migrations extract person contacts js at srv storage gardener data working dir node modules medic api migrations extract person contacts js at async series err srv storage gardener data working dir node modules medic api controllers places js at srv storage gardener data working dir node modules medic api node modules async dist async js at srv storage gardener data working dir node modules medic api node modules async dist async js at iterateecallback srv storage gardener data working dir node modules medic api node modules async dist async js at srv storage gardener data working dir node modules medic api node modules async dist async js at srv storage gardener data working dir node modules medic api node modules async dist async js at self validateplace err srv storage gardener data working dir node modules medic api controllers places js at err srv storage gardener data working dir node modules medic api controllers places js however the document when requested from the original server is curl jq id rev type clinic name redacted contact name redacted phone parent id rev type health center name redacted contact name redacted phone parent id rev type district hospital name redacted parent contact name redacted phone sent forms ancr f let s request the parent object to confirm that it s an actual document in the database henok curl jq id rev type district hospital name redacted parent contact name redacted phone
| 0
|
126,921
| 17,142,185,581
|
IssuesEvent
|
2021-07-13 10:51:43
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
Color panel should be hidden when the block's color are disabled
|
[Feature] Design Tools [Status] In Progress [Type] Bug
|
In working at https://github.com/WordPress/gutenberg/pull/33280/ and preparing https://github.com/WordPress/gutenberg/pull/33295 I've realized the color panel doesn't work as expected.
If a theme passes the following data via `theme.json`:
```json
{
"version": 1,
"settings": {
"color": {
"custom": false,
"customGradient": false,
"gradients": [],
"link": false,
"palette": []
}
}
```
the result is that the color panel is hidden.
However, let's look at a block such as the paragraph. It supports text, background, and link colors but not gradients. I'd expect that disabling only the things it supports should hide the color panel (notice that there's no `customGradient` or `gradients` in the following `theme.json`):
```json
{
"version": 1,
"settings": {
"color": {
"custom": false,
"link": false,
"palette": []
}
}
```
What happens instead is that the color panel is shown empty:

|
1.0
|
Color panel should be hidden when the block's color are disabled - In working at https://github.com/WordPress/gutenberg/pull/33280/ and preparing https://github.com/WordPress/gutenberg/pull/33295 I've realized the color panel doesn't work as expected.
If a theme passes the following data via `theme.json`:
```json
{
"version": 1,
"settings": {
"color": {
"custom": false,
"customGradient": false,
"gradients": [],
"link": false,
"palette": []
}
}
```
the result is that the color panel is hidden.
However, let's look at a block such as the paragraph. It supports text, background, and link colors but not gradients. I'd expect that disabling only the things it supports should hide the color panel (notice that there's no `customGradient` or `gradients` in the following `theme.json`):
```json
{
"version": 1,
"settings": {
"color": {
"custom": false,
"link": false,
"palette": []
}
}
```
What happens instead is that the color panel is shown empty:

|
design
|
color panel should be hidden when the block s color are disabled in working at and preparing i ve realized the color panel doesn t work as expected if a theme passes the following data via theme json json version settings color custom false customgradient false gradients link false palette the result is that the color panel is hidden however let s look at a block such as the paragraph it supports text background and link colors but not gradients i d expect that disabling only the things it supports should hide the color panel notice that there s no customgradient or gradients in the following theme json json version settings color custom false link false palette what happens instead is that the color panel is shown empty
| 1
|
94,730
| 19,577,671,159
|
IssuesEvent
|
2022-01-04 17:03:14
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
closed
|
extreme number of draw calls per frame
|
Code Performance
|
- [X] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Game uses a very large number of draw calls (5200-6500) in new-old saves, many more used when drawing ui.
**Steps To Reproduce**
```
$ apitrace Barotrauma
<open up save or create a new game>
$ qapitrace Barotrauma.trace
```
**Version**
v0.15.22.1, debian sid
**Additional information**
Running barotrauma on a crappy 2.5ghz quad core cpu (1 core contantly maxed while playing)
I decided to do some **basic** profiling to check out some causes of lag.
I used `apitrace trace ./Barotrauma` to check how many draw calls the game uses, around 120 for loading screen which is pretty good but around 6500(!) for a singleplayer campaign game on the Typhon with 6 crew members.
This results in around 20 fps (on both 1920x1080 and 1280x720), in comparison to a new game (Humpback) with around 45 fps and 5200 draw calls. I'm sure a part of this slowdown is caused by the increased draw calls, but obviously not all of it.
I also noticed that when opening containers with lots of items it gets a bit slower (an increase of ~400 draw calls for 13 items in a medical cabinet, roughly checked) which would indicate that UI drawing probably isn't being batched.
It seems like making more drawing batched would be a nice way to reduce draw calls and gain a few fps.
|
1.0
|
extreme number of draw calls per frame - - [X] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Game uses a very large number of draw calls (5200-6500) in new-old saves, many more used when drawing ui.
**Steps To Reproduce**
```
$ apitrace Barotrauma
<open up save or create a new game>
$ qapitrace Barotrauma.trace
```
**Version**
v0.15.22.1, debian sid
**Additional information**
Running barotrauma on a crappy 2.5ghz quad core cpu (1 core contantly maxed while playing)
I decided to do some **basic** profiling to check out some causes of lag.
I used `apitrace trace ./Barotrauma` to check how many draw calls the game uses, around 120 for loading screen which is pretty good but around 6500(!) for a singleplayer campaign game on the Typhon with 6 crew members.
This results in around 20 fps (on both 1920x1080 and 1280x720), in comparison to a new game (Humpback) with around 45 fps and 5200 draw calls. I'm sure a part of this slowdown is caused by the increased draw calls, but obviously not all of it.
I also noticed that when opening containers with lots of items it gets a bit slower (an increase of ~400 draw calls for 13 items in a medical cabinet, roughly checked) which would indicate that UI drawing probably isn't being batched.
It seems like making more drawing batched would be a nice way to reduce draw calls and gain a few fps.
|
non_design
|
extreme number of draw calls per frame i have searched the issue tracker to check if the issue has already been reported description game uses a very large number of draw calls in new old saves many more used when drawing ui steps to reproduce apitrace barotrauma qapitrace barotrauma trace version debian sid additional information running barotrauma on a crappy quad core cpu core contantly maxed while playing i decided to do some basic profiling to check out some causes of lag i used apitrace trace barotrauma to check how many draw calls the game uses around for loading screen which is pretty good but around for a singleplayer campaign game on the typhon with crew members this results in around fps on both and in comparison to a new game humpback with around fps and draw calls i m sure a part of this slowdown is caused by the increased draw calls but obviously not all of it i also noticed that when opening containers with lots of items it gets a bit slower an increase of draw calls for items in a medical cabinet roughly checked which would indicate that ui drawing probably isn t being batched it seems like making more drawing batched would be a nice way to reduce draw calls and gain a few fps
| 0
|
69,014
| 8,367,933,114
|
IssuesEvent
|
2018-10-04 13:36:20
|
gctools-outilsgc/design-system
|
https://api.github.com/repos/gctools-outilsgc/design-system
|
closed
|
navigation templates and layouts
|
Project: Design System [zube]: Backlog layout systems phase II
|
Guidelines and wireframes for application and site navigation
|
1.0
|
navigation templates and layouts - Guidelines and wireframes for application and site navigation
|
design
|
navigation templates and layouts guidelines and wireframes for application and site navigation
| 1
|
332,960
| 24,356,799,292
|
IssuesEvent
|
2022-10-03 08:13:00
|
pravega/pravega
|
https://api.github.com/repos/pravega/pravega
|
closed
|
Create document to recover container exhibiting Table Segment credits exhausted
|
area/segmentstore area/documentation priority/P2 area/data-recovery version/0.13.0
|
**Description of the task**
We need to create a document under the "recovery_procedures" section that describes how to recover a Segment Container for which internal metadata Table Segments cannot be initialized due to exhaustion of credits (i.e., too much data has been accumulated in Tier-1 that has not been moved to LTS).
|
1.0
|
Create document to recover container exhibiting Table Segment credits exhausted - **Description of the task**
We need to create a document under the "recovery_procedures" section that describes how to recover a Segment Container for which internal metadata Table Segments cannot be initialized due to exhaustion of credits (i.e., too much data has been accumulated in Tier-1 that has not been moved to LTS).
|
non_design
|
create document to recover container exhibiting table segment credits exhausted description of the task we need to create a document under the recovery procedures section that describes how to recover a segment container for which internal metadata table segments cannot be initialized due to exhaustion of credits i e too much data has been accumulated in tier that has not been moved to lts
| 0
|
169,047
| 26,738,209,531
|
IssuesEvent
|
2023-01-30 11:00:57
|
BulletBoxProject/BulletBox-project-FE-
|
https://api.github.com/repos/BulletBoxProject/BulletBox-project-FE-
|
closed
|
데일리로그 날짜 선택 기능
|
feat design
|
## Description
데일리로그 날짜 선택 기능 구현 작업입니다.
## Todo
- [x] 캘린더 드롭다운 생성
- [x] 날짜 선택해서 이동 기능
## Etc
기타사항
|
1.0
|
데일리로그 날짜 선택 기능 - ## Description
데일리로그 날짜 선택 기능 구현 작업입니다.
## Todo
- [x] 캘린더 드롭다운 생성
- [x] 날짜 선택해서 이동 기능
## Etc
기타사항
|
design
|
데일리로그 날짜 선택 기능 description 데일리로그 날짜 선택 기능 구현 작업입니다 todo 캘린더 드롭다운 생성 날짜 선택해서 이동 기능 etc 기타사항
| 1
|
34
| 2,492,581,956
|
IssuesEvent
|
2015-01-05 01:50:29
|
lakkatv/Lakka
|
https://api.github.com/repos/lakkatv/Lakka
|
closed
|
Handle blank config
|
bug design enhancement menu
|
For our menu to work, you have to configure paths in the retroarch config. If the right paths are not set, it is not able to do anything and even segfaults. This is a bug because most users have an empty config file when launching retroarch for the first time.
The right way to fix it I think, is to:
- display a helper at the first launch, as rgui already does
- implement configuration editing (we must do a mockup before starting coding that)
Providing a small configuration wizard while path are not set would be OK too.
Anyway, it is a lot of work.
|
1.0
|
Handle blank config - For our menu to work, you have to configure paths in the retroarch config. If the right paths are not set, it is not able to do anything and even segfaults. This is a bug because most users have an empty config file when launching retroarch for the first time.
The right way to fix it I think, is to:
- display a helper at the first launch, as rgui already does
- implement configuration editing (we must do a mockup before starting coding that)
Providing a small configuration wizard while path are not set would be OK too.
Anyway, it is a lot of work.
|
design
|
handle blank config for our menu to work you have to configure paths in the retroarch config if the right paths are not set it is not able to do anything and even segfaults this is a bug because most users have an empty config file when launching retroarch for the first time the right way to fix it i think is to display a helper at the first launch as rgui already does implement configuration editing we must do a mockup before starting coding that providing a small configuration wizard while path are not set would be ok too anyway it is a lot of work
| 1
|
609,840
| 18,888,810,406
|
IssuesEvent
|
2021-11-15 10:53:15
|
CatalogueOfLife/backend
|
https://api.github.com/repos/CatalogueOfLife/backend
|
closed
|
Search exceptions for large offsets
|
bug high priority search
|
http://api.catalogueoflife.org/dataset/2349/nameusage/search?facet=rank&facet=issue&facet=status&facet=nomStatus&facet=nameType&facet=field&facet=authorship&facet=extinct&facet=environment&limit=50&offset=4418350&reverse=false&sortBy=taxonomic&status=_NOT_NULL
```
{"type": "server", "timestamp": "2021-11-15T10:32:19,328Z", "level": "DEBUG", "component": "o.e.a.s.TransportSearchAction", "cluster.name": "col", "node.name": "node-col", "message": "[xTwUbd7mSvuFGdFtYPbj8A][col-2021-03-26][0]: Failed to execute [SearchRequest{searchType=QUERY_THEN_FETCH, indices=[prod-nu], indicesOptions=IndicesOptions[ignore_unavailable=false, allow_no_indices=true, expand_wildcards_open=true, expand_wildcards_closed=false, expand_wildcards_hidden=false, allow_aliases_to_multiple_indices=true, forbid_closed_indices=true, ignore_aliases=false, ignore_throttled=true], types=[], routing='null', preference='null', requestCache=null, scroll=null, maxConcurrentShardRequests=0, batchedReduceSize=512, preFilterShardSize=null, allowPartialSearchResults=true, localClusterAlias=null, getOrCreateAbsoluteStartMillis=-1, ccsMinimizeRoundtrips=true, source={\"from\":4418350,\"size\":50,\"query\":{\"bool\":{\"filter\":[{\"term\":{\"datasetKey\":{\"value\":2349,\"boost\":1.0}}},{\"exists\":{\"field\":\"status\",\"boost\":1.0}}],\"adjust_pure_negative\":true,\"boost\":1.0}},\"sort\":[{\"rank\":{\"order\":\"asc\"}},{\"scientificName\":{\"order\":\"asc\"}}],\"track_total_hits\":2147483647,\"aggregations\":{\"_global_\":{\"global\":{},\"aggregations\":{\"_filter_\":{\"filter\":{\"term\":{\"datasetKey\":{\"value\":2349,\"boost\":1.0}}},\"aggregations\":{\"FIELD\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"nameFields\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"ISSUE\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"issues\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"NOM_STATUS\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"nomStatus\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"RANK\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"rank\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"STATUS\":{\"filter\":{\"match_all\":{\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"status\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"NAME_TYPE\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"type\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"EXTINCT\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"extinct\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"ENVIRONMENT\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"environments\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"AUTHORSHIP\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"authorship\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}}}}}}}}}] lastShard [true]", "cluster.uuid": "VSaAkBhoQg67yXoFlbafkg", "node.id": "xTwUbd7mSvuFGdFtYPbj8A" ,
"stacktrace": ["org.elasticsearch.transport.RemoteTransportException: [node-col][130.225.43.29:9300][indices:data/read/search[phase/query]]",
"Caused by: java.lang.IllegalArgumentException: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.search.DefaultSearchContext.preProcess(DefaultSearchContext.java:192) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.query.QueryPhase.preProcess(QueryPhase.java:113) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.createContext(SearchService.java:732) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:422) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.access$500(SearchService.java:135) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService$2.lambda$onResponse$0(SearchService.java:395) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.lambda$supply$0(ActionRunnable.java:47) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable$2.doRun(ActionRunnable.java:62) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.TimedRunnable.doRun(TimedRunnable.java:33) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:732) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) [?:?]",
"at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) [?:?]",
"at java.lang.Thread.run(Thread.java:832) [?:?]"] }
{"type": "server", "timestamp": "2021-11-15T10:32:19,339Z", "level": "DEBUG", "component": "o.e.a.s.TransportSearchAction", "cluster.name": "col", "node.name": "node-col", "message": "All shards failed for phase: [query]", "cluster.uuid": "VSaAkBhoQg67yXoFlbafkg", "node.id": "xTwUbd7mSvuFGdFtYPbj8A" ,
"stacktrace": ["org.elasticsearch.ElasticsearchException$1: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.ElasticsearchException.guessRootCauses(ElasticsearchException.java:633) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:330) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:636) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.onShardFailure(AbstractSearchAsyncAction.java:415) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.access$000(AbstractSearchAsyncAction.java:59) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction$1.onFailure(AbstractSearchAsyncAction.java:264) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.SearchExecutionStatsCollector.onFailure(SearchExecutionStatsCollector.java:62) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionListenerResponseHandler.handleException(ActionListenerResponseHandler.java:48) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleException(SearchTransportService.java:404) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$6.handleException(TransportService.java:743) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleException(TransportService.java:1288) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$DirectResponseChannel.processException(TransportService.java:1397) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1371) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:50) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportChannel.sendErrorResponse(TransportChannel.java:45) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.support.ChannelActionListener.onFailure(ChannelActionListener.java:40) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.onFailure(ActionRunnable.java:77) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:28) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.TimedRunnable.doRun(TimedRunnable.java:33) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:732) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) [?:?]",
"at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) [?:?]",
"at java.lang.Thread.run(Thread.java:832) [?:?]",
"Caused by: java.lang.IllegalArgumentException: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.search.DefaultSearchContext.preProcess(DefaultSearchContext.java:192) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.query.QueryPhase.preProcess(QueryPhase.java:113) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.createContext(SearchService.java:732) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:422) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.access$500(SearchService.java:135) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService$2.lambda$onResponse$0(SearchService.java:395) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.lambda$supply$0(ActionRunnable.java:47) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable$2.doRun(ActionRunnable.java:62) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) ~[elasticsearch-7.11.2.jar:7.11.2]",
"... 6 more"] }
```
|
1.0
|
Search exceptions for large offsets - http://api.catalogueoflife.org/dataset/2349/nameusage/search?facet=rank&facet=issue&facet=status&facet=nomStatus&facet=nameType&facet=field&facet=authorship&facet=extinct&facet=environment&limit=50&offset=4418350&reverse=false&sortBy=taxonomic&status=_NOT_NULL
```
{"type": "server", "timestamp": "2021-11-15T10:32:19,328Z", "level": "DEBUG", "component": "o.e.a.s.TransportSearchAction", "cluster.name": "col", "node.name": "node-col", "message": "[xTwUbd7mSvuFGdFtYPbj8A][col-2021-03-26][0]: Failed to execute [SearchRequest{searchType=QUERY_THEN_FETCH, indices=[prod-nu], indicesOptions=IndicesOptions[ignore_unavailable=false, allow_no_indices=true, expand_wildcards_open=true, expand_wildcards_closed=false, expand_wildcards_hidden=false, allow_aliases_to_multiple_indices=true, forbid_closed_indices=true, ignore_aliases=false, ignore_throttled=true], types=[], routing='null', preference='null', requestCache=null, scroll=null, maxConcurrentShardRequests=0, batchedReduceSize=512, preFilterShardSize=null, allowPartialSearchResults=true, localClusterAlias=null, getOrCreateAbsoluteStartMillis=-1, ccsMinimizeRoundtrips=true, source={\"from\":4418350,\"size\":50,\"query\":{\"bool\":{\"filter\":[{\"term\":{\"datasetKey\":{\"value\":2349,\"boost\":1.0}}},{\"exists\":{\"field\":\"status\",\"boost\":1.0}}],\"adjust_pure_negative\":true,\"boost\":1.0}},\"sort\":[{\"rank\":{\"order\":\"asc\"}},{\"scientificName\":{\"order\":\"asc\"}}],\"track_total_hits\":2147483647,\"aggregations\":{\"_global_\":{\"global\":{},\"aggregations\":{\"_filter_\":{\"filter\":{\"term\":{\"datasetKey\":{\"value\":2349,\"boost\":1.0}}},\"aggregations\":{\"FIELD\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"nameFields\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"ISSUE\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"issues\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"NOM_STATUS\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"nomStatus\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"RANK\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"rank\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"STATUS\":{\"filter\":{\"match_all\":{\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"status\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"NAME_TYPE\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"type\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"EXTINCT\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"extinct\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"ENVIRONMENT\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"environments\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}},\"AUTHORSHIP\":{\"filter\":{\"exists\":{\"field\":\"status\",\"boost\":1.0}},\"aggregations\":{\"_values_\":{\"terms\":{\"field\":\"authorship\",\"size\":50,\"min_doc_count\":1,\"shard_min_doc_count\":0,\"show_term_doc_count_error\":false,\"order\":[{\"_count\":\"desc\"},{\"_key\":\"asc\"}]}}}}}}}}}}}] lastShard [true]", "cluster.uuid": "VSaAkBhoQg67yXoFlbafkg", "node.id": "xTwUbd7mSvuFGdFtYPbj8A" ,
"stacktrace": ["org.elasticsearch.transport.RemoteTransportException: [node-col][130.225.43.29:9300][indices:data/read/search[phase/query]]",
"Caused by: java.lang.IllegalArgumentException: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.search.DefaultSearchContext.preProcess(DefaultSearchContext.java:192) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.query.QueryPhase.preProcess(QueryPhase.java:113) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.createContext(SearchService.java:732) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:422) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.access$500(SearchService.java:135) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService$2.lambda$onResponse$0(SearchService.java:395) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.lambda$supply$0(ActionRunnable.java:47) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable$2.doRun(ActionRunnable.java:62) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.TimedRunnable.doRun(TimedRunnable.java:33) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:732) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) [?:?]",
"at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) [?:?]",
"at java.lang.Thread.run(Thread.java:832) [?:?]"] }
{"type": "server", "timestamp": "2021-11-15T10:32:19,339Z", "level": "DEBUG", "component": "o.e.a.s.TransportSearchAction", "cluster.name": "col", "node.name": "node-col", "message": "All shards failed for phase: [query]", "cluster.uuid": "VSaAkBhoQg67yXoFlbafkg", "node.id": "xTwUbd7mSvuFGdFtYPbj8A" ,
"stacktrace": ["org.elasticsearch.ElasticsearchException$1: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.ElasticsearchException.guessRootCauses(ElasticsearchException.java:633) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:330) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:636) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.onShardFailure(AbstractSearchAsyncAction.java:415) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction.access$000(AbstractSearchAsyncAction.java:59) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.AbstractSearchAsyncAction$1.onFailure(AbstractSearchAsyncAction.java:264) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.SearchExecutionStatsCollector.onFailure(SearchExecutionStatsCollector.java:62) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionListenerResponseHandler.handleException(ActionListenerResponseHandler.java:48) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleException(SearchTransportService.java:404) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$6.handleException(TransportService.java:743) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleException(TransportService.java:1288) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$DirectResponseChannel.processException(TransportService.java:1397) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1371) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:50) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.transport.TransportChannel.sendErrorResponse(TransportChannel.java:45) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.support.ChannelActionListener.onFailure(ChannelActionListener.java:40) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.onFailure(ActionRunnable.java:77) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:28) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.TimedRunnable.doRun(TimedRunnable.java:33) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:732) [elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) [elasticsearch-7.11.2.jar:7.11.2]",
"at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) [?:?]",
"at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) [?:?]",
"at java.lang.Thread.run(Thread.java:832) [?:?]",
"Caused by: java.lang.IllegalArgumentException: Result window is too large, from + size must be less than or equal to: [2000000] but was [4418400]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting.",
"at org.elasticsearch.search.DefaultSearchContext.preProcess(DefaultSearchContext.java:192) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.query.QueryPhase.preProcess(QueryPhase.java:113) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.createContext(SearchService.java:732) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:422) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService.access$500(SearchService.java:135) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.search.SearchService$2.lambda$onResponse$0(SearchService.java:395) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable.lambda$supply$0(ActionRunnable.java:47) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.action.ActionRunnable$2.doRun(ActionRunnable.java:62) ~[elasticsearch-7.11.2.jar:7.11.2]",
"at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:26) ~[elasticsearch-7.11.2.jar:7.11.2]",
"... 6 more"] }
```
|
non_design
|
search exceptions for large offsets type server timestamp level debug component o e a s transportsearchaction cluster name col node name node col message failed to execute indicesoptions indicesoptions types routing null preference null requestcache null scroll null maxconcurrentshardrequests batchedreducesize prefiltershardsize null allowpartialsearchresults true localclusteralias null getorcreateabsolutestartmillis ccsminimizeroundtrips true source from size query bool filter adjust pure negative true boost sort track total hits aggregations global global aggregations filter filter term datasetkey value boost aggregations field filter exists field status boost aggregations values terms field namefields size min doc count shard min doc count show term doc count error false order issue filter exists field status boost aggregations values terms field issues size min doc count shard min doc count show term doc count error false order nom status filter exists field status boost aggregations values terms field nomstatus size min doc count shard min doc count show term doc count error false order rank filter exists field status boost aggregations values terms field rank size min doc count shard min doc count show term doc count error false order status filter match all boost aggregations values terms field status size min doc count shard min doc count show term doc count error false order name type filter exists field status boost aggregations values terms field type size min doc count shard min doc count show term doc count error false order extinct filter exists field status boost aggregations values terms field extinct size min doc count shard min doc count show term doc count error false order environment filter exists field status boost aggregations values terms field environments size min doc count shard min doc count show term doc count error false order authorship filter exists field status boost aggregations values terms field authorship size min doc count shard min doc count show term doc count error false order lastshard cluster uuid node id stacktrace caused by java lang illegalargumentexception result window is too large from size must be less than or equal to but was see the scroll api for a more efficient way to request large data sets this limit can be set by changing the index level setting at org elasticsearch search defaultsearchcontext preprocess defaultsearchcontext java at org elasticsearch search query queryphase preprocess queryphase java at org elasticsearch search searchservice createcontext searchservice java at org elasticsearch search searchservice executequeryphase searchservice java at org elasticsearch search searchservice access searchservice java at org elasticsearch search searchservice lambda onresponse searchservice java at org elasticsearch action actionrunnable lambda supply actionrunnable java at org elasticsearch action actionrunnable dorun actionrunnable java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at org elasticsearch common util concurrent timedrunnable dorun timedrunnable java at org elasticsearch common util concurrent threadcontext contextpreservingabstractrunnable dorun threadcontext java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java type server timestamp level debug component o e a s transportsearchaction cluster name col node name node col message all shards failed for phase cluster uuid node id stacktrace but was see the scroll api for a more efficient way to request large data sets this limit can be set by changing the index level setting at org elasticsearch elasticsearchexception guessrootcauses elasticsearchexception java at org elasticsearch action search abstractsearchasyncaction executenextphase abstractsearchasyncaction java at org elasticsearch action search abstractsearchasyncaction onphasedone abstractsearchasyncaction java at org elasticsearch action search abstractsearchasyncaction onshardfailure abstractsearchasyncaction java at org elasticsearch action search abstractsearchasyncaction access abstractsearchasyncaction java at org elasticsearch action search abstractsearchasyncaction onfailure abstractsearchasyncaction java at org elasticsearch action search searchexecutionstatscollector onfailure searchexecutionstatscollector java at org elasticsearch action actionlistenerresponsehandler handleexception actionlistenerresponsehandler java at org elasticsearch action search searchtransportservice connectioncountinghandler handleexception searchtransportservice java at org elasticsearch transport transportservice handleexception transportservice java at org elasticsearch transport transportservice contextrestoreresponsehandler handleexception transportservice java at org elasticsearch transport transportservice directresponsechannel processexception transportservice java at org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java at org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java at org elasticsearch transport transportchannel senderrorresponse transportchannel java at org elasticsearch action support channelactionlistener onfailure channelactionlistener java at org elasticsearch action actionrunnable onfailure actionrunnable java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at org elasticsearch common util concurrent timedrunnable dorun timedrunnable java at org elasticsearch common util concurrent threadcontext contextpreservingabstractrunnable dorun threadcontext java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java lang illegalargumentexception result window is too large from size must be less than or equal to but was see the scroll api for a more efficient way to request large data sets this limit can be set by changing the index level setting at org elasticsearch search defaultsearchcontext preprocess defaultsearchcontext java at org elasticsearch search query queryphase preprocess queryphase java at org elasticsearch search searchservice createcontext searchservice java at org elasticsearch search searchservice executequeryphase searchservice java at org elasticsearch search searchservice access searchservice java at org elasticsearch search searchservice lambda onresponse searchservice java at org elasticsearch action actionrunnable lambda supply actionrunnable java at org elasticsearch action actionrunnable dorun actionrunnable java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java more
| 0
|
677,612
| 23,167,734,880
|
IssuesEvent
|
2022-07-30 07:50:12
|
ObsidianMC/Obsidian
|
https://api.github.com/repos/ObsidianMC/Obsidian
|
closed
|
Implement new chat message/command changes (1.19 branch)
|
enhancement help wanted good first issue priority: high networking
|
As the title suggests I'm opening up this issue for someone who wants to do this before I get to it myself.
Helpful Links:
https://wiki.vg/images/f/f4/MinecraftChat.drawio4.png
https://wiki.vg/Chat#Processing_chat
Packets that should be looked at (Classes for these packets are created already)
https://wiki.vg/Protocol#Chat_Message
https://wiki.vg/Protocol#Player_Chat_Message
https://wiki.vg/Protocol#Chat_Command
This should work reliably with online mode enabled/disabled
|
1.0
|
Implement new chat message/command changes (1.19 branch) - As the title suggests I'm opening up this issue for someone who wants to do this before I get to it myself.
Helpful Links:
https://wiki.vg/images/f/f4/MinecraftChat.drawio4.png
https://wiki.vg/Chat#Processing_chat
Packets that should be looked at (Classes for these packets are created already)
https://wiki.vg/Protocol#Chat_Message
https://wiki.vg/Protocol#Player_Chat_Message
https://wiki.vg/Protocol#Chat_Command
This should work reliably with online mode enabled/disabled
|
non_design
|
implement new chat message command changes branch as the title suggests i m opening up this issue for someone who wants to do this before i get to it myself helpful links packets that should be looked at classes for these packets are created already this should work reliably with online mode enabled disabled
| 0
|
257,957
| 22,265,919,078
|
IssuesEvent
|
2022-06-10 07:25:35
|
gravitee-io/issues
|
https://api.github.com/repos/gravitee-io/issues
|
closed
|
[gateway] conditional logging on date/Duration prevents to display API logs
|
type: bug project: APIM Support 2 p2 loop quantum status: in test
|
## :collision: Describe the bug
When enabling logging on an API with a Condition on an date/Duration, the detailed logs are not displayed and the following warning is raise in the gateway logs :
```
gio_apim_gateway-3.16.1 | 10:17:52.956 [vert.x-eventloop-thread-4] [] WARN i.g.g.c.l.p.LoggableRequestProcessor - Unexpected error while evaluating logging condition for the API e410fbd5-725e-43b5-90fb-d5725ea3b5e0 and context path /echo/ : EL1041E: After parsing a valid expression, there is still more data in the expression: 'Mar'
```
Condition is displayed as :
{#request.timestamp <= Fri Mar 25 2022 14:00:15 `GMT+0100l}
Workaround is to convert to a timestamp in Milliseconds as :
{#request.timestamp <= 1647944975000l}
## :sunrise_over_mountains: To Reproduce
Steps to reproduce the behaviour:
1. Create an API with API Key Plan
2. subscribe to it with your Default Application
3. Enable Logging on the API and add a Condition on the Date/Duration
4. Make a request on the API and check the logs
## :rainbow: Expected behaviour
A clear and concise description of what you expected to happen.
## Current behaviour
A clear and concise description of what is currently happening.
## :movie_camera: Useful information
Screenshot, video, logs, other supporting material
## :computer: Desktop:
***Please complete the following information:***
- Environment: 3.16.1
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
## :warning: Potential impacts
***Which other features may be impacted by this fix. This could be populated after fix***
***What are the impacted versions?***
## :link: Dependencies
see also https://github.com/gravitee-io/issues/issues/7329
Link a story or other related things...
|
1.0
|
[gateway] conditional logging on date/Duration prevents to display API logs - ## :collision: Describe the bug
When enabling logging on an API with a Condition on an date/Duration, the detailed logs are not displayed and the following warning is raise in the gateway logs :
```
gio_apim_gateway-3.16.1 | 10:17:52.956 [vert.x-eventloop-thread-4] [] WARN i.g.g.c.l.p.LoggableRequestProcessor - Unexpected error while evaluating logging condition for the API e410fbd5-725e-43b5-90fb-d5725ea3b5e0 and context path /echo/ : EL1041E: After parsing a valid expression, there is still more data in the expression: 'Mar'
```
Condition is displayed as :
{#request.timestamp <= Fri Mar 25 2022 14:00:15 `GMT+0100l}
Workaround is to convert to a timestamp in Milliseconds as :
{#request.timestamp <= 1647944975000l}
## :sunrise_over_mountains: To Reproduce
Steps to reproduce the behaviour:
1. Create an API with API Key Plan
2. subscribe to it with your Default Application
3. Enable Logging on the API and add a Condition on the Date/Duration
4. Make a request on the API and check the logs
## :rainbow: Expected behaviour
A clear and concise description of what you expected to happen.
## Current behaviour
A clear and concise description of what is currently happening.
## :movie_camera: Useful information
Screenshot, video, logs, other supporting material
## :computer: Desktop:
***Please complete the following information:***
- Environment: 3.16.1
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
## :warning: Potential impacts
***Which other features may be impacted by this fix. This could be populated after fix***
***What are the impacted versions?***
## :link: Dependencies
see also https://github.com/gravitee-io/issues/issues/7329
Link a story or other related things...
|
non_design
|
conditional logging on date duration prevents to display api logs collision describe the bug when enabling logging on an api with a condition on an date duration the detailed logs are not displayed and the following warning is raise in the gateway logs gio apim gateway warn i g g c l p loggablerequestprocessor unexpected error while evaluating logging condition for the api and context path echo after parsing a valid expression there is still more data in the expression mar condition is displayed as request timestamp fri mar gmt workaround is to convert to a timestamp in milliseconds as request timestamp sunrise over mountains to reproduce steps to reproduce the behaviour create an api with api key plan subscribe to it with your default application enable logging on the api and add a condition on the date duration make a request on the api and check the logs rainbow expected behaviour a clear and concise description of what you expected to happen current behaviour a clear and concise description of what is currently happening movie camera useful information screenshot video logs other supporting material computer desktop please complete the following information environment os browser warning potential impacts which other features may be impacted by this fix this could be populated after fix what are the impacted versions link dependencies see also link a story or other related things
| 0
|
100,558
| 4,098,272,332
|
IssuesEvent
|
2016-06-03 07:34:43
|
agda/agda
|
https://api.github.com/repos/agda/agda
|
closed
|
Copattern matching with catch-all is inconsistent
|
copatterns pattern-matching priority-high
|
Consider the following code:
```agda
data Bool : Set where
true false : Bool
f : Bool → Set₁
f true = Set
f = λ _ → Set
```
Agda fails to see that the pattern matching in `f` is complete:
```
Incomplete pattern matching for f. Missing cases:
f false
when checking the definition of f
```
Is this intended?
|
1.0
|
Copattern matching with catch-all is inconsistent - Consider the following code:
```agda
data Bool : Set where
true false : Bool
f : Bool → Set₁
f true = Set
f = λ _ → Set
```
Agda fails to see that the pattern matching in `f` is complete:
```
Incomplete pattern matching for f. Missing cases:
f false
when checking the definition of f
```
Is this intended?
|
non_design
|
copattern matching with catch all is inconsistent consider the following code agda data bool set where true false bool f bool → set₁ f true set f λ → set agda fails to see that the pattern matching in f is complete incomplete pattern matching for f missing cases f false when checking the definition of f is this intended
| 0
|
63,070
| 17,366,059,390
|
IssuesEvent
|
2021-07-30 07:27:30
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Riot is not respecting audible and call settings
|
A-Notifications A-VoIP P1 S-Major T-Defect
|
Hi
The Linux version of Riot desktop is not respecting the audible notifications and the p2p call settings.
I disabled audible notifications and p2p calls and I still get loud notifications with messages and calls, and I still get p2p calls. It seems to me that the settings are not respecting for such things.
I do not have any "noisy" notifications enabled either. I also tried enabling and disabling back with no luck
Debian 4.17.17-1 (2018-08-18) x86_64 GNU/Linux
matrix-react-sdk version: <local>
riot-web version: 0.16.4
olm version: 2.2.1
|
1.0
|
Riot is not respecting audible and call settings - Hi
The Linux version of Riot desktop is not respecting the audible notifications and the p2p call settings.
I disabled audible notifications and p2p calls and I still get loud notifications with messages and calls, and I still get p2p calls. It seems to me that the settings are not respecting for such things.
I do not have any "noisy" notifications enabled either. I also tried enabling and disabling back with no luck
Debian 4.17.17-1 (2018-08-18) x86_64 GNU/Linux
matrix-react-sdk version: <local>
riot-web version: 0.16.4
olm version: 2.2.1
|
non_design
|
riot is not respecting audible and call settings hi the linux version of riot desktop is not respecting the audible notifications and the call settings i disabled audible notifications and calls and i still get loud notifications with messages and calls and i still get calls it seems to me that the settings are not respecting for such things i do not have any noisy notifications enabled either i also tried enabling and disabling back with no luck debian gnu linux matrix react sdk version riot web version olm version
| 0
|
109,971
| 16,946,278,494
|
IssuesEvent
|
2021-06-28 07:16:53
|
k8-proxy/go-k8s-infra
|
https://api.github.com/repos/k8-proxy/go-k8s-infra
|
reopened
|
Securely handle Secrets across multiple clusters using industry standard practices
|
Epic P1 Security
|
As an InfoSec manageer or IT Administrator I expect Glasswall to implment recognised secrets management patterns for the handling and propogation of secrets in a open-architecture.
<br/>
Clusters should be able to validate secrets from a central location so that administrators are not expected to manually change values that need to be shared with N number of dependent entities.
|
True
|
Securely handle Secrets across multiple clusters using industry standard practices - As an InfoSec manageer or IT Administrator I expect Glasswall to implment recognised secrets management patterns for the handling and propogation of secrets in a open-architecture.
<br/>
Clusters should be able to validate secrets from a central location so that administrators are not expected to manually change values that need to be shared with N number of dependent entities.
|
non_design
|
securely handle secrets across multiple clusters using industry standard practices as an infosec manageer or it administrator i expect glasswall to implment recognised secrets management patterns for the handling and propogation of secrets in a open architecture clusters should be able to validate secrets from a central location so that administrators are not expected to manually change values that need to be shared with n number of dependent entities
| 0
|
660,761
| 21,997,105,264
|
IssuesEvent
|
2022-05-26 07:42:22
|
googleapis/python-appengine-admin
|
https://api.github.com/repos/googleapis/python-appengine-admin
|
closed
|
tests.unit.gapic.appengine_admin_v1.test_authorized_domains: test_list_authorized_domains_async_pager failed
|
type: bug priority: p1 flakybot: issue api: appengine
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 02284e0a99978cbfd3608d0204e210ee9944475f
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2fc8200f-ed4f-4d65-9a32-395fda4dd821), [Sponge](http://sponge2/2fc8200f-ed4f-4d65-9a32-395fda4dd821)
status: failed
<details><summary>Test output</summary><br><pre>@pytest.mark.asyncio
async def test_list_authorized_domains_async_pager():
client = AuthorizedDomainsAsyncClient(
credentials=ga_credentials.AnonymousCredentials,
)
# Mock the actual call within the gRPC stub, and fake the request.
with mock.patch.object(
type(client.transport.list_authorized_domains),
"__call__",
new_callable=mock.AsyncMock,
) as call:
# Set the response to a series of pages.
call.side_effect = (
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
],
next_page_token="abc",
),
appengine.ListAuthorizedDomainsResponse(
domains=[],
next_page_token="def",
),
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
],
next_page_token="ghi",
),
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
],
),
RuntimeError,
)
async_pager = await client.list_authorized_domains(
request={},
)
assert async_pager.next_page_token == "abc"
responses = []
async for response in async_pager: # pragma: no branch
responses.append(response)
assert len(responses) == 6
> assert all(isinstance(i, domain.AuthorizedDomain) for i in responses)
E assert False
E + where False = all(<generator object test_list_authorized_domains_async_pager.<locals>.<genexpr> at 0x7fcbebec0350>)
tests/unit/gapic/appengine_admin_v1/test_authorized_domains.py:976: AssertionError</pre></details>
|
1.0
|
tests.unit.gapic.appengine_admin_v1.test_authorized_domains: test_list_authorized_domains_async_pager failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 02284e0a99978cbfd3608d0204e210ee9944475f
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2fc8200f-ed4f-4d65-9a32-395fda4dd821), [Sponge](http://sponge2/2fc8200f-ed4f-4d65-9a32-395fda4dd821)
status: failed
<details><summary>Test output</summary><br><pre>@pytest.mark.asyncio
async def test_list_authorized_domains_async_pager():
client = AuthorizedDomainsAsyncClient(
credentials=ga_credentials.AnonymousCredentials,
)
# Mock the actual call within the gRPC stub, and fake the request.
with mock.patch.object(
type(client.transport.list_authorized_domains),
"__call__",
new_callable=mock.AsyncMock,
) as call:
# Set the response to a series of pages.
call.side_effect = (
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
],
next_page_token="abc",
),
appengine.ListAuthorizedDomainsResponse(
domains=[],
next_page_token="def",
),
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
],
next_page_token="ghi",
),
appengine.ListAuthorizedDomainsResponse(
domains=[
domain.AuthorizedDomain(),
domain.AuthorizedDomain(),
],
),
RuntimeError,
)
async_pager = await client.list_authorized_domains(
request={},
)
assert async_pager.next_page_token == "abc"
responses = []
async for response in async_pager: # pragma: no branch
responses.append(response)
assert len(responses) == 6
> assert all(isinstance(i, domain.AuthorizedDomain) for i in responses)
E assert False
E + where False = all(<generator object test_list_authorized_domains_async_pager.<locals>.<genexpr> at 0x7fcbebec0350>)
tests/unit/gapic/appengine_admin_v1/test_authorized_domains.py:976: AssertionError</pre></details>
|
non_design
|
tests unit gapic appengine admin test authorized domains test list authorized domains async pager failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output pytest mark asyncio async def test list authorized domains async pager client authorizeddomainsasyncclient credentials ga credentials anonymouscredentials mock the actual call within the grpc stub and fake the request with mock patch object type client transport list authorized domains call new callable mock asyncmock as call set the response to a series of pages call side effect appengine listauthorizeddomainsresponse domains domain authorizeddomain domain authorizeddomain domain authorizeddomain next page token abc appengine listauthorizeddomainsresponse domains next page token def appengine listauthorizeddomainsresponse domains domain authorizeddomain next page token ghi appengine listauthorizeddomainsresponse domains domain authorizeddomain domain authorizeddomain runtimeerror async pager await client list authorized domains request assert async pager next page token abc responses async for response in async pager pragma no branch responses append response assert len responses assert all isinstance i domain authorizeddomain for i in responses e assert false e where false all at tests unit gapic appengine admin test authorized domains py assertionerror
| 0
|
108,134
| 23,538,177,291
|
IssuesEvent
|
2022-08-20 01:26:46
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
opened
|
With 1.12.1, infinite memory gets used after doing a readability-else-after-return clang-tidy fix with vcFormat on Windows with a file with LF (\n) line endings
|
bug Language Service regression Feature: Code Formatting Feature: Code Analysis
|
Use a file with LF line endings on Windows with...
```cpp
void CheckModified(int ii) {
if (ii == 0) {
return;
}
else {
//aaa
}
}
```
```json
"C_Cpp.codeAnalysis.clangTidy.checks.enabled": [
"readability-else-after-return"
],
"C_Cpp.formatting": "vcFormat",
```
Run Code Analysis and do the Fix option.
Bug: Infinite memory gets used during some vcFormat operation -- you should close Vs Code when this happens or Windows may become unstable due to the lack of memory.
|
2.0
|
With 1.12.1, infinite memory gets used after doing a readability-else-after-return clang-tidy fix with vcFormat on Windows with a file with LF (\n) line endings - Use a file with LF line endings on Windows with...
```cpp
void CheckModified(int ii) {
if (ii == 0) {
return;
}
else {
//aaa
}
}
```
```json
"C_Cpp.codeAnalysis.clangTidy.checks.enabled": [
"readability-else-after-return"
],
"C_Cpp.formatting": "vcFormat",
```
Run Code Analysis and do the Fix option.
Bug: Infinite memory gets used during some vcFormat operation -- you should close Vs Code when this happens or Windows may become unstable due to the lack of memory.
|
non_design
|
with infinite memory gets used after doing a readability else after return clang tidy fix with vcformat on windows with a file with lf n line endings use a file with lf line endings on windows with cpp void checkmodified int ii if ii return else aaa json c cpp codeanalysis clangtidy checks enabled readability else after return c cpp formatting vcformat run code analysis and do the fix option bug infinite memory gets used during some vcformat operation you should close vs code when this happens or windows may become unstable due to the lack of memory
| 0
|
408,921
| 11,954,462,485
|
IssuesEvent
|
2020-04-03 23:44:35
|
getting-things-gnome/gtg
|
https://api.github.com/repos/getting-things-gnome/gtg
|
closed
|
Task editor windows do not have the proper children window relationship to the main window
|
bug low-hanging-fruit priority:medium reproducible-in-git
|
The task editor instances need to:
* have the "attached-to" and/or "parent" and/or "transient-for" property set correctly to be attached to the main window
* possibly have a type-hint set to be a utility (or dialog) window?
* in both cases it would also avoid having minimize/maximize buttons added by Ubuntu onto that thing
* in the case of utility window types, it would make the children windows minimize alongside the main window if the main window is minimized, if I'm not mistaken, which I've been wanting for a long time.
|
1.0
|
Task editor windows do not have the proper children window relationship to the main window - The task editor instances need to:
* have the "attached-to" and/or "parent" and/or "transient-for" property set correctly to be attached to the main window
* possibly have a type-hint set to be a utility (or dialog) window?
* in both cases it would also avoid having minimize/maximize buttons added by Ubuntu onto that thing
* in the case of utility window types, it would make the children windows minimize alongside the main window if the main window is minimized, if I'm not mistaken, which I've been wanting for a long time.
|
non_design
|
task editor windows do not have the proper children window relationship to the main window the task editor instances need to have the attached to and or parent and or transient for property set correctly to be attached to the main window possibly have a type hint set to be a utility or dialog window in both cases it would also avoid having minimize maximize buttons added by ubuntu onto that thing in the case of utility window types it would make the children windows minimize alongside the main window if the main window is minimized if i m not mistaken which i ve been wanting for a long time
| 0
|
239,996
| 26,254,310,517
|
IssuesEvent
|
2023-01-05 22:31:59
|
mpulsemobile/doccano
|
https://api.github.com/repos/mpulsemobile/doccano
|
opened
|
CVE-2021-23386 (Medium) detected in dns-packet-1.3.1.tgz
|
security vulnerability
|
## CVE-2021-23386 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dns-packet-1.3.1.tgz</b></p></summary>
<p>An abstract-encoding compliant module for encoding / decoding DNS packets</p>
<p>Library home page: <a href="https://registry.npmjs.org/dns-packet/-/dns-packet-1.3.1.tgz">https://registry.npmjs.org/dns-packet/-/dns-packet-1.3.1.tgz</a></p>
<p>Path to dependency file: /app/server/static/package.json</p>
<p>Path to vulnerable library: /app/server/static/node_modules/dns-packet/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- bonjour-3.5.0.tgz
- multicast-dns-6.2.3.tgz
- :x: **dns-packet-1.3.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package dns-packet before 5.2.2. It creates buffers with allocUnsafe and does not always fill them before forming network packets. This can expose internal application memory over unencrypted network when querying crafted invalid domain names.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23386>CVE-2021-23386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23386">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23386</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution (dns-packet): 1.3.2</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 3.3.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
True
|
CVE-2021-23386 (Medium) detected in dns-packet-1.3.1.tgz - ## CVE-2021-23386 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dns-packet-1.3.1.tgz</b></p></summary>
<p>An abstract-encoding compliant module for encoding / decoding DNS packets</p>
<p>Library home page: <a href="https://registry.npmjs.org/dns-packet/-/dns-packet-1.3.1.tgz">https://registry.npmjs.org/dns-packet/-/dns-packet-1.3.1.tgz</a></p>
<p>Path to dependency file: /app/server/static/package.json</p>
<p>Path to vulnerable library: /app/server/static/node_modules/dns-packet/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- bonjour-3.5.0.tgz
- multicast-dns-6.2.3.tgz
- :x: **dns-packet-1.3.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package dns-packet before 5.2.2. It creates buffers with allocUnsafe and does not always fill them before forming network packets. This can expose internal application memory over unencrypted network when querying crafted invalid domain names.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23386>CVE-2021-23386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23386">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23386</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution (dns-packet): 1.3.2</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 3.3.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
non_design
|
cve medium detected in dns packet tgz cve medium severity vulnerability vulnerable library dns packet tgz an abstract encoding compliant module for encoding decoding dns packets library home page a href path to dependency file app server static package json path to vulnerable library app server static node modules dns packet package json dependency hierarchy webpack dev server tgz root library bonjour tgz multicast dns tgz x dns packet tgz vulnerable library vulnerability details this affects the package dns packet before it creates buffers with allocunsafe and does not always fill them before forming network packets this can expose internal application memory over unencrypted network when querying crafted invalid domain names publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution dns packet direct dependency fix resolution webpack dev server check this box to open an automated fix pr
| 0
|
590,240
| 17,774,509,479
|
IssuesEvent
|
2021-08-30 17:24:21
|
anegostudios/VintageStory-Issues
|
https://api.github.com/repos/anegostudios/VintageStory-Issues
|
opened
|
Request World Download from Server
|
status: new priority: high
|
**Game Version:** 1.15.5
**Platform:** Windows
**Modded:** Unknown
### Description
Im having trouble with downloading the world from my server.
My game says "World Download Requested, copying in progress.
To continue, the game keepings crashing when trying to load "Your Game Server" here is the log
### How to reproduce
### Expected behavior
### Screenshots
### Logs
```
Running on 64 bit Windows with 32 GB RAM
Version: v1.15.5 (Stable)8/29/2021 12:12:38 AM: Critical error occurred
System.NullReferenceException: Object reference not set to an instance of an object.
at _7Gljh9TrtMZNo1VaLv2WsBRLGqD._IzerFHKy7VGQuDWNxLUYvJgrKKb()
at _Jm1GekeVHONpXWfUYAiJ0gvyjE._lxtmYa4SzyWlslBnrHadOIfFZpi`1._cwNjMeV6CKknPcGV2ZAAMQ2rLzf()
at _Wwq7msw90Af4XU7mxkxAUKyrS5B._vCB99EztL7cSfmplK8tyTwrQQxC(Single )
at Vintagestory.Client.NoObf.ClientPlatformWindows.window_RenderFrame(Object sender, FrameEventArgs e)
at System.EventHandler`1.Invoke(Object sender, TEventArgs e)
at OpenTK.GameWindow.RaiseRenderFrame(Double elapsed, Double& timestamp)
at OpenTK.GameWindow.DispatchRenderFrame()
at OpenTK.GameWindow.Run(Double updates_per_second, Double frames_per_second)
at _q065sSmA3CNgGuKTGnlAjR4ljPw._WkSl7yfiyr5DwPiSMvTudNAhp6(_t3ILnyYBq5OoYhhBY3JeTWgFmJDA , String[] )
at _DBu5wiQWRAHGS5G2jiCKymvQ9yE._WkSl7yfiyr5DwPiSMvTudNAhp6(ThreadStart )
-------------------------------
Event Log entries containing Vintagestory.exe, the latest 3
==================================
{ TimeGenerated = 8/26/2021 7:26:47 PM, Site = , Source = Windows Error Reporting, Message = Fault bucket 1658733256924091269, type 5
Event Name: RADAR_PRE_LEAK_64
Response: Not available
Cab Id: 0
Problem signature:
P1: Vintagestory.exe
P2: 1.15.5.0
P3: 10.0.19042.2.0.0
P4:
P5:
P6:
P7:
P8:
P9:
P10:
Attached files:
\\?\C:\Users\JOSHUA~1\AppData\Local\Temp\RDR74F3.tmp\empty.txt
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER74F4.tmp.WERInternalMetadata.xml
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER7505.tmp.xml
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER750F.tmp.csv
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER752F.tmp.txt
These files may be available here:
Analysis symbol:
Rechecking for solution: 0
Report Id: 382c07ef-0e6c-4ec0-8712-009f110df371
Report Status: 268435456
Hashed bucket: eb8d9853264150393705011bdc165385
Cab Guid: 0 }
```
|
1.0
|
Request World Download from Server - **Game Version:** 1.15.5
**Platform:** Windows
**Modded:** Unknown
### Description
Im having trouble with downloading the world from my server.
My game says "World Download Requested, copying in progress.
To continue, the game keepings crashing when trying to load "Your Game Server" here is the log
### How to reproduce
### Expected behavior
### Screenshots
### Logs
```
Running on 64 bit Windows with 32 GB RAM
Version: v1.15.5 (Stable)8/29/2021 12:12:38 AM: Critical error occurred
System.NullReferenceException: Object reference not set to an instance of an object.
at _7Gljh9TrtMZNo1VaLv2WsBRLGqD._IzerFHKy7VGQuDWNxLUYvJgrKKb()
at _Jm1GekeVHONpXWfUYAiJ0gvyjE._lxtmYa4SzyWlslBnrHadOIfFZpi`1._cwNjMeV6CKknPcGV2ZAAMQ2rLzf()
at _Wwq7msw90Af4XU7mxkxAUKyrS5B._vCB99EztL7cSfmplK8tyTwrQQxC(Single )
at Vintagestory.Client.NoObf.ClientPlatformWindows.window_RenderFrame(Object sender, FrameEventArgs e)
at System.EventHandler`1.Invoke(Object sender, TEventArgs e)
at OpenTK.GameWindow.RaiseRenderFrame(Double elapsed, Double& timestamp)
at OpenTK.GameWindow.DispatchRenderFrame()
at OpenTK.GameWindow.Run(Double updates_per_second, Double frames_per_second)
at _q065sSmA3CNgGuKTGnlAjR4ljPw._WkSl7yfiyr5DwPiSMvTudNAhp6(_t3ILnyYBq5OoYhhBY3JeTWgFmJDA , String[] )
at _DBu5wiQWRAHGS5G2jiCKymvQ9yE._WkSl7yfiyr5DwPiSMvTudNAhp6(ThreadStart )
-------------------------------
Event Log entries containing Vintagestory.exe, the latest 3
==================================
{ TimeGenerated = 8/26/2021 7:26:47 PM, Site = , Source = Windows Error Reporting, Message = Fault bucket 1658733256924091269, type 5
Event Name: RADAR_PRE_LEAK_64
Response: Not available
Cab Id: 0
Problem signature:
P1: Vintagestory.exe
P2: 1.15.5.0
P3: 10.0.19042.2.0.0
P4:
P5:
P6:
P7:
P8:
P9:
P10:
Attached files:
\\?\C:\Users\JOSHUA~1\AppData\Local\Temp\RDR74F3.tmp\empty.txt
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER74F4.tmp.WERInternalMetadata.xml
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER7505.tmp.xml
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER750F.tmp.csv
\\?\C:\ProgramData\Microsoft\Windows\WER\Temp\WER752F.tmp.txt
These files may be available here:
Analysis symbol:
Rechecking for solution: 0
Report Id: 382c07ef-0e6c-4ec0-8712-009f110df371
Report Status: 268435456
Hashed bucket: eb8d9853264150393705011bdc165385
Cab Guid: 0 }
```
|
non_design
|
request world download from server game version platform windows modded unknown description im having trouble with downloading the world from my server my game says world download requested copying in progress to continue the game keepings crashing when trying to load your game server here is the log how to reproduce expected behavior screenshots logs running on bit windows with gb ram version stable am critical error occurred system nullreferenceexception object reference not set to an instance of an object at at at single at vintagestory client noobf clientplatformwindows window renderframe object sender frameeventargs e at system eventhandler invoke object sender teventargs e at opentk gamewindow raiserenderframe double elapsed double timestamp at opentk gamewindow dispatchrenderframe at opentk gamewindow run double updates per second double frames per second at string at threadstart event log entries containing vintagestory exe the latest timegenerated pm site source windows error reporting message fault bucket type event name radar pre leak response not available cab id problem signature vintagestory exe attached files c users joshua appdata local temp tmp empty txt c programdata microsoft windows wer temp tmp werinternalmetadata xml c programdata microsoft windows wer temp tmp xml c programdata microsoft windows wer temp tmp csv c programdata microsoft windows wer temp tmp txt these files may be available here analysis symbol rechecking for solution report id report status hashed bucket cab guid
| 0
|
1,811
| 2,572,415,302
|
IssuesEvent
|
2015-02-10 22:23:08
|
kmcurry/3Scape
|
https://api.github.com/repos/kmcurry/3Scape
|
closed
|
Design Profile Page
|
Design In Progress UX
|
- [x] Sketch several concepts for profile pages
- [x] Review concepts and down-select top one or two
- [x] Render top choice
- [x] Review and iterate as needed
- [ ] Get internal feedback
- [ ] Get external feedback and UX review if practical
- [ ] Write CSS
|
1.0
|
Design Profile Page - - [x] Sketch several concepts for profile pages
- [x] Review concepts and down-select top one or two
- [x] Render top choice
- [x] Review and iterate as needed
- [ ] Get internal feedback
- [ ] Get external feedback and UX review if practical
- [ ] Write CSS
|
design
|
design profile page sketch several concepts for profile pages review concepts and down select top one or two render top choice review and iterate as needed get internal feedback get external feedback and ux review if practical write css
| 1
|
163,320
| 25,789,588,863
|
IssuesEvent
|
2022-12-10 01:25:07
|
MetaMask/metamask-extension
|
https://api.github.com/repos/MetaMask/metamask-extension
|
opened
|
Consolidate all component import paths
|
area-UI design-system IA/NAV
|
### Description
We currently have inconsistent import paths in our component documentation. We should be consistent with our import paths for components.

### Technical Details
All component examples should use short hand notation and be imported from ` '../../component-library'`
```
import { Button, FormTextField, AvatarAccount } from '../../component-library';
```
### Acceptance Critera
- All component code examples in MDX docs show the same consistent import path
|
1.0
|
Consolidate all component import paths - ### Description
We currently have inconsistent import paths in our component documentation. We should be consistent with our import paths for components.

### Technical Details
All component examples should use short hand notation and be imported from ` '../../component-library'`
```
import { Button, FormTextField, AvatarAccount } from '../../component-library';
```
### Acceptance Critera
- All component code examples in MDX docs show the same consistent import path
|
design
|
consolidate all component import paths description we currently have inconsistent import paths in our component documentation we should be consistent with our import paths for components technical details all component examples should use short hand notation and be imported from component library import button formtextfield avataraccount from component library acceptance critera all component code examples in mdx docs show the same consistent import path
| 1
|
80,662
| 10,037,615,604
|
IssuesEvent
|
2019-07-18 13:33:30
|
front-end-by-rimantas/9-grupe-portfolio-netlify
|
https://api.github.com/repos/front-end-by-rimantas/9-grupe-portfolio-netlify
|
closed
|
Create my services
|
css good first issue responsive design
|
- [x] lay services using JS
- [x] style services as needed
- [x] make it responsive
|
1.0
|
Create my services - - [x] lay services using JS
- [x] style services as needed
- [x] make it responsive
|
design
|
create my services lay services using js style services as needed make it responsive
| 1
|
465,288
| 13,381,958,252
|
IssuesEvent
|
2020-09-02 08:04:59
|
gardener/autoscaler
|
https://api.github.com/repos/gardener/autoscaler
|
closed
|
cluster-autoscaler panics when newly launched instance type is used
|
kind/bug priority/blocker
|
**What happened**:
Currently https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go is outdated compared to https://github.com/kubernetes/autoscaler/blob/master/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go.
cluster-autoscaler panics when such instance type that is not present in `azure_instance_types.go` is used with scale-from-zero feature.
**What you expected to happen**:
https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go to be updated with newly launched instance types.
**How to reproduce it (as minimally and precisely as possible)**:
1. Create a worker pool with instance type that is not maintained in https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go (for example `Standard_D2d_v4`) and `minimum=0`
```yaml
workers:
# omitted
- machine:
image:
name: coreos
version: 2512.3.0
type: Standard_D2d_v4
maxSurge: 1
maxUnavailable: 0
maximum: 1
minimum: 0
name: cpu-worker2
systemComponents:
allow: true
volume:
size: 35Gi
type: standard
```
2. Create a Pod that would trigger scale-up.
3. Ensure that cluster-autoscaler panics with:
```
$ k -n shoot--foo--bar logs cluster-autoscaler-78bc9675cb-scwvg
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x3051428]
goroutine 90 [running]:
github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm.(*McmManager).GetMachineDeploymentNodeTemplate(0xc420b70e10, 0xc420988bc0, 0x2, 0x0, 0xc4214caea0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm/mcm_manager.go:425 +0x508
github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm.(*MachineDeployment).TemplateNodeInfo(0xc420988bc0, 0xc4214caea0, 0x7ffeffe95344, 0x28)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm/mcm_cloud_provider.go:349 +0x38
github.com/gardener/autoscaler/cluster-autoscaler/core.GetNodeInfosForGroups(0xc421455460, 0x2, 0x2, 0x44582a0, 0xc421076c00, 0x447e3a0, 0xc4209f4a00, 0x0, 0x0, 0x0, ...)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/utils.go:269 +0x296
github.com/gardener/autoscaler/cluster-autoscaler/core.ScaleUp(0xc4201b9860, 0xc420d38f00, 0xc420a5cf00, 0xc420b5f0d0, 0x1, 0x1, 0xc421455460, 0x2, 0x2, 0x0, ...)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/scale_up.go:267 +0x420
github.com/gardener/autoscaler/cluster-autoscaler/core.(*StaticAutoscaler).RunOnce(0xc420b70ea0, 0xbfcab3e9d11cc4ae, 0x2590855617, 0x6b4cea0, 0x0, 0x0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/static_autoscaler.go:273 +0x1e0c
main.run(0xc4206a4000)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/main.go:294 +0x1dd
main.main.func2(0x443ae00, 0xc420dc9bc0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/main.go:373 +0x2a
created by github.com/gardener/autoscaler/cluster-autoscaler/vendor/k8s.io/client-go/tools/leaderelection.(*LeaderElector).Run
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/vendor/k8s.io/client-go/tools/leaderelection/leaderelection.go:155 +0xee
```
```
$ k -n shoot--foo--bar get po -l role=cluster-autoscaler
NAME READY STATUS RESTARTS AGE
cluster-autoscaler-78bc9675cb-scwvg 0/1 CrashLoopBackOff 6 8m52s
```
**Anything else we need to know**:
**Environment**:
|
1.0
|
cluster-autoscaler panics when newly launched instance type is used - **What happened**:
Currently https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go is outdated compared to https://github.com/kubernetes/autoscaler/blob/master/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go.
cluster-autoscaler panics when such instance type that is not present in `azure_instance_types.go` is used with scale-from-zero feature.
**What you expected to happen**:
https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go to be updated with newly launched instance types.
**How to reproduce it (as minimally and precisely as possible)**:
1. Create a worker pool with instance type that is not maintained in https://github.com/gardener/autoscaler/blob/machine-controller-manager/cluster-autoscaler/cloudprovider/azure/azure_instance_types.go (for example `Standard_D2d_v4`) and `minimum=0`
```yaml
workers:
# omitted
- machine:
image:
name: coreos
version: 2512.3.0
type: Standard_D2d_v4
maxSurge: 1
maxUnavailable: 0
maximum: 1
minimum: 0
name: cpu-worker2
systemComponents:
allow: true
volume:
size: 35Gi
type: standard
```
2. Create a Pod that would trigger scale-up.
3. Ensure that cluster-autoscaler panics with:
```
$ k -n shoot--foo--bar logs cluster-autoscaler-78bc9675cb-scwvg
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x3051428]
goroutine 90 [running]:
github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm.(*McmManager).GetMachineDeploymentNodeTemplate(0xc420b70e10, 0xc420988bc0, 0x2, 0x0, 0xc4214caea0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm/mcm_manager.go:425 +0x508
github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm.(*MachineDeployment).TemplateNodeInfo(0xc420988bc0, 0xc4214caea0, 0x7ffeffe95344, 0x28)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/cloudprovider/mcm/mcm_cloud_provider.go:349 +0x38
github.com/gardener/autoscaler/cluster-autoscaler/core.GetNodeInfosForGroups(0xc421455460, 0x2, 0x2, 0x44582a0, 0xc421076c00, 0x447e3a0, 0xc4209f4a00, 0x0, 0x0, 0x0, ...)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/utils.go:269 +0x296
github.com/gardener/autoscaler/cluster-autoscaler/core.ScaleUp(0xc4201b9860, 0xc420d38f00, 0xc420a5cf00, 0xc420b5f0d0, 0x1, 0x1, 0xc421455460, 0x2, 0x2, 0x0, ...)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/scale_up.go:267 +0x420
github.com/gardener/autoscaler/cluster-autoscaler/core.(*StaticAutoscaler).RunOnce(0xc420b70ea0, 0xbfcab3e9d11cc4ae, 0x2590855617, 0x6b4cea0, 0x0, 0x0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/core/static_autoscaler.go:273 +0x1e0c
main.run(0xc4206a4000)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/main.go:294 +0x1dd
main.main.func2(0x443ae00, 0xc420dc9bc0)
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/main.go:373 +0x2a
created by github.com/gardener/autoscaler/cluster-autoscaler/vendor/k8s.io/client-go/tools/leaderelection.(*LeaderElector).Run
/tmp/build/80754af9/git-gardener_autoscaler-machine-controller-manager_machine-controller-manager/tmp/src/github.com/gardener/autoscaler/cluster-autoscaler/vendor/k8s.io/client-go/tools/leaderelection/leaderelection.go:155 +0xee
```
```
$ k -n shoot--foo--bar get po -l role=cluster-autoscaler
NAME READY STATUS RESTARTS AGE
cluster-autoscaler-78bc9675cb-scwvg 0/1 CrashLoopBackOff 6 8m52s
```
**Anything else we need to know**:
**Environment**:
|
non_design
|
cluster autoscaler panics when newly launched instance type is used what happened currently is outdated compared to cluster autoscaler panics when such instance type that is not present in azure instance types go is used with scale from zero feature what you expected to happen to be updated with newly launched instance types how to reproduce it as minimally and precisely as possible create a worker pool with instance type that is not maintained in for example standard and minimum yaml workers omitted machine image name coreos version type standard maxsurge maxunavailable maximum minimum name cpu systemcomponents allow true volume size type standard create a pod that would trigger scale up ensure that cluster autoscaler panics with k n shoot foo bar logs cluster autoscaler scwvg panic runtime error invalid memory address or nil pointer dereference goroutine github com gardener autoscaler cluster autoscaler cloudprovider mcm mcmmanager getmachinedeploymentnodetemplate tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler cloudprovider mcm mcm manager go github com gardener autoscaler cluster autoscaler cloudprovider mcm machinedeployment templatenodeinfo tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler cloudprovider mcm mcm cloud provider go github com gardener autoscaler cluster autoscaler core getnodeinfosforgroups tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler core utils go github com gardener autoscaler cluster autoscaler core scaleup tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler core scale up go github com gardener autoscaler cluster autoscaler core staticautoscaler runonce tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler core static autoscaler go main run tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler main go main main tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler main go created by github com gardener autoscaler cluster autoscaler vendor io client go tools leaderelection leaderelector run tmp build git gardener autoscaler machine controller manager machine controller manager tmp src github com gardener autoscaler cluster autoscaler vendor io client go tools leaderelection leaderelection go k n shoot foo bar get po l role cluster autoscaler name ready status restarts age cluster autoscaler scwvg crashloopbackoff anything else we need to know environment
| 0
|
66,179
| 7,968,625,694
|
IssuesEvent
|
2018-07-16 04:41:06
|
quoeamaster/queutil
|
https://api.github.com/repos/quoeamaster/queutil
|
closed
|
add logging utility features
|
design feature released
|
**logging** is essential to all kinds of systems, a simple yet flexible logging architecture could solve things out in an efficient way.
Features involved:
* able to add logger(s) to the system; each logger would help to log down messages in a particular scenario (e.g. rolling-file logging, rdbms logging, console logging)
* provide lifecycle hooks / apis to add necessary resource claiming and resource release operations
|
1.0
|
add logging utility features - **logging** is essential to all kinds of systems, a simple yet flexible logging architecture could solve things out in an efficient way.
Features involved:
* able to add logger(s) to the system; each logger would help to log down messages in a particular scenario (e.g. rolling-file logging, rdbms logging, console logging)
* provide lifecycle hooks / apis to add necessary resource claiming and resource release operations
|
design
|
add logging utility features logging is essential to all kinds of systems a simple yet flexible logging architecture could solve things out in an efficient way features involved able to add logger s to the system each logger would help to log down messages in a particular scenario e g rolling file logging rdbms logging console logging provide lifecycle hooks apis to add necessary resource claiming and resource release operations
| 1
|
391,389
| 26,890,849,922
|
IssuesEvent
|
2023-02-06 08:48:46
|
TypeCobolTeam/TypeCobol
|
https://api.github.com/repos/TypeCobolTeam/TypeCobol
|
closed
|
Create better functional documentation for TypeCobol (FR + EN)
|
Documentation
|
We already have some docs on the Wiki + one private Tutorial.
We need to improve the documentation.
For our enterprise we need a doc in English and French.
|
1.0
|
Create better functional documentation for TypeCobol (FR + EN) - We already have some docs on the Wiki + one private Tutorial.
We need to improve the documentation.
For our enterprise we need a doc in English and French.
|
non_design
|
create better functional documentation for typecobol fr en we already have some docs on the wiki one private tutorial we need to improve the documentation for our enterprise we need a doc in english and french
| 0
|
6,578
| 7,693,392,766
|
IssuesEvent
|
2018-05-18 03:20:08
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
VM Extensions Example gives error on Network Interface
|
bug service/virtual-machine-extensions
|
### Terraform Version
Terraform v0.10.8
### Affected Resource(s)
N/A
### Terraform Configuration Files
```hcl
resource "random_id" "server" {
keepers = {
azi_id = 1
}
byte_length = 8
}
resource "azurerm_resource_group" "test" {
name = "acctestrg"
location = "West US 2"
}
resource "azurerm_virtual_network" "test" {
name = "acctvn"
address_space = ["10.0.0.0/16"]
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
}
resource "azurerm_subnet" "test" {
name = "acctsub"
resource_group_name = "${azurerm_resource_group.test.name}"
virtual_network_name = "${azurerm_virtual_network.test.name}"
address_prefix = "10.0.2.0/24"
}
resource "azurerm_network_interface" "test" {
name = "acctni"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
ip_configuration {
name = "testconfiguration1"
subnet_id = "${azurerm_subnet.test.id}"
private_ip_address_allocation = "dynamic"
}
}
resource "azurerm_storage_account" "test" {
name = "${random_id.server.hex}"
resource_group_name = "${azurerm_resource_group.test.name}"
location = "West US 2"
account_tier = "Standard"
account_replication_type = "LRS"
tags {
environment = "staging"
}
}
resource "azurerm_storage_container" "test" {
name = "vhds"
resource_group_name = "${azurerm_resource_group.test.name}"
storage_account_name = "${azurerm_storage_account.test.name}"
container_access_type = "private"
}
resource "azurerm_virtual_machine" "test" {
name = "acctvm"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
network_interface_ids = ["${azurerm_network_interface.test.id}"]
vm_size = "Standard_A0"
storage_image_reference {
publisher = "Canonical"
offer = "UbuntuServer"
sku = "16.04-LTS"
version = "latest"
}
storage_os_disk {
name = "myosdisk1"
vhd_uri = "${azurerm_storage_account.test.primary_blob_endpoint}${azurerm_storage_container.test.name}/myosdisk1.vhd"
caching = "ReadWrite"
create_option = "FromImage"
}
os_profile {
computer_name = "hostname"
admin_username = "testadmin"
admin_password = "Password1234!"
}
os_profile_linux_config {
disable_password_authentication = false
}
tags {
environment = "staging"
}
}
resource "azurerm_virtual_machine_extension" "test" {
name = "hostname"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
virtual_machine_name = "${azurerm_virtual_machine.test.name}"
publisher = "Microsoft.OSTCExtensions"
type = "CustomScriptForLinux"
type_handler_version = "1.2"
settings = <<SETTINGS
{
"commandToExecute": "hostname"
}
SETTINGS
tags {
environment = "Production"
}
}
```
### Debug Output
N/A
### Panic Output
N/A
### Expected Behavior
Resources Created
### Actual Behavior
Error message
Error: Error applying plan:
1 error(s) occurred:
* azurerm_network_interface.test: 1 error(s) occurred:
* azurerm_network_interface.test: network.InterfacesClient#CreateOrUpdate: Failure responding to request: StatusCode=400 -- Original Error: autorest/azure: Service returned an error. Status=400 Code="InvalidResourceReference" Message="Resource /subscriptions/959965bb-d9df-4c6f-825e-37e1090d817d/resourceGroups/acctestrg/providers/Microsoft.Network/virtualNetworks/acctvn/subnets/acctsub referenced by resource /subscriptions/959965bb-d9df-4c6f-825e-37e1090d817d/resourceGroups/acctestrg/providers/Microsoft.Network/networkInterfaces/acctni was not found. Please make sure that the referenced resource exists, and that both resources are in the same region." Details=[]
Terraform does not automatically rollback in the face of errors.
Instead, your Terraform state file has been partially updated with
any resources that successfully completed. Please address the error
above and apply again to incrementally change your infrastructure.
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform apply`
### Important Factoids
### References
|
1.0
|
VM Extensions Example gives error on Network Interface - ### Terraform Version
Terraform v0.10.8
### Affected Resource(s)
N/A
### Terraform Configuration Files
```hcl
resource "random_id" "server" {
keepers = {
azi_id = 1
}
byte_length = 8
}
resource "azurerm_resource_group" "test" {
name = "acctestrg"
location = "West US 2"
}
resource "azurerm_virtual_network" "test" {
name = "acctvn"
address_space = ["10.0.0.0/16"]
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
}
resource "azurerm_subnet" "test" {
name = "acctsub"
resource_group_name = "${azurerm_resource_group.test.name}"
virtual_network_name = "${azurerm_virtual_network.test.name}"
address_prefix = "10.0.2.0/24"
}
resource "azurerm_network_interface" "test" {
name = "acctni"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
ip_configuration {
name = "testconfiguration1"
subnet_id = "${azurerm_subnet.test.id}"
private_ip_address_allocation = "dynamic"
}
}
resource "azurerm_storage_account" "test" {
name = "${random_id.server.hex}"
resource_group_name = "${azurerm_resource_group.test.name}"
location = "West US 2"
account_tier = "Standard"
account_replication_type = "LRS"
tags {
environment = "staging"
}
}
resource "azurerm_storage_container" "test" {
name = "vhds"
resource_group_name = "${azurerm_resource_group.test.name}"
storage_account_name = "${azurerm_storage_account.test.name}"
container_access_type = "private"
}
resource "azurerm_virtual_machine" "test" {
name = "acctvm"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
network_interface_ids = ["${azurerm_network_interface.test.id}"]
vm_size = "Standard_A0"
storage_image_reference {
publisher = "Canonical"
offer = "UbuntuServer"
sku = "16.04-LTS"
version = "latest"
}
storage_os_disk {
name = "myosdisk1"
vhd_uri = "${azurerm_storage_account.test.primary_blob_endpoint}${azurerm_storage_container.test.name}/myosdisk1.vhd"
caching = "ReadWrite"
create_option = "FromImage"
}
os_profile {
computer_name = "hostname"
admin_username = "testadmin"
admin_password = "Password1234!"
}
os_profile_linux_config {
disable_password_authentication = false
}
tags {
environment = "staging"
}
}
resource "azurerm_virtual_machine_extension" "test" {
name = "hostname"
location = "West US 2"
resource_group_name = "${azurerm_resource_group.test.name}"
virtual_machine_name = "${azurerm_virtual_machine.test.name}"
publisher = "Microsoft.OSTCExtensions"
type = "CustomScriptForLinux"
type_handler_version = "1.2"
settings = <<SETTINGS
{
"commandToExecute": "hostname"
}
SETTINGS
tags {
environment = "Production"
}
}
```
### Debug Output
N/A
### Panic Output
N/A
### Expected Behavior
Resources Created
### Actual Behavior
Error message
Error: Error applying plan:
1 error(s) occurred:
* azurerm_network_interface.test: 1 error(s) occurred:
* azurerm_network_interface.test: network.InterfacesClient#CreateOrUpdate: Failure responding to request: StatusCode=400 -- Original Error: autorest/azure: Service returned an error. Status=400 Code="InvalidResourceReference" Message="Resource /subscriptions/959965bb-d9df-4c6f-825e-37e1090d817d/resourceGroups/acctestrg/providers/Microsoft.Network/virtualNetworks/acctvn/subnets/acctsub referenced by resource /subscriptions/959965bb-d9df-4c6f-825e-37e1090d817d/resourceGroups/acctestrg/providers/Microsoft.Network/networkInterfaces/acctni was not found. Please make sure that the referenced resource exists, and that both resources are in the same region." Details=[]
Terraform does not automatically rollback in the face of errors.
Instead, your Terraform state file has been partially updated with
any resources that successfully completed. Please address the error
above and apply again to incrementally change your infrastructure.
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform apply`
### Important Factoids
### References
|
non_design
|
vm extensions example gives error on network interface terraform version terraform affected resource s n a terraform configuration files hcl resource random id server keepers azi id byte length resource azurerm resource group test name acctestrg location west us resource azurerm virtual network test name acctvn address space location west us resource group name azurerm resource group test name resource azurerm subnet test name acctsub resource group name azurerm resource group test name virtual network name azurerm virtual network test name address prefix resource azurerm network interface test name acctni location west us resource group name azurerm resource group test name ip configuration name subnet id azurerm subnet test id private ip address allocation dynamic resource azurerm storage account test name random id server hex resource group name azurerm resource group test name location west us account tier standard account replication type lrs tags environment staging resource azurerm storage container test name vhds resource group name azurerm resource group test name storage account name azurerm storage account test name container access type private resource azurerm virtual machine test name acctvm location west us resource group name azurerm resource group test name network interface ids vm size standard storage image reference publisher canonical offer ubuntuserver sku lts version latest storage os disk name vhd uri azurerm storage account test primary blob endpoint azurerm storage container test name vhd caching readwrite create option fromimage os profile computer name hostname admin username testadmin admin password os profile linux config disable password authentication false tags environment staging resource azurerm virtual machine extension test name hostname location west us resource group name azurerm resource group test name virtual machine name azurerm virtual machine test name publisher microsoft ostcextensions type customscriptforlinux type handler version settings settings commandtoexecute hostname settings tags environment production debug output n a panic output n a expected behavior resources created actual behavior error message error error applying plan error s occurred azurerm network interface test error s occurred azurerm network interface test network interfacesclient createorupdate failure responding to request statuscode original error autorest azure service returned an error status code invalidresourcereference message resource subscriptions resourcegroups acctestrg providers microsoft network virtualnetworks acctvn subnets acctsub referenced by resource subscriptions resourcegroups acctestrg providers microsoft network networkinterfaces acctni was not found please make sure that the referenced resource exists and that both resources are in the same region details terraform does not automatically rollback in the face of errors instead your terraform state file has been partially updated with any resources that successfully completed please address the error above and apply again to incrementally change your infrastructure steps to reproduce please list the steps required to reproduce the issue for example terraform apply important factoids references
| 0
|
94,002
| 11,841,120,550
|
IssuesEvent
|
2020-03-23 20:10:30
|
patternfly/patternfly-design
|
https://api.github.com/repos/patternfly/patternfly-design
|
closed
|
Accordion box-shadow
|
Enhancement Visual Design
|
From https://github.com/patternfly/patternfly-next/issues/2375, we added a variation to the accordion component that removes the CSS `box-shadow`. After chatting with @mceledonia about it, we came to the conclusion that the shadow is not necessary and could be removed.
In this issue, I would like to determine if we should:
1. Leave the component as it is with the `box-shadow` enabled, and offer a variation to remove it.
2. Remove the existing `box-shadow` (since this is the most likely use case), and offer a variation to add it back as it is now.
3. Remove the `box-shadow`, and if a user would like a shadow around the accordion, they can place it inside a card component.
4. Something else?
Re: patternfly/patternfly-next#3, it's worth noting that the existing `box-shadow` on the accordion is `--pf-global--BoxShadow--md`, and the card's `box-shadow` is `--pf-global--BoxShadow--sm`.
cc @mcarrano
|
1.0
|
Accordion box-shadow - From https://github.com/patternfly/patternfly-next/issues/2375, we added a variation to the accordion component that removes the CSS `box-shadow`. After chatting with @mceledonia about it, we came to the conclusion that the shadow is not necessary and could be removed.
In this issue, I would like to determine if we should:
1. Leave the component as it is with the `box-shadow` enabled, and offer a variation to remove it.
2. Remove the existing `box-shadow` (since this is the most likely use case), and offer a variation to add it back as it is now.
3. Remove the `box-shadow`, and if a user would like a shadow around the accordion, they can place it inside a card component.
4. Something else?
Re: patternfly/patternfly-next#3, it's worth noting that the existing `box-shadow` on the accordion is `--pf-global--BoxShadow--md`, and the card's `box-shadow` is `--pf-global--BoxShadow--sm`.
cc @mcarrano
|
design
|
accordion box shadow from we added a variation to the accordion component that removes the css box shadow after chatting with mceledonia about it we came to the conclusion that the shadow is not necessary and could be removed in this issue i would like to determine if we should leave the component as it is with the box shadow enabled and offer a variation to remove it remove the existing box shadow since this is the most likely use case and offer a variation to add it back as it is now remove the box shadow and if a user would like a shadow around the accordion they can place it inside a card component something else re patternfly patternfly next it s worth noting that the existing box shadow on the accordion is pf global boxshadow md and the card s box shadow is pf global boxshadow sm cc mcarrano
| 1
|
1,172
| 2,532,695,066
|
IssuesEvent
|
2015-01-23 17:49:01
|
ThibaultLatrille/ControverSciences
|
https://api.github.com/repos/ThibaultLatrille/ControverSciences
|
closed
|
Icone
|
**** urgent design
|
Mettre icone pour les pages
Sur la page http://www.controversciences.org/
Par : T. Latrille
Navigateur : chrome modern linux webkit
|
1.0
|
Icone - Mettre icone pour les pages
Sur la page http://www.controversciences.org/
Par : T. Latrille
Navigateur : chrome modern linux webkit
|
design
|
icone mettre icone pour les pages sur la page par t latrille navigateur chrome modern linux webkit
| 1
|
218,066
| 16,748,612,618
|
IssuesEvent
|
2021-06-11 19:05:02
|
ESV-20/CAHSI-Bayamon
|
https://api.github.com/repos/ESV-20/CAHSI-Bayamon
|
opened
|
Deliver Database E-R Model Diagram
|
documentation
|
_Design Entity Relation Model Diagram to fully understand the purpose for a database to the project._
|
1.0
|
Deliver Database E-R Model Diagram - _Design Entity Relation Model Diagram to fully understand the purpose for a database to the project._
|
non_design
|
deliver database e r model diagram design entity relation model diagram to fully understand the purpose for a database to the project
| 0
|
20,869
| 2,631,873,583
|
IssuesEvent
|
2015-03-07 15:04:35
|
GrannyCookies/scratchext2
|
https://api.github.com/repos/GrannyCookies/scratchext2
|
opened
|
2.0 Online Editor Glitch
|
bug low priority
|
Just found a glitch in the 2.0 Editor. I am running both the Chrome Extension and the Tampermonkey Userscript.

Oh goodness it annoys me!
Can someone fix?
|
1.0
|
2.0 Online Editor Glitch - Just found a glitch in the 2.0 Editor. I am running both the Chrome Extension and the Tampermonkey Userscript.

Oh goodness it annoys me!
Can someone fix?
|
non_design
|
online editor glitch just found a glitch in the editor i am running both the chrome extension and the tampermonkey userscript oh goodness it annoys me can someone fix
| 0
|
125,854
| 16,845,458,251
|
IssuesEvent
|
2021-06-19 11:35:34
|
lukihd/scanReader
|
https://api.github.com/repos/lukihd/scanReader
|
opened
|
Sketch of the application
|
design
|
# Sketch design of all componnents
- [ ] Home
- [ ] Manga
- [ ] Reader
- [ ] Submit
- [ ] Settings
|
1.0
|
Sketch of the application - # Sketch design of all componnents
- [ ] Home
- [ ] Manga
- [ ] Reader
- [ ] Submit
- [ ] Settings
|
design
|
sketch of the application sketch design of all componnents home manga reader submit settings
| 1
|
157,206
| 24,632,869,982
|
IssuesEvent
|
2022-10-17 04:48:14
|
zachyuen/fa22-cse110-lab3
|
https://api.github.com/repos/zachyuen/fa22-cse110-lab3
|
closed
|
Create meeting minutes template
|
design/template
|
**Lab #:**3
**Describe problem**
I need to create a meeting minutes template for the lab.
|
1.0
|
Create meeting minutes template - **Lab #:**3
**Describe problem**
I need to create a meeting minutes template for the lab.
|
design
|
create meeting minutes template lab describe problem i need to create a meeting minutes template for the lab
| 1
|
9,202
| 4,442,141,227
|
IssuesEvent
|
2016-08-19 12:20:20
|
aria2/aria2
|
https://api.github.com/repos/aria2/aria2
|
closed
|
Link failure with ld.gold, -Wl,--as-needed and --enable-libaria2
|
bug build
|
When aria2 is built with `--enable-libaria2` most of the relevant code is built into the shared library. However, `src/Makefile.am` puts almost all linked dependency libraries into `LDADD` which applies to executables only. Therefore, most of the NEEDED entries land in the executable rather than the library.
This has two consequences:
1. The library lacks proper dependencies and therefore can't be used by a program unless it links to all its dependency libraries directly.
2. If aria2c is linked using ld.gold (the new binutils linker) with `-Wl,--as-needed`, the linker detects that aria2c executable does not use most of the passed `-l` libraries and removes them. As a result, aria2c is broken since neither the program nor the library links to necessary dependency libraries.
I think the correct solution here is to move all the dependency libraries from `LDADD` to `libaria2_la_LIBADD`.
|
1.0
|
Link failure with ld.gold, -Wl,--as-needed and --enable-libaria2 - When aria2 is built with `--enable-libaria2` most of the relevant code is built into the shared library. However, `src/Makefile.am` puts almost all linked dependency libraries into `LDADD` which applies to executables only. Therefore, most of the NEEDED entries land in the executable rather than the library.
This has two consequences:
1. The library lacks proper dependencies and therefore can't be used by a program unless it links to all its dependency libraries directly.
2. If aria2c is linked using ld.gold (the new binutils linker) with `-Wl,--as-needed`, the linker detects that aria2c executable does not use most of the passed `-l` libraries and removes them. As a result, aria2c is broken since neither the program nor the library links to necessary dependency libraries.
I think the correct solution here is to move all the dependency libraries from `LDADD` to `libaria2_la_LIBADD`.
|
non_design
|
link failure with ld gold wl as needed and enable when is built with enable most of the relevant code is built into the shared library however src makefile am puts almost all linked dependency libraries into ldadd which applies to executables only therefore most of the needed entries land in the executable rather than the library this has two consequences the library lacks proper dependencies and therefore can t be used by a program unless it links to all its dependency libraries directly if is linked using ld gold the new binutils linker with wl as needed the linker detects that executable does not use most of the passed l libraries and removes them as a result is broken since neither the program nor the library links to necessary dependency libraries i think the correct solution here is to move all the dependency libraries from ldadd to la libadd
| 0
|
171,730
| 27,169,549,150
|
IssuesEvent
|
2023-02-17 18:05:22
|
briangormanly/agora
|
https://api.github.com/repos/briangormanly/agora
|
closed
|
Investigate / Integrate Google Identity services for web
|
help wanted good first issue Software Archecture Planning / Design Infrastructure SSO
|
Use googles new API (current one is scheduled for deprecation in 23)
https://developers.google.com/identity/gsi/web/guides/overview
|
1.0
|
Investigate / Integrate Google Identity services for web - Use googles new API (current one is scheduled for deprecation in 23)
https://developers.google.com/identity/gsi/web/guides/overview
|
design
|
investigate integrate google identity services for web use googles new api current one is scheduled for deprecation in
| 1
|
153,635
| 24,166,328,977
|
IssuesEvent
|
2022-09-22 15:18:23
|
hypha-dao/dho-web-client
|
https://api.github.com/repos/hypha-dao/dho-web-client
|
closed
|
Remove decimals from numbers
|
Design
|
Summary:
As a DAO user I don't want to see decimals in the DAO because the numbers are too long and confusing:
AC:
Remove decimals behind the dot on following numbers:
- [x] Proposal detail page (token for Utility, Cash & Voice)
- [x] Proposal creation wizard (token for Utility, Cash & Voice)
- [x] Wallet widget on profile
- [x] Wallet widget on wallet page
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/75991832/190535089-1c8dca96-5056-481d-9143-7bec1bba9869.png">
<img width="579" alt="image" src="https://user-images.githubusercontent.com/75991832/190535191-8f671de4-2cec-4f96-b6ba-276fac222e7b.png">
<img width="1163" alt="image" src="https://user-images.githubusercontent.com/75991832/190535432-a3873043-ed85-4259-89ee-98a76666941d.png">
Leave it with decimals on the wallet page table:
<img width="907" alt="image" src="https://user-images.githubusercontent.com/75991832/190535235-bac51a5f-a3e4-489c-a1dd-14f2d8578e8b.png">
|
1.0
|
Remove decimals from numbers - Summary:
As a DAO user I don't want to see decimals in the DAO because the numbers are too long and confusing:
AC:
Remove decimals behind the dot on following numbers:
- [x] Proposal detail page (token for Utility, Cash & Voice)
- [x] Proposal creation wizard (token for Utility, Cash & Voice)
- [x] Wallet widget on profile
- [x] Wallet widget on wallet page
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/75991832/190535089-1c8dca96-5056-481d-9143-7bec1bba9869.png">
<img width="579" alt="image" src="https://user-images.githubusercontent.com/75991832/190535191-8f671de4-2cec-4f96-b6ba-276fac222e7b.png">
<img width="1163" alt="image" src="https://user-images.githubusercontent.com/75991832/190535432-a3873043-ed85-4259-89ee-98a76666941d.png">
Leave it with decimals on the wallet page table:
<img width="907" alt="image" src="https://user-images.githubusercontent.com/75991832/190535235-bac51a5f-a3e4-489c-a1dd-14f2d8578e8b.png">
|
design
|
remove decimals from numbers summary as a dao user i don t want to see decimals in the dao because the numbers are too long and confusing ac remove decimals behind the dot on following numbers proposal detail page token for utility cash voice proposal creation wizard token for utility cash voice wallet widget on profile wallet widget on wallet page img width alt image src img width alt image src img width alt image src leave it with decimals on the wallet page table img width alt image src
| 1
|
24,260
| 5,041,344,980
|
IssuesEvent
|
2016-12-19 10:00:54
|
salesforce-ux/design-system
|
https://api.github.com/repos/salesforce-ux/design-system
|
closed
|
SLDS.com / Grid / Order
|
bug documentation
|
https://www.lightningdesignsystem.com/components/utilities/grid/#flavor-order
There's a bug with the example:
- large: orders correctly (reversed: 3, 2, 1)
- medium: orders correctly (1, 2, 3)
- small: orders the same as medium instead of (3, 1, 2)...
(Using Chrome 53.0.2785.143, Firefox 45.0.1, Opera 40.0.2308.81)
_I have no idea how to add labels or anything, but it should be labeled as a bug._
|
1.0
|
SLDS.com / Grid / Order - https://www.lightningdesignsystem.com/components/utilities/grid/#flavor-order
There's a bug with the example:
- large: orders correctly (reversed: 3, 2, 1)
- medium: orders correctly (1, 2, 3)
- small: orders the same as medium instead of (3, 1, 2)...
(Using Chrome 53.0.2785.143, Firefox 45.0.1, Opera 40.0.2308.81)
_I have no idea how to add labels or anything, but it should be labeled as a bug._
|
non_design
|
slds com grid order there s a bug with the example large orders correctly reversed medium orders correctly small orders the same as medium instead of using chrome firefox opera i have no idea how to add labels or anything but it should be labeled as a bug
| 0
|
115,196
| 14,702,535,080
|
IssuesEvent
|
2021-01-04 13:45:21
|
EmbarkStudios/opensource-website
|
https://api.github.com/repos/EmbarkStudios/opensource-website
|
closed
|
Create a nice view for newsletter archives and signup
|
design enhancement good first issue help wanted
|
**Describe the solution you'd like**
Currently, the newsletter is displayed as a basic HTML list on the homepage. We should have a separate page (embark.dev/newsletter) that shows recent newsletters in a nice way, and allows users to sign up. (this is currently done through the basic mailchimp page)
|
1.0
|
Create a nice view for newsletter archives and signup - **Describe the solution you'd like**
Currently, the newsletter is displayed as a basic HTML list on the homepage. We should have a separate page (embark.dev/newsletter) that shows recent newsletters in a nice way, and allows users to sign up. (this is currently done through the basic mailchimp page)
|
design
|
create a nice view for newsletter archives and signup describe the solution you d like currently the newsletter is displayed as a basic html list on the homepage we should have a separate page embark dev newsletter that shows recent newsletters in a nice way and allows users to sign up this is currently done through the basic mailchimp page
| 1
|
56,729
| 15,343,044,901
|
IssuesEvent
|
2021-02-27 18:38:27
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
The crystallball distribution entropy is sometimes minus infinity
|
defect scipy.stats
|
<!--
Thank you for taking the time to file a bug report.
Please fill in the fields below, deleting the sections that
don't apply to your issue. You can view the final output
by clicking the preview button above.
Note: This is a comment, and won't appear in the output.
-->
The crystalball entropy function returns -inf for valid shape parameters.
#### Reproducing code example:
<!--
f you place your code between the triple backticks below,
it will be rendered as a code block.
-->
```
from scipy.stats import crystalball
crystalball.entropy(2, 3)
```
#### Error message:
<!-- If any, paste the *full* error message inside a code block
as above (starting from line Traceback)
-->
```
>>> cb.entropy(2,3)
array(-inf)
```
#### Scipy/Numpy/Python version information:
```
1.7.0.dev0+5ac5f63 1.19.4 sys.version_info(major=3, minor=9, micro=1, releaselevel='final', serial=0)
```
|
1.0
|
The crystallball distribution entropy is sometimes minus infinity - <!--
Thank you for taking the time to file a bug report.
Please fill in the fields below, deleting the sections that
don't apply to your issue. You can view the final output
by clicking the preview button above.
Note: This is a comment, and won't appear in the output.
-->
The crystalball entropy function returns -inf for valid shape parameters.
#### Reproducing code example:
<!--
f you place your code between the triple backticks below,
it will be rendered as a code block.
-->
```
from scipy.stats import crystalball
crystalball.entropy(2, 3)
```
#### Error message:
<!-- If any, paste the *full* error message inside a code block
as above (starting from line Traceback)
-->
```
>>> cb.entropy(2,3)
array(-inf)
```
#### Scipy/Numpy/Python version information:
```
1.7.0.dev0+5ac5f63 1.19.4 sys.version_info(major=3, minor=9, micro=1, releaselevel='final', serial=0)
```
|
non_design
|
the crystallball distribution entropy is sometimes minus infinity thank you for taking the time to file a bug report please fill in the fields below deleting the sections that don t apply to your issue you can view the final output by clicking the preview button above note this is a comment and won t appear in the output the crystalball entropy function returns inf for valid shape parameters reproducing code example f you place your code between the triple backticks below it will be rendered as a code block from scipy stats import crystalball crystalball entropy error message if any paste the full error message inside a code block as above starting from line traceback cb entropy array inf scipy numpy python version information sys version info major minor micro releaselevel final serial
| 0
|
11,769
| 7,447,011,814
|
IssuesEvent
|
2018-03-28 11:01:21
|
nerdalize/nerd
|
https://api.github.com/repos/nerdalize/nerd
|
opened
|
Unhelpful error when specifying invalid kubeconfig path
|
usability
|
## Expected Behavior
When specifying an invalid kubeconfig path, it should return a proper error and not a "not logged in" one
## Actual Behavior

|
True
|
Unhelpful error when specifying invalid kubeconfig path - ## Expected Behavior
When specifying an invalid kubeconfig path, it should return a proper error and not a "not logged in" one
## Actual Behavior

|
non_design
|
unhelpful error when specifying invalid kubeconfig path expected behavior when specifying an invalid kubeconfig path it should return a proper error and not a not logged in one actual behavior
| 0
|
115,876
| 14,900,759,781
|
IssuesEvent
|
2021-01-21 15:44:44
|
fremtind/jokul
|
https://api.github.com/repos/fremtind/jokul
|
closed
|
Feil: For Designere peker til en lenke som ikke finnes lengre.
|
🎨 design 🐝 bug 👽portal 📚 documentation
|
**Feilbeskrivelse**
"Klikk på lenken i sidemenyen", det er ikke lengre noen lenke i sidemenyen som kan ta deg til figma

**For å reprodusere**
Fortell oss hvordan vi kan gjenskape feilen (det hjelper oss med å fikse den):
1. Gå til https://jokul.fremtind.no/komigang/design
**Forventet oppførsel**
Det er en lenke, eller hvertfall ikke referanser i teksten til en lenke som ikke finnes.
|
1.0
|
Feil: For Designere peker til en lenke som ikke finnes lengre. - **Feilbeskrivelse**
"Klikk på lenken i sidemenyen", det er ikke lengre noen lenke i sidemenyen som kan ta deg til figma

**For å reprodusere**
Fortell oss hvordan vi kan gjenskape feilen (det hjelper oss med å fikse den):
1. Gå til https://jokul.fremtind.no/komigang/design
**Forventet oppførsel**
Det er en lenke, eller hvertfall ikke referanser i teksten til en lenke som ikke finnes.
|
design
|
feil for designere peker til en lenke som ikke finnes lengre feilbeskrivelse klikk på lenken i sidemenyen det er ikke lengre noen lenke i sidemenyen som kan ta deg til figma for å reprodusere fortell oss hvordan vi kan gjenskape feilen det hjelper oss med å fikse den gå til forventet oppførsel det er en lenke eller hvertfall ikke referanser i teksten til en lenke som ikke finnes
| 1
|
137,323
| 20,115,759,472
|
IssuesEvent
|
2022-02-07 19:20:21
|
influxdata/ui
|
https://api.github.com/repos/influxdata/ui
|
opened
|
Take Inventory of Honeybadger Calls: Determine Appropriate Amplitude Events
|
team/design
|
Take Inventory of Honeybadger Calls: Determine Appropriate Amplitude Events
|
1.0
|
Take Inventory of Honeybadger Calls: Determine Appropriate Amplitude Events - Take Inventory of Honeybadger Calls: Determine Appropriate Amplitude Events
|
design
|
take inventory of honeybadger calls determine appropriate amplitude events take inventory of honeybadger calls determine appropriate amplitude events
| 1
|
40,045
| 5,169,836,730
|
IssuesEvent
|
2017-01-18 02:44:04
|
c4gnv/c4gnv.github.io
|
https://api.github.com/repos/c4gnv/c4gnv.github.io
|
closed
|
Review admin site designs
|
design
|
Hey there, team! Please take a look at the visual designs I've put together for our c4gnv.com site and leave your feedback here.
Home: https://www.dropbox.com/s/lccvsow6gdfew1m/home.png?dl=0
Get Started: https://www.dropbox.com/s/twa0av2v33fcqjc/get-started.png?dl=0
|
1.0
|
Review admin site designs - Hey there, team! Please take a look at the visual designs I've put together for our c4gnv.com site and leave your feedback here.
Home: https://www.dropbox.com/s/lccvsow6gdfew1m/home.png?dl=0
Get Started: https://www.dropbox.com/s/twa0av2v33fcqjc/get-started.png?dl=0
|
design
|
review admin site designs hey there team please take a look at the visual designs i ve put together for our com site and leave your feedback here home get started
| 1
|
438,475
| 12,639,481,283
|
IssuesEvent
|
2020-06-16 00:02:04
|
seccomp/libseccomp-golang
|
https://api.github.com/repos/seccomp/libseccomp-golang
|
opened
|
BUG: rename the "master" branch to "main" #255
|
bug priority/medium
|
Similar to [PR #246](https://github.com/seccomp/libseccomp/pull/246) we need to change the name of the "master" branch to "main", or similar. This is quite easy, and via a branch rename, it can be done without losing any of the git history; there is no good reason to *not* do this.
The steps are simple:
1. Rename the local git repostory: `git branch -m master main`
2. Push the "main" branch to GitHub: `git push origin main`
3. Switch the default branch on GitHub to "main" (see Settings -> Branches)
4. Merge any outstanding PRs against "master" into "main"
5. Once all PR's referencing "master" have been merged, delete the "master" branch: `git push origin :master`
|
1.0
|
BUG: rename the "master" branch to "main" #255 - Similar to [PR #246](https://github.com/seccomp/libseccomp/pull/246) we need to change the name of the "master" branch to "main", or similar. This is quite easy, and via a branch rename, it can be done without losing any of the git history; there is no good reason to *not* do this.
The steps are simple:
1. Rename the local git repostory: `git branch -m master main`
2. Push the "main" branch to GitHub: `git push origin main`
3. Switch the default branch on GitHub to "main" (see Settings -> Branches)
4. Merge any outstanding PRs against "master" into "main"
5. Once all PR's referencing "master" have been merged, delete the "master" branch: `git push origin :master`
|
non_design
|
bug rename the master branch to main similar to we need to change the name of the master branch to main or similar this is quite easy and via a branch rename it can be done without losing any of the git history there is no good reason to not do this the steps are simple rename the local git repostory git branch m master main push the main branch to github git push origin main switch the default branch on github to main see settings branches merge any outstanding prs against master into main once all pr s referencing master have been merged delete the master branch git push origin master
| 0
|
517,288
| 15,001,435,536
|
IssuesEvent
|
2021-01-30 00:08:47
|
IDAES/idaes-pse
|
https://api.github.com/repos/IDAES/idaes-pse
|
closed
|
Units problem with drum.py model when updating to Pyomo master
|
Priority:High bug
|
@jsiirola and I are testing idaes tests with the Pyomo master before we bump the IDAES Pyomo version. We encountered the following error with the test_drum.py in the power generation unit model library.
> Units problem with expression fs.unit.control_volume.volume[0.0] - (((asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter))) + 1.5707963267948966*rad)*(0.5*fs.unit.drum_diameter)**2 + (0.5*fs.unit.drum_diameter)*cos(asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter)))*(fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter)))*fs.unit.drum_length
> Error in units when checking fs.unit.volume_eqn[0.0]
> Error in units when checking fs.unit
> Error in units when checking fs
> Error in units when checking unknown
> ================================================================================================================================= slowest 100 durations =================================================================================================================================
> 0.83s call idaes/power_generation/unit_models/tests/test_drum.py::test_units
> 0.20s call idaes/power_generation/unit_models/tests/test_drum.py::test_initialize_drum
> 0.10s setup idaes/power_generation/unit_models/tests/test_drum.py::test_basic_build
> 0.09s call idaes/power_generation/unit_models/tests/test_drum.py::test_run_drum
>
> (8 durations < 0.005s hidden. Use -vv to show these durations.)
> ================================================================================================================================ short test summary info ================================================================================================================================
> FAILED test_drum.py::test_units - pyomo.core.base.units_container.InconsistentUnitsError: Error in units found in expression: ((asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter))) + 1.5707963267948966*rad)*(0.5*fs.unit.drum_diameter)**2 + (...
To reproduce this error, you will need the following:
1. Current Pyomo master
2. `units-updates` from @jsiirola's idaes fork: https://github.com/jsiirola/idaes-pse/tree/units-updates
Once fixed, I think it is best to push changes back to @jsiirola's branch and we can create a PR from there.
|
1.0
|
Units problem with drum.py model when updating to Pyomo master - @jsiirola and I are testing idaes tests with the Pyomo master before we bump the IDAES Pyomo version. We encountered the following error with the test_drum.py in the power generation unit model library.
> Units problem with expression fs.unit.control_volume.volume[0.0] - (((asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter))) + 1.5707963267948966*rad)*(0.5*fs.unit.drum_diameter)**2 + (0.5*fs.unit.drum_diameter)*cos(asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter)))*(fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter)))*fs.unit.drum_length
> Error in units when checking fs.unit.volume_eqn[0.0]
> Error in units when checking fs.unit
> Error in units when checking fs
> Error in units when checking unknown
> ================================================================================================================================= slowest 100 durations =================================================================================================================================
> 0.83s call idaes/power_generation/unit_models/tests/test_drum.py::test_units
> 0.20s call idaes/power_generation/unit_models/tests/test_drum.py::test_initialize_drum
> 0.10s setup idaes/power_generation/unit_models/tests/test_drum.py::test_basic_build
> 0.09s call idaes/power_generation/unit_models/tests/test_drum.py::test_run_drum
>
> (8 durations < 0.005s hidden. Use -vv to show these durations.)
> ================================================================================================================================ short test summary info ================================================================================================================================
> FAILED test_drum.py::test_units - pyomo.core.base.units_container.InconsistentUnitsError: Error in units found in expression: ((asin((fs.unit.drum_level[0.0] - (0.5*fs.unit.drum_diameter))/(0.5*fs.unit.drum_diameter))) + 1.5707963267948966*rad)*(0.5*fs.unit.drum_diameter)**2 + (...
To reproduce this error, you will need the following:
1. Current Pyomo master
2. `units-updates` from @jsiirola's idaes fork: https://github.com/jsiirola/idaes-pse/tree/units-updates
Once fixed, I think it is best to push changes back to @jsiirola's branch and we can create a PR from there.
|
non_design
|
units problem with drum py model when updating to pyomo master jsiirola and i are testing idaes tests with the pyomo master before we bump the idaes pyomo version we encountered the following error with the test drum py in the power generation unit model library units problem with expression fs unit control volume volume asin fs unit drum level fs unit drum diameter fs unit drum diameter rad fs unit drum diameter fs unit drum diameter cos asin fs unit drum level fs unit drum diameter fs unit drum diameter fs unit drum level fs unit drum diameter fs unit drum length error in units when checking fs unit volume eqn error in units when checking fs unit error in units when checking fs error in units when checking unknown slowest durations call idaes power generation unit models tests test drum py test units call idaes power generation unit models tests test drum py test initialize drum setup idaes power generation unit models tests test drum py test basic build call idaes power generation unit models tests test drum py test run drum durations hidden use vv to show these durations short test summary info failed test drum py test units pyomo core base units container inconsistentunitserror error in units found in expression asin fs unit drum level fs unit drum diameter fs unit drum diameter rad fs unit drum diameter to reproduce this error you will need the following current pyomo master units updates from jsiirola s idaes fork once fixed i think it is best to push changes back to jsiirola s branch and we can create a pr from there
| 0
|
42,613
| 5,502,783,498
|
IssuesEvent
|
2017-03-16 01:01:59
|
kubernetes-incubator/bootkube
|
https://api.github.com/repos/kubernetes-incubator/bootkube
|
closed
|
self hosted etcd: checkpoint iptables on master nodes
|
kind/design kind/enhancement priority/P1
|
self hosted etcd relies on service IP to work correctly. Kuberetes API server contact etcd pod by service IP (load balancing + hide the actual etcd pod IP which is subject to change).
Service IP relies on API server to be restored after a machine reboot. If we restart all API servers at the same time, service IP is not recoverable.
To solve this chicken and egg issue, we have to checkpoint the iptables (which do all the heavy-lifting for service IP).
I have tried the IP table checkpoint approach, it works well at least for the hack/multi-node example.
/cc @aaronlevy
|
1.0
|
self hosted etcd: checkpoint iptables on master nodes - self hosted etcd relies on service IP to work correctly. Kuberetes API server contact etcd pod by service IP (load balancing + hide the actual etcd pod IP which is subject to change).
Service IP relies on API server to be restored after a machine reboot. If we restart all API servers at the same time, service IP is not recoverable.
To solve this chicken and egg issue, we have to checkpoint the iptables (which do all the heavy-lifting for service IP).
I have tried the IP table checkpoint approach, it works well at least for the hack/multi-node example.
/cc @aaronlevy
|
design
|
self hosted etcd checkpoint iptables on master nodes self hosted etcd relies on service ip to work correctly kuberetes api server contact etcd pod by service ip load balancing hide the actual etcd pod ip which is subject to change service ip relies on api server to be restored after a machine reboot if we restart all api servers at the same time service ip is not recoverable to solve this chicken and egg issue we have to checkpoint the iptables which do all the heavy lifting for service ip i have tried the ip table checkpoint approach it works well at least for the hack multi node example cc aaronlevy
| 1
|
114,766
| 14,633,060,870
|
IssuesEvent
|
2020-12-24 00:31:17
|
keepid/keepid_client
|
https://api.github.com/repos/keepid/keepid_client
|
opened
|
Create breakpoint sizing on XL/Lg/Md/Sm/Xs screens
|
Design Component
|
Make breakpoints for these modals. Remember for small screens (sm and xs) the modal will resize to the whole screen
|
1.0
|
Create breakpoint sizing on XL/Lg/Md/Sm/Xs screens - Make breakpoints for these modals. Remember for small screens (sm and xs) the modal will resize to the whole screen
|
design
|
create breakpoint sizing on xl lg md sm xs screens make breakpoints for these modals remember for small screens sm and xs the modal will resize to the whole screen
| 1
|
104,848
| 4,226,100,885
|
IssuesEvent
|
2016-07-02 07:42:27
|
The-Compiler/qutebrowser
|
https://api.github.com/repos/The-Compiler/qutebrowser
|
closed
|
Don't reload page when url didn't change with `:edit-url`
|
easy priority: 2 - low
|
Everything is in the title, since I bound `e` to `:edit-url` and sometimes press it by mistake, I would like to be able to quickly undo this (so I just hit `ZZ` in vim), but this will then reload the page, which I don't want since the URL hasn't changed.
Do you agree ?
|
1.0
|
Don't reload page when url didn't change with `:edit-url` - Everything is in the title, since I bound `e` to `:edit-url` and sometimes press it by mistake, I would like to be able to quickly undo this (so I just hit `ZZ` in vim), but this will then reload the page, which I don't want since the URL hasn't changed.
Do you agree ?
|
non_design
|
don t reload page when url didn t change with edit url everything is in the title since i bound e to edit url and sometimes press it by mistake i would like to be able to quickly undo this so i just hit zz in vim but this will then reload the page which i don t want since the url hasn t changed do you agree
| 0
|
789
| 2,905,112,497
|
IssuesEvent
|
2015-06-18 21:41:55
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
Publish signing key and security team key to keybase.io
|
A-infrastructure
|
keybase.io is a seemingly-popular new service for validating GPG keys. I'd like to publish our keys there but haven't gotten around to it.
|
1.0
|
Publish signing key and security team key to keybase.io - keybase.io is a seemingly-popular new service for validating GPG keys. I'd like to publish our keys there but haven't gotten around to it.
|
non_design
|
publish signing key and security team key to keybase io keybase io is a seemingly popular new service for validating gpg keys i d like to publish our keys there but haven t gotten around to it
| 0
|
182,025
| 30,779,321,220
|
IssuesEvent
|
2023-07-31 08:55:47
|
opencollective/opencollective
|
https://api.github.com/repos/opencollective/opencollective
|
opened
|
OCR: Handle currency mismatches
|
frontend enhancement needs design
|
Part of https://github.com/opencollective/opencollective/issues/6865
Require https://github.com/opencollective/opencollective/issues/6903
In https://github.com/opencollective/opencollective/issues/6903, we'll start with not prefilling amounts in case they're expressed with a currency that is different from the expense.
Following up on that, we would like to be smart and try to convert the amount to the desired currency, but we need the UX to be really clear to make sure the expense submitter double-checks the amount and knows which exchange rate was used.
|
1.0
|
OCR: Handle currency mismatches - Part of https://github.com/opencollective/opencollective/issues/6865
Require https://github.com/opencollective/opencollective/issues/6903
In https://github.com/opencollective/opencollective/issues/6903, we'll start with not prefilling amounts in case they're expressed with a currency that is different from the expense.
Following up on that, we would like to be smart and try to convert the amount to the desired currency, but we need the UX to be really clear to make sure the expense submitter double-checks the amount and knows which exchange rate was used.
|
design
|
ocr handle currency mismatches part of require in we ll start with not prefilling amounts in case they re expressed with a currency that is different from the expense following up on that we would like to be smart and try to convert the amount to the desired currency but we need the ux to be really clear to make sure the expense submitter double checks the amount and knows which exchange rate was used
| 1
|
177,931
| 29,193,313,400
|
IssuesEvent
|
2023-05-19 22:55:21
|
prettierlichess/prettierlichess
|
https://api.github.com/repos/prettierlichess/prettierlichess
|
closed
|
Board is oversized
|
design request
|

I downloaded the latest ver of the extension, it fixed the player window discrepancy but now the board is oversized. If you can fix this would be appreciated, thanks in advance!
|
1.0
|
Board is oversized - 
I downloaded the latest ver of the extension, it fixed the player window discrepancy but now the board is oversized. If you can fix this would be appreciated, thanks in advance!
|
design
|
board is oversized i downloaded the latest ver of the extension it fixed the player window discrepancy but now the board is oversized if you can fix this would be appreciated thanks in advance
| 1
|
174,706
| 27,712,188,144
|
IssuesEvent
|
2023-03-14 14:52:44
|
hpi-swa-lab/BP2021RH1
|
https://api.github.com/repos/hpi-swa-lab/BP2021RH1
|
opened
|
Guide
|
U-low C-design-prototype
|
Es sollte einen Guide durch die Seite geben. Das kann auch in Form von Tooltips während der Nutzung passieren.
|
1.0
|
Guide - Es sollte einen Guide durch die Seite geben. Das kann auch in Form von Tooltips während der Nutzung passieren.
|
design
|
guide es sollte einen guide durch die seite geben das kann auch in form von tooltips während der nutzung passieren
| 1
|
155,079
| 24,397,940,565
|
IssuesEvent
|
2022-10-04 21:10:42
|
dotnet/efcore
|
https://api.github.com/repos/dotnet/efcore
|
closed
|
Migration is not smart enough to combine mutually exclusive settings
|
closed-by-design customer-reported
|
I am not sure whether it is a bug or an already-known issue that is left as it is (because it is not harmful).
The following
```csharp
class Person
{
public int Id { get; set; }
public string? FullName { get; set; }
public string? Biography { get; set; }
}
```
```csharp
class MyContext : DbContext
{
public MyContext(DbContextOptions<MyContext> options) : base(options) { }
protected override void ConfigureConventions(ModelConfigurationBuilder mcb)
{
mcb.Properties<string>().HaveMaxLength(150);
}
protected override void OnModelCreating(ModelBuilder mb)
{
mb.Entity<Person>().Property(p => p.Biography).HasColumnType("nvarchar(max)");
}
}
```
```csharp
class Program : IDesignTimeDbContextFactory<MyContext>
{
public MyContext CreateDbContext(string[] args)
{
string constr = "server=.;database=EFCoreTestDb;integrated security=true";
var dcob = new DbContextOptionsBuilder<MyContext>();
dcob.UseSqlServer(constr);
return new MyContext(dcob.Options);
}
static void Main() { }
}
```
produce a migration as follows:
```csharp
protected override void Up(MigrationBuilder migrationBuilder)
{
migrationBuilder.CreateTable(
name: "Person",
columns: table => new
{
Id = table.Column<int>(type: "int", nullable: false)
.Annotation("SqlServer:Identity", "1, 1"),
FullName = table.Column<string>(type: "nvarchar(150)", maxLength: 150, nullable: true),
Biography = table.Column<string>(type: "nvarchar(max)", maxLength: 150, nullable: true)
},
constraints: table =>
{
table.PrimaryKey("PK_Person", x => x.Id);
});
}
```
As you can see, the migration still assigns `maxLength: 150` to `Biography` of type `nvarchar(max)`. It means that migration is not smart enough to infer that `nvarchar(max)` with `maxLength: 150` does not make sense.
Fortunately, the generated database is correct, the field `Biography` is of type `nvarchar(max)`.
|
1.0
|
Migration is not smart enough to combine mutually exclusive settings - I am not sure whether it is a bug or an already-known issue that is left as it is (because it is not harmful).
The following
```csharp
class Person
{
public int Id { get; set; }
public string? FullName { get; set; }
public string? Biography { get; set; }
}
```
```csharp
class MyContext : DbContext
{
public MyContext(DbContextOptions<MyContext> options) : base(options) { }
protected override void ConfigureConventions(ModelConfigurationBuilder mcb)
{
mcb.Properties<string>().HaveMaxLength(150);
}
protected override void OnModelCreating(ModelBuilder mb)
{
mb.Entity<Person>().Property(p => p.Biography).HasColumnType("nvarchar(max)");
}
}
```
```csharp
class Program : IDesignTimeDbContextFactory<MyContext>
{
public MyContext CreateDbContext(string[] args)
{
string constr = "server=.;database=EFCoreTestDb;integrated security=true";
var dcob = new DbContextOptionsBuilder<MyContext>();
dcob.UseSqlServer(constr);
return new MyContext(dcob.Options);
}
static void Main() { }
}
```
produce a migration as follows:
```csharp
protected override void Up(MigrationBuilder migrationBuilder)
{
migrationBuilder.CreateTable(
name: "Person",
columns: table => new
{
Id = table.Column<int>(type: "int", nullable: false)
.Annotation("SqlServer:Identity", "1, 1"),
FullName = table.Column<string>(type: "nvarchar(150)", maxLength: 150, nullable: true),
Biography = table.Column<string>(type: "nvarchar(max)", maxLength: 150, nullable: true)
},
constraints: table =>
{
table.PrimaryKey("PK_Person", x => x.Id);
});
}
```
As you can see, the migration still assigns `maxLength: 150` to `Biography` of type `nvarchar(max)`. It means that migration is not smart enough to infer that `nvarchar(max)` with `maxLength: 150` does not make sense.
Fortunately, the generated database is correct, the field `Biography` is of type `nvarchar(max)`.
|
design
|
migration is not smart enough to combine mutually exclusive settings i am not sure whether it is a bug or an already known issue that is left as it is because it is not harmful the following csharp class person public int id get set public string fullname get set public string biography get set csharp class mycontext dbcontext public mycontext dbcontextoptions options base options protected override void configureconventions modelconfigurationbuilder mcb mcb properties havemaxlength protected override void onmodelcreating modelbuilder mb mb entity property p p biography hascolumntype nvarchar max csharp class program idesigntimedbcontextfactory public mycontext createdbcontext string args string constr server database efcoretestdb integrated security true var dcob new dbcontextoptionsbuilder dcob usesqlserver constr return new mycontext dcob options static void main produce a migration as follows csharp protected override void up migrationbuilder migrationbuilder migrationbuilder createtable name person columns table new id table column type int nullable false annotation sqlserver identity fullname table column type nvarchar maxlength nullable true biography table column type nvarchar max maxlength nullable true constraints table table primarykey pk person x x id as you can see the migration still assigns maxlength to biography of type nvarchar max it means that migration is not smart enough to infer that nvarchar max with maxlength does not make sense fortunately the generated database is correct the field biography is of type nvarchar max
| 1
|
143,276
| 21,993,500,196
|
IssuesEvent
|
2022-05-26 02:11:42
|
harryodubhghaill/CI-Portforlio-4-blogsocial
|
https://api.github.com/repos/harryodubhghaill/CI-Portforlio-4-blogsocial
|
closed
|
Design database structure
|
SysAdmin Design
|
Design a relational database structure to allow for the full functionality of the site.
- [ ] User
- [ ] Post
- [ ] Comment
- [ ] Group
|
1.0
|
Design database structure - Design a relational database structure to allow for the full functionality of the site.
- [ ] User
- [ ] Post
- [ ] Comment
- [ ] Group
|
design
|
design database structure design a relational database structure to allow for the full functionality of the site user post comment group
| 1
|
90,650
| 11,424,360,830
|
IssuesEvent
|
2020-02-03 17:35:14
|
are-you-still-watching/web-app
|
https://api.github.com/repos/are-you-still-watching/web-app
|
opened
|
Create profile page buttons
|
design
|
Create profile page buttons, including:
- [ ] netflix
- [ ] crave
- [ ] disney +
- [ ] prime video
|
1.0
|
Create profile page buttons - Create profile page buttons, including:
- [ ] netflix
- [ ] crave
- [ ] disney +
- [ ] prime video
|
design
|
create profile page buttons create profile page buttons including netflix crave disney prime video
| 1
|
173,643
| 27,503,716,541
|
IssuesEvent
|
2023-03-05 23:53:35
|
penumbra-zone/penumbra
|
https://api.github.com/repos/penumbra-zone/penumbra
|
closed
|
Specify note contents
|
A-shielded-crypto E-medium C-design
|
Fill in this section of the protocol spec; settle on (and write up) a choice of `leadByte` method.
|
1.0
|
Specify note contents - Fill in this section of the protocol spec; settle on (and write up) a choice of `leadByte` method.
|
design
|
specify note contents fill in this section of the protocol spec settle on and write up a choice of leadbyte method
| 1
|
63,731
| 7,740,452,286
|
IssuesEvent
|
2018-05-28 21:49:53
|
vtex/styleguide
|
https://api.github.com/repos/vtex/styleguide
|
closed
|
Empty state
|
Design Done
|
Here's the Empty State pattern we use in the Credit Control module.
(It's already implemented, so it's just a matter of extracting it :)

What problem it solves
===
Shows _something_ instead of _nothing_.
Explains what should be in that particular space, and encourages the designer/developer to instruct the user what to do.
According to [Swetha Suresh](https://medium.com/@swethasuresh1108/the-empty-states-design-mantra-91c56eb88b3b), an Empty State:
1. Informs about the empty screen, the whats, whys, and hows
2. Prompt towards taking an action to populate the platform
3. Educates about the platform and all its features
Design decisions
===
- A muted color ~and non-primary button~ (thx @rsimoens) are used so this component has a weak prominence in the screen.
- The container borders helps show the space that would be normally filled in the screen if it wasn't "empty". This reduces cognitive load on consequent user visits.
- Although Empty States [in the wild](http://emptystat.es/) often include a personalized icon, we decided not to include it here for now. In the Credit Control case we had half a dozen empty states possible, and it would be cumbersome to search for custom icons for each one, and we couldn't come up with a reusable pattern for that. In addition to that, an icon, even if with muted colors, adds a lot of visual weight to this section, which in our case wasn't desirable.
- The rounded corners are simple decorative elements that makes this component fit better with VTEX visual patterns.
Example of application
===

Props
===
The idea is that all fields (heading, body and call-to-action) are optional, however at least a heading or a body should be present for it to make sense.
Optionally the call-to-action can be a link, let's say to a VTEX Help article for example.
Figma link
===
https://www.figma.com/file/a94lX91ZmYGIDBxvb8shwLn8/VTEX-Design?node-id=191%3A1823
|
1.0
|
Empty state - Here's the Empty State pattern we use in the Credit Control module.
(It's already implemented, so it's just a matter of extracting it :)

What problem it solves
===
Shows _something_ instead of _nothing_.
Explains what should be in that particular space, and encourages the designer/developer to instruct the user what to do.
According to [Swetha Suresh](https://medium.com/@swethasuresh1108/the-empty-states-design-mantra-91c56eb88b3b), an Empty State:
1. Informs about the empty screen, the whats, whys, and hows
2. Prompt towards taking an action to populate the platform
3. Educates about the platform and all its features
Design decisions
===
- A muted color ~and non-primary button~ (thx @rsimoens) are used so this component has a weak prominence in the screen.
- The container borders helps show the space that would be normally filled in the screen if it wasn't "empty". This reduces cognitive load on consequent user visits.
- Although Empty States [in the wild](http://emptystat.es/) often include a personalized icon, we decided not to include it here for now. In the Credit Control case we had half a dozen empty states possible, and it would be cumbersome to search for custom icons for each one, and we couldn't come up with a reusable pattern for that. In addition to that, an icon, even if with muted colors, adds a lot of visual weight to this section, which in our case wasn't desirable.
- The rounded corners are simple decorative elements that makes this component fit better with VTEX visual patterns.
Example of application
===

Props
===
The idea is that all fields (heading, body and call-to-action) are optional, however at least a heading or a body should be present for it to make sense.
Optionally the call-to-action can be a link, let's say to a VTEX Help article for example.
Figma link
===
https://www.figma.com/file/a94lX91ZmYGIDBxvb8shwLn8/VTEX-Design?node-id=191%3A1823
|
design
|
empty state here s the empty state pattern we use in the credit control module it s already implemented so it s just a matter of extracting it what problem it solves shows something instead of nothing explains what should be in that particular space and encourages the designer developer to instruct the user what to do according to an empty state informs about the empty screen the whats whys and hows prompt towards taking an action to populate the platform educates about the platform and all its features design decisions a muted color and non primary button thx rsimoens are used so this component has a weak prominence in the screen the container borders helps show the space that would be normally filled in the screen if it wasn t empty this reduces cognitive load on consequent user visits although empty states often include a personalized icon we decided not to include it here for now in the credit control case we had half a dozen empty states possible and it would be cumbersome to search for custom icons for each one and we couldn t come up with a reusable pattern for that in addition to that an icon even if with muted colors adds a lot of visual weight to this section which in our case wasn t desirable the rounded corners are simple decorative elements that makes this component fit better with vtex visual patterns example of application props the idea is that all fields heading body and call to action are optional however at least a heading or a body should be present for it to make sense optionally the call to action can be a link let s say to a vtex help article for example figma link
| 1
|
103,966
| 22,534,119,974
|
IssuesEvent
|
2022-06-25 01:19:36
|
macder/medusa-fulfillment-shippo
|
https://api.github.com/repos/macder/medusa-fulfillment-shippo
|
closed
|
Dev - automate config source references
|
code improvement chore
|
eliminate the manual step of changing the config reference (standalone localhost vs medusa package)
|
1.0
|
Dev - automate config source references - eliminate the manual step of changing the config reference (standalone localhost vs medusa package)
|
non_design
|
dev automate config source references eliminate the manual step of changing the config reference standalone localhost vs medusa package
| 0
|
91,381
| 11,498,792,194
|
IssuesEvent
|
2020-02-12 12:44:34
|
liqd/adhocracy-plus
|
https://api.github.com/repos/liqd/adhocracy-plus
|
closed
|
changes in organisation page
|
Type: UX/UI or design
|
URL: https://aplus-dev.liqd.net/teststadt/
Comment/Question:
If possible, I would increase the image height cause most of the pictures aren't that stretched. The layout had 345px height.
I would also leave the project description text using the same paragraph style as the about text on top.
I know I already mentioned, but just to leave it registered, the CTA is missing. :)
Also, If I send you the shadow style precisely can we change that as well?
Thanks!

|
1.0
|
changes in organisation page - URL: https://aplus-dev.liqd.net/teststadt/
Comment/Question:
If possible, I would increase the image height cause most of the pictures aren't that stretched. The layout had 345px height.
I would also leave the project description text using the same paragraph style as the about text on top.
I know I already mentioned, but just to leave it registered, the CTA is missing. :)
Also, If I send you the shadow style precisely can we change that as well?
Thanks!

|
design
|
changes in organisation page url comment question if possible i would increase the image height cause most of the pictures aren t that stretched the layout had height i would also leave the project description text using the same paragraph style as the about text on top i know i already mentioned but just to leave it registered the cta is missing also if i send you the shadow style precisely can we change that as well thanks
| 1
|
84,928
| 10,573,271,105
|
IssuesEvent
|
2019-10-07 11:34:36
|
fac-17/Generation-Change
|
https://api.github.com/repos/fac-17/Generation-Change
|
closed
|
prepare presentation on design week
|
design important 1
|
- [ ] prepare demo for design week figmas
- [ ] talk about what went well and what could have gone better
- [ ] show things that we've learnt
- [ ] where we go from here
|
1.0
|
prepare presentation on design week - - [ ] prepare demo for design week figmas
- [ ] talk about what went well and what could have gone better
- [ ] show things that we've learnt
- [ ] where we go from here
|
design
|
prepare presentation on design week prepare demo for design week figmas talk about what went well and what could have gone better show things that we ve learnt where we go from here
| 1
|
111,893
| 14,169,463,610
|
IssuesEvent
|
2020-11-12 13:16:16
|
ajency/Dhanda-App
|
https://api.github.com/repos/ajency/Dhanda-App
|
closed
|
Opening Balance Page: The font size on the second button is too big. Make it the same size as " Create staff "
|
Assigned to QA Priority: High UI/ design bug
|

|
1.0
|
Opening Balance Page: The font size on the second button is too big. Make it the same size as " Create staff " -

|
design
|
opening balance page the font size on the second button is too big make it the same size as create staff
| 1
|
149,159
| 11,884,393,376
|
IssuesEvent
|
2020-03-27 17:34:02
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
PortResever doesn't work correctly for netcoreapp5.0
|
Area:Test
|
PortReserver has the following code, setting TimeSpan.Zero for a CancellationTokenSource:
[` var tryOnceCts = new CancellationTokenSource(TimeSpan.Zero);`](https://github.com/NuGet/NuGet.Client/blob/dev/test/TestUtilities/Test.Utility/TestServer/PortReserver.cs#L60)
In netcore5.0, the task will get cancelled, so it will never get a chance to run task on this port. Then it will keep trying to increase the port until it exceeds the max value(65535):
`typeof(System.InvalidOperationException): Exceeded port range`
|
1.0
|
PortResever doesn't work correctly for netcoreapp5.0 - PortReserver has the following code, setting TimeSpan.Zero for a CancellationTokenSource:
[` var tryOnceCts = new CancellationTokenSource(TimeSpan.Zero);`](https://github.com/NuGet/NuGet.Client/blob/dev/test/TestUtilities/Test.Utility/TestServer/PortReserver.cs#L60)
In netcore5.0, the task will get cancelled, so it will never get a chance to run task on this port. Then it will keep trying to increase the port until it exceeds the max value(65535):
`typeof(System.InvalidOperationException): Exceeded port range`
|
non_design
|
portresever doesn t work correctly for portreserver has the following code setting timespan zero for a cancellationtokensource in the task will get cancelled so it will never get a chance to run task on this port then it will keep trying to increase the port until it exceeds the max value typeof system invalidoperationexception exceeded port range
| 0
|
218,527
| 24,376,033,267
|
IssuesEvent
|
2022-10-04 01:02:03
|
TIBCOSoftware/jasperreports-server-ce
|
https://api.github.com/repos/TIBCOSoftware/jasperreports-server-ce
|
opened
|
CVE-2022-42004 (Medium) detected in jackson-databind-2.13.2.2.jar, jackson-databind-2.13.2.jar
|
security vulnerability
|
## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.13.2.2.jar</b>, <b>jackson-databind-2.13.2.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.13.2.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /jasperserver/buildomatic/lib/jackson-databind-2.13.2.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.13.2.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.13.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /jasperserver/buildomatic/pom.xml</p>
<p>Path to vulnerable library: /wnloadResource_KSJHRQ/20220802194008/jackson-databind-2.13.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.13.2.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/jasperreports-server-ce/commit/e1b47cd38e2251ab73815346ec28c1fee1b43487">e1b47cd38e2251ab73815346ec28c1fee1b43487</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
True
|
CVE-2022-42004 (Medium) detected in jackson-databind-2.13.2.2.jar, jackson-databind-2.13.2.jar - ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.13.2.2.jar</b>, <b>jackson-databind-2.13.2.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.13.2.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /jasperserver/buildomatic/lib/jackson-databind-2.13.2.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.13.2.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.13.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /jasperserver/buildomatic/pom.xml</p>
<p>Path to vulnerable library: /wnloadResource_KSJHRQ/20220802194008/jackson-databind-2.13.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.13.2.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/jasperreports-server-ce/commit/e1b47cd38e2251ab73815346ec28c1fee1b43487">e1b47cd38e2251ab73815346ec28c1fee1b43487</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
non_design
|
cve medium detected in jackson databind jar jackson databind jar cve medium severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library jasperserver buildomatic lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file jasperserver buildomatic pom xml path to vulnerable library wnloadresource ksjhrq jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in beandeserializer deserializefromarray to prevent use of deeply nested arrays an application is vulnerable only with certain customized choices for deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind check this box to open an automated fix pr
| 0
|
80,725
| 3,573,544,439
|
IssuesEvent
|
2016-01-27 07:12:02
|
antang/NewCapstoneProject
|
https://api.github.com/repos/antang/NewCapstoneProject
|
closed
|
Bug - Second player also get bonus score
|
Bug Medium priority
|
When second player complete all phases then they get bonus score.
|
1.0
|
Bug - Second player also get bonus score - When second player complete all phases then they get bonus score.
|
non_design
|
bug second player also get bonus score when second player complete all phases then they get bonus score
| 0
|
163,195
| 25,772,238,420
|
IssuesEvent
|
2022-12-09 09:01:03
|
PowerShell/PowerShell
|
https://api.github.com/repos/PowerShell/PowerShell
|
closed
|
ConvertFrom-JSON incorrectly deserializes dates to DateTime
|
Issue-Question Resolution-By Design WG-Cmdlets-Utility
|
When `ConvertFrom-JSON` guesses something in the input data can be converted to a `DateTime`, it (silently) does so.
But the conversion doesn't correctly interprets some strings.
## Steps to reproduce
I'm running this on a system that's on UTC+2 (CEST).
```
Write-Host("PowerShell version: " + $PSVersionTable.PSVersion.ToString())
$date = "2020-09-07T09:44:13.769Z"
Write-Host ("Original string: " + $date)
Write-Host("Cast to Datetime: " + [datetime] $date)
$json = ('[{"start":"' + $date + '"}]')
Write-Host("JSON data: " + $json)
$data = $json | ConvertFrom-Json
Write-Host($data[0].start.GetType().Name + " resulting from ConvertFrom-JSON: " + $data[0].start)
```
## Expected behavior
```
PowerShell version: 7.0.3
Original string: 2020-09-07T09:44:13.769Z
Cast to Datetime: 09/07/2020 11:44:13
JSON data: [{"start":"2020-09-07T09:44:13.769Z"}]
DateTime resulting from ConvertFrom-JSON: 09/07/2020 11:44:13
```
## Actual behavior
```
PowerShell version: 7.0.3
Original string: 2020-09-07T09:44:13.769Z
Cast to Datetime: 09/07/2020 11:44:13
JSON data: [{"start":"2020-09-07T09:44:13.769Z"}]
DateTime resulting from ConvertFrom-JSON: 09/07/2020 09:44:13
```
## Environment data
```
Name Value
---- -----
PSVersion 7.0.3
PSEdition Core
GitCommitId 7.0.3
OS Microsoft Windows 10.0.17763
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
The same issue is still present in 7.1.0-preview.6.
|
1.0
|
ConvertFrom-JSON incorrectly deserializes dates to DateTime - When `ConvertFrom-JSON` guesses something in the input data can be converted to a `DateTime`, it (silently) does so.
But the conversion doesn't correctly interprets some strings.
## Steps to reproduce
I'm running this on a system that's on UTC+2 (CEST).
```
Write-Host("PowerShell version: " + $PSVersionTable.PSVersion.ToString())
$date = "2020-09-07T09:44:13.769Z"
Write-Host ("Original string: " + $date)
Write-Host("Cast to Datetime: " + [datetime] $date)
$json = ('[{"start":"' + $date + '"}]')
Write-Host("JSON data: " + $json)
$data = $json | ConvertFrom-Json
Write-Host($data[0].start.GetType().Name + " resulting from ConvertFrom-JSON: " + $data[0].start)
```
## Expected behavior
```
PowerShell version: 7.0.3
Original string: 2020-09-07T09:44:13.769Z
Cast to Datetime: 09/07/2020 11:44:13
JSON data: [{"start":"2020-09-07T09:44:13.769Z"}]
DateTime resulting from ConvertFrom-JSON: 09/07/2020 11:44:13
```
## Actual behavior
```
PowerShell version: 7.0.3
Original string: 2020-09-07T09:44:13.769Z
Cast to Datetime: 09/07/2020 11:44:13
JSON data: [{"start":"2020-09-07T09:44:13.769Z"}]
DateTime resulting from ConvertFrom-JSON: 09/07/2020 09:44:13
```
## Environment data
```
Name Value
---- -----
PSVersion 7.0.3
PSEdition Core
GitCommitId 7.0.3
OS Microsoft Windows 10.0.17763
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
The same issue is still present in 7.1.0-preview.6.
|
design
|
convertfrom json incorrectly deserializes dates to datetime when convertfrom json guesses something in the input data can be converted to a datetime it silently does so but the conversion doesn t correctly interprets some strings steps to reproduce i m running this on a system that s on utc cest write host powershell version psversiontable psversion tostring date write host original string date write host cast to datetime date json write host json data json data json convertfrom json write host data start gettype name resulting from convertfrom json data start expected behavior powershell version original string cast to datetime json data datetime resulting from convertfrom json actual behavior powershell version original string cast to datetime json data datetime resulting from convertfrom json environment data name value psversion psedition core gitcommitid os microsoft windows platform pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion the same issue is still present in preview
| 1
|
174,998
| 27,770,976,489
|
IssuesEvent
|
2023-03-16 14:26:55
|
department-of-veterans-affairs/vets-design-system-documentation
|
https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation
|
closed
|
Conversion to USWDS v3 - Button component - Development
|
vsp-design-system-team va-button dst-engineering
|
## Description
Adapt the va-button component to match the [USWDS Button](https://designsystem.digital.gov/components/button/) component and import and use the USWDS CSS for this component.
If this is a pattern or component that is already in existence, check the Design System Team backlog for outstanding issues. If you find any, link to them in a comment on this ticket. If possible, address any outstanding issues with this new version and link to this issue from the original issue. If not, indicate that in the original issue.
## Details
Design documents: [add links to any design documents]
See the [Textarea discovery](https://github.com/department-of-veterans-affairs/component-library/pull/493) for an example
According to the [plan](https://vfs.atlassian.net/wiki/spaces/DST/pages/2366275704/Plan+-+Aligning+VA+Design+System+components+with+their+USWDS+v3+equivalents), we will be using a prop if necessary to differentiate between our version of this component and the USWDS version. It may not be necessary for this component, this can be decided on a case-by-case basis
[USWDS Button component](https://designsystem.digital.gov/components/button/)
[VA DS Button component](https://design.va.gov/components/button/)
## Tasks
- [ ] Create a version of this component that imports and uses the USWDS CSS
- [ ] Make any structural changes necessary and add a prop that determines which version of the component will be used
- [ ] Check with Natalie to make sure the component and the design match
- [ ] Add a USWDS story to Storybook
- [ ] Merge component
- [ ] Create a new release of component-library
- [ ] Update component-library dependency in vets-design-system-documentation to get the updated component-docs.json
## Acceptance Criteria
- [ ] USWDS version has been created and added to Storybook
- [ ] Design.va.gov has the latest version of component-library
|
1.0
|
Conversion to USWDS v3 - Button component - Development - ## Description
Adapt the va-button component to match the [USWDS Button](https://designsystem.digital.gov/components/button/) component and import and use the USWDS CSS for this component.
If this is a pattern or component that is already in existence, check the Design System Team backlog for outstanding issues. If you find any, link to them in a comment on this ticket. If possible, address any outstanding issues with this new version and link to this issue from the original issue. If not, indicate that in the original issue.
## Details
Design documents: [add links to any design documents]
See the [Textarea discovery](https://github.com/department-of-veterans-affairs/component-library/pull/493) for an example
According to the [plan](https://vfs.atlassian.net/wiki/spaces/DST/pages/2366275704/Plan+-+Aligning+VA+Design+System+components+with+their+USWDS+v3+equivalents), we will be using a prop if necessary to differentiate between our version of this component and the USWDS version. It may not be necessary for this component, this can be decided on a case-by-case basis
[USWDS Button component](https://designsystem.digital.gov/components/button/)
[VA DS Button component](https://design.va.gov/components/button/)
## Tasks
- [ ] Create a version of this component that imports and uses the USWDS CSS
- [ ] Make any structural changes necessary and add a prop that determines which version of the component will be used
- [ ] Check with Natalie to make sure the component and the design match
- [ ] Add a USWDS story to Storybook
- [ ] Merge component
- [ ] Create a new release of component-library
- [ ] Update component-library dependency in vets-design-system-documentation to get the updated component-docs.json
## Acceptance Criteria
- [ ] USWDS version has been created and added to Storybook
- [ ] Design.va.gov has the latest version of component-library
|
design
|
conversion to uswds button component development description adapt the va button component to match the component and import and use the uswds css for this component if this is a pattern or component that is already in existence check the design system team backlog for outstanding issues if you find any link to them in a comment on this ticket if possible address any outstanding issues with this new version and link to this issue from the original issue if not indicate that in the original issue details design documents see the for an example according to the we will be using a prop if necessary to differentiate between our version of this component and the uswds version it may not be necessary for this component this can be decided on a case by case basis tasks create a version of this component that imports and uses the uswds css make any structural changes necessary and add a prop that determines which version of the component will be used check with natalie to make sure the component and the design match add a uswds story to storybook merge component create a new release of component library update component library dependency in vets design system documentation to get the updated component docs json acceptance criteria uswds version has been created and added to storybook design va gov has the latest version of component library
| 1
|
274,971
| 20,885,956,349
|
IssuesEvent
|
2022-03-23 05:11:27
|
lucasdesouza2/Desenvolvimento-Java
|
https://api.github.com/repos/lucasdesouza2/Desenvolvimento-Java
|
closed
|
Pesquisar sobre Kanban
|
documentation
|
> Adicionar uma descrição na própria issue sobre o entendimento obtido do assunto.
|
1.0
|
Pesquisar sobre Kanban - > Adicionar uma descrição na própria issue sobre o entendimento obtido do assunto.
|
non_design
|
pesquisar sobre kanban adicionar uma descrição na própria issue sobre o entendimento obtido do assunto
| 0
|
95,414
| 11,991,666,785
|
IssuesEvent
|
2020-04-08 08:46:20
|
wellcomecollection/wellcomecollection.org
|
https://api.github.com/repos/wellcomecollection/wellcomecollection.org
|
closed
|
Works & Viewer Page component annotations
|
feature needs:design
|
# What is it and who's it for?
Annotated versions of the Works page and Viewer page under Collections for the digital team
# User story
In order to _(why? accomplish a goal/get a benefit)_
understand what each component/page is under Collections
As a _(who? user/persona)_
digital team member
I need/want/expect _(what? to do this thing)_
I need to know what the common language/terminology is for the different components and define what the purpose of each thing is
## Acceptance criteria
# Implementation
## Testing and tracking
- Have you talked to Tacey?
- Test plan
- Will you need to set up additional tracking?
- What is the expected life span for tracking?
## Measurement
- Related KPIs (e.g. breadth of collection accessed, reading time, etc.)
## Other considerations
<---Delete as appropriate--->
- Impact on URL design
- Cardigan documentation
|
1.0
|
Works & Viewer Page component annotations - # What is it and who's it for?
Annotated versions of the Works page and Viewer page under Collections for the digital team
# User story
In order to _(why? accomplish a goal/get a benefit)_
understand what each component/page is under Collections
As a _(who? user/persona)_
digital team member
I need/want/expect _(what? to do this thing)_
I need to know what the common language/terminology is for the different components and define what the purpose of each thing is
## Acceptance criteria
# Implementation
## Testing and tracking
- Have you talked to Tacey?
- Test plan
- Will you need to set up additional tracking?
- What is the expected life span for tracking?
## Measurement
- Related KPIs (e.g. breadth of collection accessed, reading time, etc.)
## Other considerations
<---Delete as appropriate--->
- Impact on URL design
- Cardigan documentation
|
design
|
works viewer page component annotations what is it and who s it for annotated versions of the works page and viewer page under collections for the digital team user story in order to why accomplish a goal get a benefit understand what each component page is under collections as a who user persona digital team member i need want expect what to do this thing i need to know what the common language terminology is for the different components and define what the purpose of each thing is acceptance criteria implementation testing and tracking have you talked to tacey test plan will you need to set up additional tracking what is the expected life span for tracking measurement related kpis e g breadth of collection accessed reading time etc other considerations impact on url design cardigan documentation
| 1
|
74,242
| 9,007,880,356
|
IssuesEvent
|
2019-02-05 00:54:55
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
BottomNavigationbar Logic TextField is not touchable
|
a: text input f: material design framework
|
Steps To Reproduce
```dart
class HomePage extends StatefulWidget {
@override
_HomePageState createState() => _HomePageState();
}
class _HomePageState extends State<HomePage> {
final int _pageCount = 2;
int _pageIndex = 0;
@override
Widget build(BuildContext context) {
return Scaffold(
body: _body(),
bottomNavigationBar: _bottomNavigationBar(),
);
}
Widget _body() {
return Stack(
children: List<Widget>.generate(_pageCount, (int index) {
return IgnorePointer(
ignoring: index != _pageIndex,
child: Opacity(
opacity: _pageIndex == index ? 1.0 : 0.0,
child: Navigator(
onGenerateRoute: (RouteSettings settings) {
return new MaterialPageRoute(
builder: (_) => _page(index),
settings: settings,
);
},
),
),
);
}),
);
}
Widget _page(int index) {
switch (index) {
case 0:
return Page1();
case 1:
return Page2();
}
throw "Invalid index $index";
}
BottomNavigationBar _bottomNavigationBar() {
final theme = Theme.of(context);
return new BottomNavigationBar(
fixedColor: theme.accentColor,
currentIndex: _pageIndex,
items: [
BottomNavigationBarItem(
icon: Icon(Icons.list),
title: Text("Page 1"),
),
BottomNavigationBarItem(
icon: Icon(Icons.account_circle),
title: Text("Page 2"),
),
],
onTap: (int index) {
setState(() {
_pageIndex = index;
});
},
);
}
}
```
In the above code If one Of the Page has a `TextInputfield` And i am currently on the active page, but i am not able to tap the text input all other buttons are working, the keyboard doesn't show up, or i don't thinks it's being clicked.
In the attached i am not able to tap on write message text box, i have moved its position also the issue still persist.
<img width="267" alt="screen shot 2019-01-20 at 1 37 04 am" src="https://user-images.githubusercontent.com/1706515/51431777-f6345980-1c53-11e9-8c31-bb6d0ed92091.png">
|
1.0
|
BottomNavigationbar Logic TextField is not touchable - Steps To Reproduce
```dart
class HomePage extends StatefulWidget {
@override
_HomePageState createState() => _HomePageState();
}
class _HomePageState extends State<HomePage> {
final int _pageCount = 2;
int _pageIndex = 0;
@override
Widget build(BuildContext context) {
return Scaffold(
body: _body(),
bottomNavigationBar: _bottomNavigationBar(),
);
}
Widget _body() {
return Stack(
children: List<Widget>.generate(_pageCount, (int index) {
return IgnorePointer(
ignoring: index != _pageIndex,
child: Opacity(
opacity: _pageIndex == index ? 1.0 : 0.0,
child: Navigator(
onGenerateRoute: (RouteSettings settings) {
return new MaterialPageRoute(
builder: (_) => _page(index),
settings: settings,
);
},
),
),
);
}),
);
}
Widget _page(int index) {
switch (index) {
case 0:
return Page1();
case 1:
return Page2();
}
throw "Invalid index $index";
}
BottomNavigationBar _bottomNavigationBar() {
final theme = Theme.of(context);
return new BottomNavigationBar(
fixedColor: theme.accentColor,
currentIndex: _pageIndex,
items: [
BottomNavigationBarItem(
icon: Icon(Icons.list),
title: Text("Page 1"),
),
BottomNavigationBarItem(
icon: Icon(Icons.account_circle),
title: Text("Page 2"),
),
],
onTap: (int index) {
setState(() {
_pageIndex = index;
});
},
);
}
}
```
In the above code If one Of the Page has a `TextInputfield` And i am currently on the active page, but i am not able to tap the text input all other buttons are working, the keyboard doesn't show up, or i don't thinks it's being clicked.
In the attached i am not able to tap on write message text box, i have moved its position also the issue still persist.
<img width="267" alt="screen shot 2019-01-20 at 1 37 04 am" src="https://user-images.githubusercontent.com/1706515/51431777-f6345980-1c53-11e9-8c31-bb6d0ed92091.png">
|
design
|
bottomnavigationbar logic textfield is not touchable steps to reproduce dart class homepage extends statefulwidget override homepagestate createstate homepagestate class homepagestate extends state final int pagecount int pageindex override widget build buildcontext context return scaffold body body bottomnavigationbar bottomnavigationbar widget body return stack children list generate pagecount int index return ignorepointer ignoring index pageindex child opacity opacity pageindex index child navigator ongenerateroute routesettings settings return new materialpageroute builder page index settings settings widget page int index switch index case return case return throw invalid index index bottomnavigationbar bottomnavigationbar final theme theme of context return new bottomnavigationbar fixedcolor theme accentcolor currentindex pageindex items bottomnavigationbaritem icon icon icons list title text page bottomnavigationbaritem icon icon icons account circle title text page ontap int index setstate pageindex index in the above code if one of the page has a textinputfield and i am currently on the active page but i am not able to tap the text input all other buttons are working the keyboard doesn t show up or i don t thinks it s being clicked in the attached i am not able to tap on write message text box i have moved its position also the issue still persist img width alt screen shot at am src
| 1
|
64,867
| 7,846,612,154
|
IssuesEvent
|
2018-06-19 15:56:17
|
skycoin/skycoin-web
|
https://api.github.com/repos/skycoin/skycoin-web
|
closed
|
Indicate the status of the secondary pages
|
UI/UX design
|
Currently, when there are no transactions, the History page simply shows the gray background and a slightly darker line. It would be more convenient to show a component with a message for the user like this:

Additionally, the same component could be used while the information is being loaded, it would only be necessary to replace the text and the icon (with the spinner). In this way the user would have more information about what the wallet is doing, which would improve the UX.
The idea would be to use the component not only on the History page, but also on the Blockchain, Outputs and Pending Transactions pages. To be able to add it correctly in the Blockchain page, it would be necessary to wait for `blockchainService.lastBlock` and `blockchainService.coinSupply` to finish, before displaying the information on the screen.
The icon used in the image is https://material.io/tools/icons/?icon=announcement&style=baseline, its color is `#eaeaea` and have a size of about 50x50 px. The The color of the text is `#a9aaac`. The content has a margin of 50px with respect to the navigation bar (a similar margin should be used at the bottom).
|
1.0
|
Indicate the status of the secondary pages - Currently, when there are no transactions, the History page simply shows the gray background and a slightly darker line. It would be more convenient to show a component with a message for the user like this:

Additionally, the same component could be used while the information is being loaded, it would only be necessary to replace the text and the icon (with the spinner). In this way the user would have more information about what the wallet is doing, which would improve the UX.
The idea would be to use the component not only on the History page, but also on the Blockchain, Outputs and Pending Transactions pages. To be able to add it correctly in the Blockchain page, it would be necessary to wait for `blockchainService.lastBlock` and `blockchainService.coinSupply` to finish, before displaying the information on the screen.
The icon used in the image is https://material.io/tools/icons/?icon=announcement&style=baseline, its color is `#eaeaea` and have a size of about 50x50 px. The The color of the text is `#a9aaac`. The content has a margin of 50px with respect to the navigation bar (a similar margin should be used at the bottom).
|
design
|
indicate the status of the secondary pages currently when there are no transactions the history page simply shows the gray background and a slightly darker line it would be more convenient to show a component with a message for the user like this additionally the same component could be used while the information is being loaded it would only be necessary to replace the text and the icon with the spinner in this way the user would have more information about what the wallet is doing which would improve the ux the idea would be to use the component not only on the history page but also on the blockchain outputs and pending transactions pages to be able to add it correctly in the blockchain page it would be necessary to wait for blockchainservice lastblock and blockchainservice coinsupply to finish before displaying the information on the screen the icon used in the image is its color is eaeaea and have a size of about px the the color of the text is the content has a margin of with respect to the navigation bar a similar margin should be used at the bottom
| 1
|
133,991
| 19,012,854,634
|
IssuesEvent
|
2021-11-23 11:13:30
|
ZcashFoundation/zebra
|
https://api.github.com/repos/ZcashFoundation/zebra
|
closed
|
Tracking: block header validation
|
C-design A-consensus C-tracking-issue Epic
|
As part of block validation (#428), we need to validate all the fields in the block header:
**Parsing**
- [x] version:
- the block version number MUST be greater than or equal to 4.
- [x] solution_size:
- fixed value of 1344.
- validate during parsing (#478)
- the solution size field is not present in zebra-chain's parsed Block struct.
- Like other serialized fields of type compactSize uint, the solutionSize field MUST be encoded with the minimum number of bytes (3 in this case), and other encodings MUST be rejected. (This check is already performed in Zebra's compactsize code.)
- [x] block size:
- the size of a block MUST be less than or equal to 2000000 bytes.
- validate during parsing (#485).
- also limits the size of transactions in parsed blocks.
- the block size isn't encoded in the serialised data.
- the block size field is not present in zebra-chain's parsed Block struct.
**Semantic Validation**
Before validating the rest of the block:
- [x] time (node clock): (PR #492)
- a full validator MUST NOT accept blocks with nTime more than two hours in the future according to its clock. This is not strictly a consensus rule because it is nondeterministic, and clock time varies between nodes. Also note that a block that is rejected by this rule at a given point in time may later be accepted.
- as an optimisation, we could await future times that will happen soon.
- but for now, let's reject and drop the block.
- [x] height (semantic):
- the block height MUST be encoded as the first item in the coinbase transaction’s scriptSig, as specified in [BIP-34]. The format of the height is “serialized CScript” – the first byte is the number of bytes in the number, and the following bytes are the signed little-endian representation of the number.
- [x] solution:
- must represent a valid Equihash solution.
- see #578, #550, #607.
- [x] difficulty_threshold: encoded number of difficulty bits (#572, #799, #802)
- decode from Bitcoin's "compact u256" format (#799)
- block header hash: the block header hash is interpreted as a 256-bit integer represented in little-endian byte order, which MUST be less than or equal to the target threshold given by ToTarget(bits).
- [x] nonce: arbitrary value.
- implicitly validated as part of the `difficulty_threshold` hash check.
- implicitly validated as part of each child block's `previous_block_hash` check
Before validating transactions:
- [x] merkle_root_hash: #906
- the merkle root is derived from the hashes of all transactions included in this block, ensuring that none of those transactions can be modified without modifying the header.
- verify this field before validating any transactions
**Contextual Validation**
Depends on previous block - needed for chain reorganisation:
- [x] previous_block_hash: #737
- must be equal to a previous block's header hash.
- validate after the previous block is validated.
- the previous_block_hash for the genesis block is all zeroes (Bitcoin `null`). We should add this constant to the consensus parameters.
- [x] height (contextual): #737
- Each block in a block chain has a block height. The block height of the genesis block is 0, and the block height of each subsequent block in the block chain increments by 1.
At this point, we are doing enough validation to perform chaining. So we want to implement chain reorganisation (#452) here, because it affects the rest of the design. (We might also want to implement checkpointing (#429), but it's a lot simpler, because it's just an alternate validation path.)
Depends on a range of earlier blocks:
- [x] time (previous block times): #1413
- for each block other than the genesis block , nTime MUST be strictly greater than the median-time-past of that block.
- for each block at block height 2 or greater on the production network, or block height 653606 or greater on the test network, nTime MUST be less than or equal to the median-time-past of that block plus 90 · 60 seconds.
- validate after the PoWMedianBlockSpan previous blocks are validated.
- check for overflow when summing times from multiple blocks.
- [x] difficulty_threshold (threshold value): (#802)
- must be equal to ThresholdBits(height).
- validate after the `max(PoWMedianBlockSpan, PoWAveragingWindow)` previous blocks are validated.
- check for overflow when summing bits from multiple blocks.
- [x] light_client_hash (#856, #881)
**Open Questions**
- [x] where does zebra-consensus store blocks, while it is waiting for the previous blocks?
- see https://github.com/ZcashFoundation/zebra/issues/477#issuecomment-646350910
- [ ] TODO: Other rules inherited from Bitcoin.
|
1.0
|
Tracking: block header validation - As part of block validation (#428), we need to validate all the fields in the block header:
**Parsing**
- [x] version:
- the block version number MUST be greater than or equal to 4.
- [x] solution_size:
- fixed value of 1344.
- validate during parsing (#478)
- the solution size field is not present in zebra-chain's parsed Block struct.
- Like other serialized fields of type compactSize uint, the solutionSize field MUST be encoded with the minimum number of bytes (3 in this case), and other encodings MUST be rejected. (This check is already performed in Zebra's compactsize code.)
- [x] block size:
- the size of a block MUST be less than or equal to 2000000 bytes.
- validate during parsing (#485).
- also limits the size of transactions in parsed blocks.
- the block size isn't encoded in the serialised data.
- the block size field is not present in zebra-chain's parsed Block struct.
**Semantic Validation**
Before validating the rest of the block:
- [x] time (node clock): (PR #492)
- a full validator MUST NOT accept blocks with nTime more than two hours in the future according to its clock. This is not strictly a consensus rule because it is nondeterministic, and clock time varies between nodes. Also note that a block that is rejected by this rule at a given point in time may later be accepted.
- as an optimisation, we could await future times that will happen soon.
- but for now, let's reject and drop the block.
- [x] height (semantic):
- the block height MUST be encoded as the first item in the coinbase transaction’s scriptSig, as specified in [BIP-34]. The format of the height is “serialized CScript” – the first byte is the number of bytes in the number, and the following bytes are the signed little-endian representation of the number.
- [x] solution:
- must represent a valid Equihash solution.
- see #578, #550, #607.
- [x] difficulty_threshold: encoded number of difficulty bits (#572, #799, #802)
- decode from Bitcoin's "compact u256" format (#799)
- block header hash: the block header hash is interpreted as a 256-bit integer represented in little-endian byte order, which MUST be less than or equal to the target threshold given by ToTarget(bits).
- [x] nonce: arbitrary value.
- implicitly validated as part of the `difficulty_threshold` hash check.
- implicitly validated as part of each child block's `previous_block_hash` check
Before validating transactions:
- [x] merkle_root_hash: #906
- the merkle root is derived from the hashes of all transactions included in this block, ensuring that none of those transactions can be modified without modifying the header.
- verify this field before validating any transactions
**Contextual Validation**
Depends on previous block - needed for chain reorganisation:
- [x] previous_block_hash: #737
- must be equal to a previous block's header hash.
- validate after the previous block is validated.
- the previous_block_hash for the genesis block is all zeroes (Bitcoin `null`). We should add this constant to the consensus parameters.
- [x] height (contextual): #737
- Each block in a block chain has a block height. The block height of the genesis block is 0, and the block height of each subsequent block in the block chain increments by 1.
At this point, we are doing enough validation to perform chaining. So we want to implement chain reorganisation (#452) here, because it affects the rest of the design. (We might also want to implement checkpointing (#429), but it's a lot simpler, because it's just an alternate validation path.)
Depends on a range of earlier blocks:
- [x] time (previous block times): #1413
- for each block other than the genesis block , nTime MUST be strictly greater than the median-time-past of that block.
- for each block at block height 2 or greater on the production network, or block height 653606 or greater on the test network, nTime MUST be less than or equal to the median-time-past of that block plus 90 · 60 seconds.
- validate after the PoWMedianBlockSpan previous blocks are validated.
- check for overflow when summing times from multiple blocks.
- [x] difficulty_threshold (threshold value): (#802)
- must be equal to ThresholdBits(height).
- validate after the `max(PoWMedianBlockSpan, PoWAveragingWindow)` previous blocks are validated.
- check for overflow when summing bits from multiple blocks.
- [x] light_client_hash (#856, #881)
**Open Questions**
- [x] where does zebra-consensus store blocks, while it is waiting for the previous blocks?
- see https://github.com/ZcashFoundation/zebra/issues/477#issuecomment-646350910
- [ ] TODO: Other rules inherited from Bitcoin.
|
design
|
tracking block header validation as part of block validation we need to validate all the fields in the block header parsing version the block version number must be greater than or equal to solution size fixed value of validate during parsing the solution size field is not present in zebra chain s parsed block struct like other serialized fields of type compactsize uint the solutionsize field must be encoded with the minimum number of bytes in this case and other encodings must be rejected this check is already performed in zebra s compactsize code block size the size of a block must be less than or equal to bytes validate during parsing also limits the size of transactions in parsed blocks the block size isn t encoded in the serialised data the block size field is not present in zebra chain s parsed block struct semantic validation before validating the rest of the block time node clock pr a full validator must not accept blocks with ntime more than two hours in the future according to its clock this is not strictly a consensus rule because it is nondeterministic and clock time varies between nodes also note that a block that is rejected by this rule at a given point in time may later be accepted as an optimisation we could await future times that will happen soon but for now let s reject and drop the block height semantic the block height must be encoded as the first item in the coinbase transaction’s scriptsig as specified in the format of the height is “serialized cscript” – the first byte is the number of bytes in the number and the following bytes are the signed little endian representation of the number solution must represent a valid equihash solution see difficulty threshold encoded number of difficulty bits decode from bitcoin s compact format block header hash the block header hash is interpreted as a bit integer represented in little endian byte order which must be less than or equal to the target threshold given by totarget bits nonce arbitrary value implicitly validated as part of the difficulty threshold hash check implicitly validated as part of each child block s previous block hash check before validating transactions merkle root hash the merkle root is derived from the hashes of all transactions included in this block ensuring that none of those transactions can be modified without modifying the header verify this field before validating any transactions contextual validation depends on previous block needed for chain reorganisation previous block hash must be equal to a previous block s header hash validate after the previous block is validated the previous block hash for the genesis block is all zeroes bitcoin null we should add this constant to the consensus parameters height contextual each block in a block chain has a block height the block height of the genesis block is and the block height of each subsequent block in the block chain increments by at this point we are doing enough validation to perform chaining so we want to implement chain reorganisation here because it affects the rest of the design we might also want to implement checkpointing but it s a lot simpler because it s just an alternate validation path depends on a range of earlier blocks time previous block times for each block other than the genesis block ntime must be strictly greater than the median time past of that block for each block at block height or greater on the production network or block height or greater on the test network ntime must be less than or equal to the median time past of that block plus · seconds validate after the powmedianblockspan previous blocks are validated check for overflow when summing times from multiple blocks difficulty threshold threshold value must be equal to thresholdbits height validate after the max powmedianblockspan powaveragingwindow previous blocks are validated check for overflow when summing bits from multiple blocks light client hash open questions where does zebra consensus store blocks while it is waiting for the previous blocks see todo other rules inherited from bitcoin
| 1
|
118,437
| 15,295,895,918
|
IssuesEvent
|
2021-02-24 05:50:25
|
oshi/oshi
|
https://api.github.com/repos/oshi/oshi
|
opened
|
Add ProcessHandle methods to OSProcess in Java11 branch
|
design discussion new feature performance
|
JDK9 included a new [ProcessHandle class](https://docs.oracle.com/en/java/javase/11/docs/api/java.base/java/lang/ProcessHandle.html) which we can leverage to improve the `OSProcess` class. It provides the opportunity to be a bit more performant internally and expose a few more methods.
Here's some of my thoughts. Looking for community feedback.
- Use the `allProcesses()` and `children()` methods to improve performance of existing code
- Add a `getProcessHandle()` method
- Add a new `getDescendantProcesses()` method similar to `getChildProcesses()` (using the `descendants()` method including the children's children, etc. (This could be done in the main API too, less efficiently).
Additional thought:
- The PID should be a `long` to match. In the API I'll need to add a newly named method (`getPid()`?) for `getProcessId()` and deprecate the old one.
- Unfortunately the "Info" doesn't give more than we currently have (and often less). I'll look into whether it can make things more efficient, but don't want to prematurely optimize.
Alternately: do nothing, let users do `ProcessHandle.of(pid)` themselves. :)
|
1.0
|
Add ProcessHandle methods to OSProcess in Java11 branch - JDK9 included a new [ProcessHandle class](https://docs.oracle.com/en/java/javase/11/docs/api/java.base/java/lang/ProcessHandle.html) which we can leverage to improve the `OSProcess` class. It provides the opportunity to be a bit more performant internally and expose a few more methods.
Here's some of my thoughts. Looking for community feedback.
- Use the `allProcesses()` and `children()` methods to improve performance of existing code
- Add a `getProcessHandle()` method
- Add a new `getDescendantProcesses()` method similar to `getChildProcesses()` (using the `descendants()` method including the children's children, etc. (This could be done in the main API too, less efficiently).
Additional thought:
- The PID should be a `long` to match. In the API I'll need to add a newly named method (`getPid()`?) for `getProcessId()` and deprecate the old one.
- Unfortunately the "Info" doesn't give more than we currently have (and often less). I'll look into whether it can make things more efficient, but don't want to prematurely optimize.
Alternately: do nothing, let users do `ProcessHandle.of(pid)` themselves. :)
|
design
|
add processhandle methods to osprocess in branch included a new which we can leverage to improve the osprocess class it provides the opportunity to be a bit more performant internally and expose a few more methods here s some of my thoughts looking for community feedback use the allprocesses and children methods to improve performance of existing code add a getprocesshandle method add a new getdescendantprocesses method similar to getchildprocesses using the descendants method including the children s children etc this could be done in the main api too less efficiently additional thought the pid should be a long to match in the api i ll need to add a newly named method getpid for getprocessid and deprecate the old one unfortunately the info doesn t give more than we currently have and often less i ll look into whether it can make things more efficient but don t want to prematurely optimize alternately do nothing let users do processhandle of pid themselves
| 1
|
151,918
| 23,891,356,771
|
IssuesEvent
|
2022-09-08 11:45:12
|
gruntwork-io/cloud-nuke
|
https://api.github.com/repos/gruntwork-io/cloud-nuke
|
closed
|
New feature: ability to delete IAM roles
|
enhancement needs design p:backlog
|
To assist with hitting the IAM roles limit in AWS - we could use the ability to delete IAM roles - but to also provide a list of roles that should be preserved and not deleted.
|
1.0
|
New feature: ability to delete IAM roles - To assist with hitting the IAM roles limit in AWS - we could use the ability to delete IAM roles - but to also provide a list of roles that should be preserved and not deleted.
|
design
|
new feature ability to delete iam roles to assist with hitting the iam roles limit in aws we could use the ability to delete iam roles but to also provide a list of roles that should be preserved and not deleted
| 1
|
100,697
| 12,547,014,834
|
IssuesEvent
|
2020-06-05 21:52:29
|
ubyssey/ubyssey.ca
|
https://api.github.com/repos/ubyssey/ubyssey.ca
|
closed
|
Web Team sticker
|
3 design
|
Would be amazing if we can get our own custom laptop sticker for web team.
For example, [rbnb project](https://twitter.com/robinson_es/status/915220688717455360) at Airbnb have their own sticker with consistant brand design language, or even [something like Supreme](https://dribbble.com/shots/3688897-Shopify-Stickers)

Or the company called Mercari have their python team sticker

|
1.0
|
Web Team sticker - Would be amazing if we can get our own custom laptop sticker for web team.
For example, [rbnb project](https://twitter.com/robinson_es/status/915220688717455360) at Airbnb have their own sticker with consistant brand design language, or even [something like Supreme](https://dribbble.com/shots/3688897-Shopify-Stickers)

Or the company called Mercari have their python team sticker

|
design
|
web team sticker would be amazing if we can get our own custom laptop sticker for web team for example at airbnb have their own sticker with consistant brand design language or even or the company called mercari have their python team sticker
| 1
|
31,531
| 2,733,664,925
|
IssuesEvent
|
2015-04-17 15:13:25
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
opened
|
Setup prometheus server and dashboard
|
area/performance priority/P2 team/cluster
|
@wojtek-t @mwielgus
It'd be nice to setup prometheus server so that we can query/graph metrics that we export. This doesn't block other work, but would make it easier to debug performance issues.
|
1.0
|
Setup prometheus server and dashboard - @wojtek-t @mwielgus
It'd be nice to setup prometheus server so that we can query/graph metrics that we export. This doesn't block other work, but would make it easier to debug performance issues.
|
non_design
|
setup prometheus server and dashboard wojtek t mwielgus it d be nice to setup prometheus server so that we can query graph metrics that we export this doesn t block other work but would make it easier to debug performance issues
| 0
|
232
| 2,684,171,072
|
IssuesEvent
|
2015-03-28 18:32:51
|
cnb/News_Manager-getsimplecms
|
https://api.github.com/repos/cnb/News_Manager-getsimplecms
|
closed
|
fix display of private post checkbox/label (edit options)
|
compatibility
|
Due to some changes in GS 3.1 (css, javascript) the private post checkbox (in edit post, options) is not aligned properly, and the label doesn't turn red if the checkbox is ticked.
GS 3.0:

GS 3.1+ :

|
True
|
fix display of private post checkbox/label (edit options) - Due to some changes in GS 3.1 (css, javascript) the private post checkbox (in edit post, options) is not aligned properly, and the label doesn't turn red if the checkbox is ticked.
GS 3.0:

GS 3.1+ :

|
non_design
|
fix display of private post checkbox label edit options due to some changes in gs css javascript the private post checkbox in edit post options is not aligned properly and the label doesn t turn red if the checkbox is ticked gs gs
| 0
|
588,145
| 17,648,829,670
|
IssuesEvent
|
2021-08-20 10:11:51
|
status-im/StatusQ
|
https://api.github.com/repos/status-im/StatusQ
|
opened
|
Support dragging and dropping chat items within chat lists
|
type: feature module: components priority 2: required
|
This is actually the first step to make #227 a thing.
Also already implemented in #339, this issue is primarily for tracking purposes.
|
1.0
|
Support dragging and dropping chat items within chat lists - This is actually the first step to make #227 a thing.
Also already implemented in #339, this issue is primarily for tracking purposes.
|
non_design
|
support dragging and dropping chat items within chat lists this is actually the first step to make a thing also already implemented in this issue is primarily for tracking purposes
| 0
|
175,059
| 27,782,163,927
|
IssuesEvent
|
2023-03-16 22:10:09
|
eiksch/statev_v2_issues
|
https://api.github.com/repos/eiksch/statev_v2_issues
|
closed
|
Antrag auf Garage (Mirrorpark)
|
gamedesign solved
|
Ronny Roberts & Jim Hill
Sehr geehrtes StateV-Team,
wäre es möglich beim Haus "house_Vinewood_9" eine Hausgarage hinzuzufügen? (Bild im Anhang)
Unter dem "Carport" wäre es recht passend, leider nur ein wenig Müll im weg.
Der Weg von der Straße würde auch perfekt hinführen mit dem runtergelassenen Bordstein.
Vielen Dank für eure Überprüfung
Freundliche Grüße und Schönen Abend
Pregon
P.S Der Müll unter dem "Carport" kann auch gern weg wenn das geht ^^
Bild:

|
1.0
|
Antrag auf Garage (Mirrorpark) - Ronny Roberts & Jim Hill
Sehr geehrtes StateV-Team,
wäre es möglich beim Haus "house_Vinewood_9" eine Hausgarage hinzuzufügen? (Bild im Anhang)
Unter dem "Carport" wäre es recht passend, leider nur ein wenig Müll im weg.
Der Weg von der Straße würde auch perfekt hinführen mit dem runtergelassenen Bordstein.
Vielen Dank für eure Überprüfung
Freundliche Grüße und Schönen Abend
Pregon
P.S Der Müll unter dem "Carport" kann auch gern weg wenn das geht ^^
Bild:

|
design
|
antrag auf garage mirrorpark ronny roberts jim hill sehr geehrtes statev team wäre es möglich beim haus house vinewood eine hausgarage hinzuzufügen bild im anhang unter dem carport wäre es recht passend leider nur ein wenig müll im weg der weg von der straße würde auch perfekt hinführen mit dem runtergelassenen bordstein vielen dank für eure überprüfung freundliche grüße und schönen abend pregon p s der müll unter dem carport kann auch gern weg wenn das geht bild
| 1
|
165,441
| 26,171,446,133
|
IssuesEvent
|
2023-01-02 00:29:05
|
authelia/authelia
|
https://api.github.com/repos/authelia/authelia
|
opened
|
IndieAuth Identity Provider
|
priority/4/normal type/feature status/needs-design
|
### Description
Much like #189 I'd love to see IndieAuth as a first-class identity provider in Authelia. This would allow me to use Authelia in quite a number of (growing) places that are supporting IndieAuth.
If anyone is interested in this as well, I _could_ help implement this by borrowing from [yarn](https://git.mills.io/yarnsocial/yarn)'s implementation.
### Use Case
Numerous, but primarily to use Authelia itself as an identity provider for any supported website or service that supports IndieAuth.
### Details
A user wising to authenticate themselves to a website or web app (or anything else really) that supports [IndieAuth](https://indieauth.com/), would enter the address/url of their Authelia instance. For example as a user I would enter `auth.mydomain.tld`. If I'm not already authenticated to Authelia, I will be asked to sign-in with my usual Authelia credentials. Once signed in, Authelia will ask me whether I want to approve the login request for the website/app. Once approved authorisation tokens are generated and I'm then signed in with my Authelia identity. (See below for specific details on the flow and implementation).
### Documentation
See [IndieAuth](https://indieauth.com/) and on the [IndieWeb Wiki: IndieAuth](https://indieweb.org/IndieAuth)
See also [building a provider](https://indieweb.org/authorization-endpoint). There are a [handful of Go libraries](https://github.com/search?q=language%3AGo+indieauth), however I am planning on separating out the ones used in [Yarn.social](https://yarn.social)'s backend [yarnd](https://git.mills.io/yarnsocial/yarn) -- I'll post links here once I've done that (working on client side at the moment)
### Pre-Submission Checklist
- [X] I agree to follow the [Code of Conduct](http://www.authelia.com/code-of-conduct)
- [X] I have checked for related issues and checked the documentation
|
1.0
|
IndieAuth Identity Provider - ### Description
Much like #189 I'd love to see IndieAuth as a first-class identity provider in Authelia. This would allow me to use Authelia in quite a number of (growing) places that are supporting IndieAuth.
If anyone is interested in this as well, I _could_ help implement this by borrowing from [yarn](https://git.mills.io/yarnsocial/yarn)'s implementation.
### Use Case
Numerous, but primarily to use Authelia itself as an identity provider for any supported website or service that supports IndieAuth.
### Details
A user wising to authenticate themselves to a website or web app (or anything else really) that supports [IndieAuth](https://indieauth.com/), would enter the address/url of their Authelia instance. For example as a user I would enter `auth.mydomain.tld`. If I'm not already authenticated to Authelia, I will be asked to sign-in with my usual Authelia credentials. Once signed in, Authelia will ask me whether I want to approve the login request for the website/app. Once approved authorisation tokens are generated and I'm then signed in with my Authelia identity. (See below for specific details on the flow and implementation).
### Documentation
See [IndieAuth](https://indieauth.com/) and on the [IndieWeb Wiki: IndieAuth](https://indieweb.org/IndieAuth)
See also [building a provider](https://indieweb.org/authorization-endpoint). There are a [handful of Go libraries](https://github.com/search?q=language%3AGo+indieauth), however I am planning on separating out the ones used in [Yarn.social](https://yarn.social)'s backend [yarnd](https://git.mills.io/yarnsocial/yarn) -- I'll post links here once I've done that (working on client side at the moment)
### Pre-Submission Checklist
- [X] I agree to follow the [Code of Conduct](http://www.authelia.com/code-of-conduct)
- [X] I have checked for related issues and checked the documentation
|
design
|
indieauth identity provider description much like i d love to see indieauth as a first class identity provider in authelia this would allow me to use authelia in quite a number of growing places that are supporting indieauth if anyone is interested in this as well i could help implement this by borrowing from implementation use case numerous but primarily to use authelia itself as an identity provider for any supported website or service that supports indieauth details a user wising to authenticate themselves to a website or web app or anything else really that supports would enter the address url of their authelia instance for example as a user i would enter auth mydomain tld if i m not already authenticated to authelia i will be asked to sign in with my usual authelia credentials once signed in authelia will ask me whether i want to approve the login request for the website app once approved authorisation tokens are generated and i m then signed in with my authelia identity see below for specific details on the flow and implementation documentation see and on the see also there are a however i am planning on separating out the ones used in backend i ll post links here once i ve done that working on client side at the moment pre submission checklist i agree to follow the i have checked for related issues and checked the documentation
| 1
|
120,861
| 15,818,033,923
|
IssuesEvent
|
2021-04-05 15:26:56
|
tomgilder/routemaster
|
https://api.github.com/repos/tomgilder/routemaster
|
closed
|
API question: should we use RouteData or just a string?
|
API design
|
`Routemaster` current extends `RouterDelegate<RouteData>`, but it could potentially just extend `RouterDelegate<String>`.
`RouteData` currently just wraps a route string with nothing else. If there's nothing going to be added to this, we should just use a string.
* Is there a reason to provide an object other than String?
* Might we want to add extra data to the routing info later?
* Could this play a part in state restoration?
|
1.0
|
API question: should we use RouteData or just a string? - `Routemaster` current extends `RouterDelegate<RouteData>`, but it could potentially just extend `RouterDelegate<String>`.
`RouteData` currently just wraps a route string with nothing else. If there's nothing going to be added to this, we should just use a string.
* Is there a reason to provide an object other than String?
* Might we want to add extra data to the routing info later?
* Could this play a part in state restoration?
|
design
|
api question should we use routedata or just a string routemaster current extends routerdelegate but it could potentially just extend routerdelegate routedata currently just wraps a route string with nothing else if there s nothing going to be added to this we should just use a string is there a reason to provide an object other than string might we want to add extra data to the routing info later could this play a part in state restoration
| 1
|
723,621
| 24,903,252,926
|
IssuesEvent
|
2022-10-29 00:40:39
|
bireme/thesaurus-wp-plugin
|
https://api.github.com/repos/bireme/thesaurus-wp-plugin
|
closed
|
Na geração do DeCS do Portal, caso o descritor em inglês não tiver uma ScopeNote, mover a ScopeNote do PEP de ConceptUI mais próximo para o ScopeNote do conceito preferido desse descritor
|
enhancement priority 3
|
Hoje encontrei um descritor novo do MeSH cujo descritor (Sanitary Surveys, Water Supply) estava sem definição na planilha produzida em outubro de 2020 pela NLM para tradução. Mas esse descritor tem um PEP (conceito não preferido) e esse tem uma definição. Aí, eu vi que no MeSH browser essa definição está no conceito preferido, isto é no descritor e não no PEP (ou seja, de duas uma: ou a NLM tem um procedimento que identifica se o descritor está sem definição move a do seu primeiro PEP para ele, apenas no Browser, ou eles atualizaram o registro, alterando fisicamente a nota para o conceito preferido.)
Apesar ser um erro conceitual (pois o conteúdo dessa definição cobre o significado do PEP e não do descritor, ligeiramente mais específico), para nós é bom, pois assim no nosso sistema de consulta a definição vai aparecer. Senão o usuário só enxergaria a definição se abrisse o conceito. Geralmente os descritores do MeSH possuem definição, mas quando não possuem, geralmente apenas um de seus PEPs possuem definição.
Por isso, para resolver para os outros casos em que isso acontece, quando for possível, solicito que para o DeCS do portal de consulta no WP, "se o MH não tiver definição, mas algum de seus PEPs tiverem, mostrar a do primeiro PEP por ordem numérica de ConceptUI do descritor. Obrigado.
|
1.0
|
Na geração do DeCS do Portal, caso o descritor em inglês não tiver uma ScopeNote, mover a ScopeNote do PEP de ConceptUI mais próximo para o ScopeNote do conceito preferido desse descritor - Hoje encontrei um descritor novo do MeSH cujo descritor (Sanitary Surveys, Water Supply) estava sem definição na planilha produzida em outubro de 2020 pela NLM para tradução. Mas esse descritor tem um PEP (conceito não preferido) e esse tem uma definição. Aí, eu vi que no MeSH browser essa definição está no conceito preferido, isto é no descritor e não no PEP (ou seja, de duas uma: ou a NLM tem um procedimento que identifica se o descritor está sem definição move a do seu primeiro PEP para ele, apenas no Browser, ou eles atualizaram o registro, alterando fisicamente a nota para o conceito preferido.)
Apesar ser um erro conceitual (pois o conteúdo dessa definição cobre o significado do PEP e não do descritor, ligeiramente mais específico), para nós é bom, pois assim no nosso sistema de consulta a definição vai aparecer. Senão o usuário só enxergaria a definição se abrisse o conceito. Geralmente os descritores do MeSH possuem definição, mas quando não possuem, geralmente apenas um de seus PEPs possuem definição.
Por isso, para resolver para os outros casos em que isso acontece, quando for possível, solicito que para o DeCS do portal de consulta no WP, "se o MH não tiver definição, mas algum de seus PEPs tiverem, mostrar a do primeiro PEP por ordem numérica de ConceptUI do descritor. Obrigado.
|
non_design
|
na geração do decs do portal caso o descritor em inglês não tiver uma scopenote mover a scopenote do pep de conceptui mais próximo para o scopenote do conceito preferido desse descritor hoje encontrei um descritor novo do mesh cujo descritor sanitary surveys water supply estava sem definição na planilha produzida em outubro de pela nlm para tradução mas esse descritor tem um pep conceito não preferido e esse tem uma definição aí eu vi que no mesh browser essa definição está no conceito preferido isto é no descritor e não no pep ou seja de duas uma ou a nlm tem um procedimento que identifica se o descritor está sem definição move a do seu primeiro pep para ele apenas no browser ou eles atualizaram o registro alterando fisicamente a nota para o conceito preferido apesar ser um erro conceitual pois o conteúdo dessa definição cobre o significado do pep e não do descritor ligeiramente mais específico para nós é bom pois assim no nosso sistema de consulta a definição vai aparecer senão o usuário só enxergaria a definição se abrisse o conceito geralmente os descritores do mesh possuem definição mas quando não possuem geralmente apenas um de seus peps possuem definição por isso para resolver para os outros casos em que isso acontece quando for possível solicito que para o decs do portal de consulta no wp se o mh não tiver definição mas algum de seus peps tiverem mostrar a do primeiro pep por ordem numérica de conceptui do descritor obrigado
| 0
|
15,242
| 19,179,995,635
|
IssuesEvent
|
2021-12-04 07:50:01
|
varabyte/kobweb
|
https://api.github.com/repos/varabyte/kobweb
|
closed
|
Audit ktor plugins
|
process
|
When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
|
1.0
|
Audit ktor plugins - When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
|
non_design
|
audit ktor plugins when i created the initial project i selected a bunch of plugins that seemed relevant but right now they may not be being used and some of them are generating compiler warnings
| 0
|
170,666
| 26,995,984,249
|
IssuesEvent
|
2023-02-10 01:05:37
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
do not set FatalError handlers in unit tests
|
Area-IDE Test Need Design Review Integration-Test
|
looks like we are setting those handlers in unit test making unit test that checks error case to hit the code and fail the test.
|
1.0
|
do not set FatalError handlers in unit tests - looks like we are setting those handlers in unit test making unit test that checks error case to hit the code and fail the test.
|
design
|
do not set fatalerror handlers in unit tests looks like we are setting those handlers in unit test making unit test that checks error case to hit the code and fail the test
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.