Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
137,437
| 20,146,898,010
|
IssuesEvent
|
2022-02-09 08:33:23
|
Joystream/atlas
|
https://api.github.com/repos/Joystream/atlas
|
opened
|
Comments Design: Video page
|
design
|
### User stories
1. As a content viewer, I want to see other people's comments under a video I'm watching.
2. As a content viewer, I want to react to other people's comments under a video I'm watching.
3. As a content viewer, I want to sort other people's comments under a video I'm watching.
4. As a content viewer I want to leave a comment under a video I'm watching.
5. As a content viewer, I want to reply to other people's comments under a video I'm watching.
|
1.0
|
Comments Design: Video page - ### User stories
1. As a content viewer, I want to see other people's comments under a video I'm watching.
2. As a content viewer, I want to react to other people's comments under a video I'm watching.
3. As a content viewer, I want to sort other people's comments under a video I'm watching.
4. As a content viewer I want to leave a comment under a video I'm watching.
5. As a content viewer, I want to reply to other people's comments under a video I'm watching.
|
non_process
|
comments design video page user stories as a content viewer i want to see other people s comments under a video i m watching as a content viewer i want to react to other people s comments under a video i m watching as a content viewer i want to sort other people s comments under a video i m watching as a content viewer i want to leave a comment under a video i m watching as a content viewer i want to reply to other people s comments under a video i m watching
| 0
|
21,718
| 30,220,022,201
|
IssuesEvent
|
2023-07-05 18:35:13
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
opened
|
Error on reviveTerminalProcesses in remote
|
bug remote terminal-process
|
Repro:
1. Open OSS
2. Open new test resolver window
3. Close non-test resolver window
4. Close all terminals
5. Exit application
6. Open OSS
7. Open pty host log
I see an error being returned from `reviveTerminalProcesses`:
```
2023-07-05 11:32:09.225 [trace] [RPC Request] PtyService#reviveTerminalProcesses...
2023-07-05 11:32:09.226 [error] [RPC Response] PtyService#reviveTerminalProcesses Cannot read properties of undefined (reading 'cwd')
```
|
1.0
|
Error on reviveTerminalProcesses in remote - Repro:
1. Open OSS
2. Open new test resolver window
3. Close non-test resolver window
4. Close all terminals
5. Exit application
6. Open OSS
7. Open pty host log
I see an error being returned from `reviveTerminalProcesses`:
```
2023-07-05 11:32:09.225 [trace] [RPC Request] PtyService#reviveTerminalProcesses...
2023-07-05 11:32:09.226 [error] [RPC Response] PtyService#reviveTerminalProcesses Cannot read properties of undefined (reading 'cwd')
```
|
process
|
error on reviveterminalprocesses in remote repro open oss open new test resolver window close non test resolver window close all terminals exit application open oss open pty host log i see an error being returned from reviveterminalprocesses ptyservice reviveterminalprocesses ptyservice reviveterminalprocesses cannot read properties of undefined reading cwd
| 1
|
282,184
| 21,315,470,098
|
IssuesEvent
|
2022-04-16 07:34:42
|
Cyolune/pe
|
https://api.github.com/repos/Cyolune/pe
|
opened
|
Large sequence diagram in Adding Meetings
|
severity.VeryLow type.DocumentationBug
|
The sequence diagram found in Adding Meetings in the DG is large and that makes it complicated. An alternative would be to use reference frames for sections like the time parsing

<!--session: 1650084272383-7e8d4aac-5cf7-47b0-afc2-c94d25d092da-->
<!--Version: Web v3.4.2-->
|
1.0
|
Large sequence diagram in Adding Meetings - The sequence diagram found in Adding Meetings in the DG is large and that makes it complicated. An alternative would be to use reference frames for sections like the time parsing

<!--session: 1650084272383-7e8d4aac-5cf7-47b0-afc2-c94d25d092da-->
<!--Version: Web v3.4.2-->
|
non_process
|
large sequence diagram in adding meetings the sequence diagram found in adding meetings in the dg is large and that makes it complicated an alternative would be to use reference frames for sections like the time parsing
| 0
|
4,986
| 7,818,563,967
|
IssuesEvent
|
2018-06-13 12:43:57
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Side-by-side support of different bazel versions
|
category: misc > release / binary type: process
|
> ATTENTION! Please read and follow:
> - if this is a _question_ about how to build / test / query / deploy using Bazel, ask it on StackOverflow instead: https://stackoverflow.com/questions/tagged/bazel
> - if this is a _discussion starter_, send it to bazel-discuss@googlegroups.com
> - if this is a _bug_ or _feature request_, fill the form below as best as you can.
### Description of the problem / feature request:
Support for multiple versions installed side-by-side by default. Maybe a built in way to switch between them. My apologies if this belongs in the `discussion starter` category.
### Feature requests: what underlying problem are you trying to solve with this feature?
Concurrent work on multiple projects using different versions of bazel. See [this PR](https://github.com/Homebrew/homebrew-core/pull/22470) against `homebrew` for further context.
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
N/A
### What operating system are you running Bazel on?
macOS.
### What's the output of `bazel info release`?
N/A
### If `bazel info release` returns "development version" or "(@non-git)", tell us how you built Bazel.
N/A
### What's the output of `git remote get-url origin ; git rev-parse master ; git rev-parse HEAD` ?
N/A
### Have you found anything relevant by searching the web?
Yep, a general discussion can be found in this [email thread](https://groups.google.com/forum/#!topic/bazel-discuss/7FdIOJdFMyc).
### Any other information, logs, or outputs that you want to share?
N/A
|
1.0
|
Side-by-side support of different bazel versions - > ATTENTION! Please read and follow:
> - if this is a _question_ about how to build / test / query / deploy using Bazel, ask it on StackOverflow instead: https://stackoverflow.com/questions/tagged/bazel
> - if this is a _discussion starter_, send it to bazel-discuss@googlegroups.com
> - if this is a _bug_ or _feature request_, fill the form below as best as you can.
### Description of the problem / feature request:
Support for multiple versions installed side-by-side by default. Maybe a built in way to switch between them. My apologies if this belongs in the `discussion starter` category.
### Feature requests: what underlying problem are you trying to solve with this feature?
Concurrent work on multiple projects using different versions of bazel. See [this PR](https://github.com/Homebrew/homebrew-core/pull/22470) against `homebrew` for further context.
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
N/A
### What operating system are you running Bazel on?
macOS.
### What's the output of `bazel info release`?
N/A
### If `bazel info release` returns "development version" or "(@non-git)", tell us how you built Bazel.
N/A
### What's the output of `git remote get-url origin ; git rev-parse master ; git rev-parse HEAD` ?
N/A
### Have you found anything relevant by searching the web?
Yep, a general discussion can be found in this [email thread](https://groups.google.com/forum/#!topic/bazel-discuss/7FdIOJdFMyc).
### Any other information, logs, or outputs that you want to share?
N/A
|
process
|
side by side support of different bazel versions attention please read and follow if this is a question about how to build test query deploy using bazel ask it on stackoverflow instead if this is a discussion starter send it to bazel discuss googlegroups com if this is a bug or feature request fill the form below as best as you can description of the problem feature request support for multiple versions installed side by side by default maybe a built in way to switch between them my apologies if this belongs in the discussion starter category feature requests what underlying problem are you trying to solve with this feature concurrent work on multiple projects using different versions of bazel see against homebrew for further context bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible n a what operating system are you running bazel on macos what s the output of bazel info release n a if bazel info release returns development version or non git tell us how you built bazel n a what s the output of git remote get url origin git rev parse master git rev parse head n a have you found anything relevant by searching the web yep a general discussion can be found in this any other information logs or outputs that you want to share n a
| 1
|
10,785
| 8,720,922,679
|
IssuesEvent
|
2018-12-08 17:38:36
|
aspnet/AspNetCore
|
https://api.github.com/repos/aspnet/AspNetCore
|
closed
|
AspNetCore-Tooling symbol publishing isn't working
|
3 - Done area-infrastructure task
|
See: https://devdiv.visualstudio.com/DevDiv/_git/VS/pullrequest/154449?_a=overview
Need to investigate why symbol publishing from our tooling builds isn't working for VS's build requirements.
|
1.0
|
AspNetCore-Tooling symbol publishing isn't working - See: https://devdiv.visualstudio.com/DevDiv/_git/VS/pullrequest/154449?_a=overview
Need to investigate why symbol publishing from our tooling builds isn't working for VS's build requirements.
|
non_process
|
aspnetcore tooling symbol publishing isn t working see need to investigate why symbol publishing from our tooling builds isn t working for vs s build requirements
| 0
|
11,480
| 14,349,479,166
|
IssuesEvent
|
2020-11-29 16:46:08
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
opened
|
[FR] torch.load should support loading directly to pinned/shared memory
|
module: multiprocessing module: serialization
|
## 🚀 Feature
The current code can be modified to do so without large changes. Loading to these special memory pages can be very helpful, e.g., when the loaded tensors are to be shared among processes.
|
1.0
|
[FR] torch.load should support loading directly to pinned/shared memory - ## 🚀 Feature
The current code can be modified to do so without large changes. Loading to these special memory pages can be very helpful, e.g., when the loaded tensors are to be shared among processes.
|
process
|
torch load should support loading directly to pinned shared memory 🚀 feature the current code can be modified to do so without large changes loading to these special memory pages can be very helpful e g when the loaded tensors are to be shared among processes
| 1
|
98,294
| 20,670,015,383
|
IssuesEvent
|
2022-03-10 00:26:08
|
pywbem/pywbem
|
https://api.github.com/repos/pywbem/pywbem
|
closed
|
Jobs manager extension -> save the code
|
area: code type: enhancement resolution: partly fixed
|
Provide a service similiar in nature to the existing subscription manager to manage CIM Jobs in WBEM servers
|
1.0
|
Jobs manager extension -> save the code - Provide a service similiar in nature to the existing subscription manager to manage CIM Jobs in WBEM servers
|
non_process
|
jobs manager extension save the code provide a service similiar in nature to the existing subscription manager to manage cim jobs in wbem servers
| 0
|
264,619
| 28,209,220,040
|
IssuesEvent
|
2023-04-05 01:40:30
|
Trinadh465/linux_4.19.72_CVE-2023-42896
|
https://api.github.com/repos/Trinadh465/linux_4.19.72_CVE-2023-42896
|
closed
|
CVE-2022-26490 (High) detected in linuxlinux-4.19.279 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-26490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux_4.19.72_CVE-2023-42896/commit/bab817ff3ba663d8c0b4d4ec9cb6c7069f775527">bab817ff3ba663d8c0b4d4ec9cb6c7069f775527</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
st21nfca_connectivity_event_received in drivers/nfc/st21nfca/se.c in the Linux kernel through 5.16.12 has EVT_TRANSACTION buffer overflows because of untrusted length parameters.
<p>Publish Date: 2022-03-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-26490>CVE-2022-26490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-26490">https://www.linuxkernelcves.com/cves/CVE-2022-26490</a></p>
<p>Release Date: 2022-03-06</p>
<p>Fix Resolution: v4.9.309,v4.14.274,v4.19.237,v5.4.188,v5.10.109,v5.15.32,v5.16.18</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-26490 (High) detected in linuxlinux-4.19.279 - autoclosed - ## CVE-2022-26490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux_4.19.72_CVE-2023-42896/commit/bab817ff3ba663d8c0b4d4ec9cb6c7069f775527">bab817ff3ba663d8c0b4d4ec9cb6c7069f775527</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
st21nfca_connectivity_event_received in drivers/nfc/st21nfca/se.c in the Linux kernel through 5.16.12 has EVT_TRANSACTION buffer overflows because of untrusted length parameters.
<p>Publish Date: 2022-03-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-26490>CVE-2022-26490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-26490">https://www.linuxkernelcves.com/cves/CVE-2022-26490</a></p>
<p>Release Date: 2022-03-06</p>
<p>Fix Resolution: v4.9.309,v4.14.274,v4.19.237,v5.4.188,v5.10.109,v5.15.32,v5.16.18</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details connectivity event received in drivers nfc se c in the linux kernel through has evt transaction buffer overflows because of untrusted length parameters publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
2,410
| 5,196,598,859
|
IssuesEvent
|
2017-01-23 13:22:37
|
raphym/Simulation_of_message_routing_by_intelligent_agents
|
https://api.github.com/repos/raphym/Simulation_of_message_routing_by_intelligent_agents
|
opened
|
Node / Quorum / Map modification with id
|
being processed
|
When i discover that a node is a quorum , so instead to add the new node (quorum) in the end of the vector of all elements of the map , i have to replace the old node with the new into the same place in the vector.
So for create a node with the same id that the old i change the constructor of the node class and quorum class.
|
1.0
|
Node / Quorum / Map modification with id - When i discover that a node is a quorum , so instead to add the new node (quorum) in the end of the vector of all elements of the map , i have to replace the old node with the new into the same place in the vector.
So for create a node with the same id that the old i change the constructor of the node class and quorum class.
|
process
|
node quorum map modification with id when i discover that a node is a quorum so instead to add the new node quorum in the end of the vector of all elements of the map i have to replace the old node with the new into the same place in the vector so for create a node with the same id that the old i change the constructor of the node class and quorum class
| 1
|
15,867
| 20,036,245,689
|
IssuesEvent
|
2022-02-02 12:13:33
|
googlefonts/noto-fonts
|
https://api.github.com/repos/googlefonts/noto-fonts
|
closed
|
Too many files in directories
|
Noto-Process-Issue
|
There are really too many files in the directory of fonts (notably the hinted and unhinted ones, possibly as well for some staged versions)
I really suggest using one directory by font family, or by ISO 15924 script code (like "Latn" for the base fonts, "Zsym" for symbols, "Szye" for emojis), and put in them all related font variants (bold, italic, various weights, cursive/unjoined).
This would limit the number of directories to less than 300 instead of having more than 1800 files (in hinted fonts) that are hard to find and nearly one half fonts hidden to users.
And this would allow working on fonts having common properties, related to the script they support.
----
Note that Noto font names are not always consistant, some of them using abbreviations, some of them using spaces between words, some not: why don't you use the "Unicode Alias" (as displayed in the ISO 15924 RA site and defined in Unicode), just replacing the underscore separator by a space, so that these names (except Latin which is ommited) are directly predictable from Unicode properties, and just trying to prefix "Noto Sans" or "Noto Serif" before these names, without having to constantly check the long list of Noto fonts to define the CSS stylesheets ?
Also I suggest that ALL fonts include in their internal list of names a short synonym/alias based on the 4-letter ISO 15924 script code like "Noto Sans ISOLatn", "Noto Sans ISOCyrl", "Noto Sans ISOHans", "Noto Sans ISOArab", ISO Sans ISOAran", ...
* For CJK fonts, the ISO15924 codes to use should be: Hans, Hant, Jpan, Kore (rather than "CJK CN/TW/JP/KR"); the "Jpan" font containing also Hiragana and Katakana, the "Kore" font containing also Hangul and Jamos, the "Hans" and Hant" possibly both containing Bopomofo (or only one in "Hanb" ?); but you may want to separate "Hrkt" and "Bpmf". The "Hani" code would not be used. A new ISO 15924 code ("Hanv") could be requested for Vietnamese Chu Nôm if it must differ significantly from "Hant" (my opinion is that those differences are small, and could be handled in the "Hant" font as language variant).
* For the two variants of Arabic, the ISO 15924 code is Arab for the simpler Western style (joining horizontally), Aran for the Eastern/Persian/Urdu/Nastaliq style (joining diagonally, so that it's much more complex to design for positioning/styling and managing more kinds of junctions).
* For Latin, the "Latf" and "Latg" codes could be used in font names for these wellknown style variants,
* Another ISO 15924 code ("Latj" ?) should be requested for the Latin cursive/joining style (and in my opinion, for Latin, we should request to Unicode the allocation of an explicit "cursive joiner" character like the one that exists in Unicode for Arabic (it could be used as well for Devanagari and all other LTR scripts, the cursive joiner defined for Arabic could be used for any other RTL script if they need it for their own cursive style : it would be convenient to allow better presentation of joining scripts with justified text, where a renderer cannot simply increase the gap between letters without inserting visible joiner glyphs; however the joiner glyphs must be mapped in fonts and there's currently no simple way to do that cleanly like we can do that for Arabic). If there's a joiner glyph defined in the font, its default metric should be the average width of one letter in the script (so this is a visible character, and by default in non-cursive styles, it should be a stroke on the default baseline, not above it, not below it; in cursive styles, the stroke will be adapted to the joining location of letters on both sides and the font will remap some glyph variants for various heights, the renderer will then insert as many glyphs as needed to fil lthe gap; this glyph can also be implicitly divided by the renderer in three parts (left/central/right) to compose what is needed. OpenType however could allow defining metrics for each letters so that they indicate which part on the left or right is a joiner to the next letter, so that this part can be replaced, or these lateral subglyphs could be references to a joining type; the renderer will then compute the matrix of joining types to infer the accurate joiner to use between them, while rendering only the central part of the glyph. the other parts being drawn by the rules on joining types (exactly like for Arabic)
|
1.0
|
Too many files in directories - There are really too many files in the directory of fonts (notably the hinted and unhinted ones, possibly as well for some staged versions)
I really suggest using one directory by font family, or by ISO 15924 script code (like "Latn" for the base fonts, "Zsym" for symbols, "Szye" for emojis), and put in them all related font variants (bold, italic, various weights, cursive/unjoined).
This would limit the number of directories to less than 300 instead of having more than 1800 files (in hinted fonts) that are hard to find and nearly one half fonts hidden to users.
And this would allow working on fonts having common properties, related to the script they support.
----
Note that Noto font names are not always consistant, some of them using abbreviations, some of them using spaces between words, some not: why don't you use the "Unicode Alias" (as displayed in the ISO 15924 RA site and defined in Unicode), just replacing the underscore separator by a space, so that these names (except Latin which is ommited) are directly predictable from Unicode properties, and just trying to prefix "Noto Sans" or "Noto Serif" before these names, without having to constantly check the long list of Noto fonts to define the CSS stylesheets ?
Also I suggest that ALL fonts include in their internal list of names a short synonym/alias based on the 4-letter ISO 15924 script code like "Noto Sans ISOLatn", "Noto Sans ISOCyrl", "Noto Sans ISOHans", "Noto Sans ISOArab", ISO Sans ISOAran", ...
* For CJK fonts, the ISO15924 codes to use should be: Hans, Hant, Jpan, Kore (rather than "CJK CN/TW/JP/KR"); the "Jpan" font containing also Hiragana and Katakana, the "Kore" font containing also Hangul and Jamos, the "Hans" and Hant" possibly both containing Bopomofo (or only one in "Hanb" ?); but you may want to separate "Hrkt" and "Bpmf". The "Hani" code would not be used. A new ISO 15924 code ("Hanv") could be requested for Vietnamese Chu Nôm if it must differ significantly from "Hant" (my opinion is that those differences are small, and could be handled in the "Hant" font as language variant).
* For the two variants of Arabic, the ISO 15924 code is Arab for the simpler Western style (joining horizontally), Aran for the Eastern/Persian/Urdu/Nastaliq style (joining diagonally, so that it's much more complex to design for positioning/styling and managing more kinds of junctions).
* For Latin, the "Latf" and "Latg" codes could be used in font names for these wellknown style variants,
* Another ISO 15924 code ("Latj" ?) should be requested for the Latin cursive/joining style (and in my opinion, for Latin, we should request to Unicode the allocation of an explicit "cursive joiner" character like the one that exists in Unicode for Arabic (it could be used as well for Devanagari and all other LTR scripts, the cursive joiner defined for Arabic could be used for any other RTL script if they need it for their own cursive style : it would be convenient to allow better presentation of joining scripts with justified text, where a renderer cannot simply increase the gap between letters without inserting visible joiner glyphs; however the joiner glyphs must be mapped in fonts and there's currently no simple way to do that cleanly like we can do that for Arabic). If there's a joiner glyph defined in the font, its default metric should be the average width of one letter in the script (so this is a visible character, and by default in non-cursive styles, it should be a stroke on the default baseline, not above it, not below it; in cursive styles, the stroke will be adapted to the joining location of letters on both sides and the font will remap some glyph variants for various heights, the renderer will then insert as many glyphs as needed to fil lthe gap; this glyph can also be implicitly divided by the renderer in three parts (left/central/right) to compose what is needed. OpenType however could allow defining metrics for each letters so that they indicate which part on the left or right is a joiner to the next letter, so that this part can be replaced, or these lateral subglyphs could be references to a joining type; the renderer will then compute the matrix of joining types to infer the accurate joiner to use between them, while rendering only the central part of the glyph. the other parts being drawn by the rules on joining types (exactly like for Arabic)
|
process
|
too many files in directories there are really too many files in the directory of fonts notably the hinted and unhinted ones possibly as well for some staged versions i really suggest using one directory by font family or by iso script code like latn for the base fonts zsym for symbols szye for emojis and put in them all related font variants bold italic various weights cursive unjoined this would limit the number of directories to less than instead of having more than files in hinted fonts that are hard to find and nearly one half fonts hidden to users and this would allow working on fonts having common properties related to the script they support note that noto font names are not always consistant some of them using abbreviations some of them using spaces between words some not why don t you use the unicode alias as displayed in the iso ra site and defined in unicode just replacing the underscore separator by a space so that these names except latin which is ommited are directly predictable from unicode properties and just trying to prefix noto sans or noto serif before these names without having to constantly check the long list of noto fonts to define the css stylesheets also i suggest that all fonts include in their internal list of names a short synonym alias based on the letter iso script code like noto sans isolatn noto sans isocyrl noto sans isohans noto sans isoarab iso sans isoaran for cjk fonts the codes to use should be hans hant jpan kore rather than cjk cn tw jp kr the jpan font containing also hiragana and katakana the kore font containing also hangul and jamos the hans and hant possibly both containing bopomofo or only one in hanb but you may want to separate hrkt and bpmf the hani code would not be used a new iso code hanv could be requested for vietnamese chu nôm if it must differ significantly from hant my opinion is that those differences are small and could be handled in the hant font as language variant for the two variants of arabic the iso code is arab for the simpler western style joining horizontally aran for the eastern persian urdu nastaliq style joining diagonally so that it s much more complex to design for positioning styling and managing more kinds of junctions for latin the latf and latg codes could be used in font names for these wellknown style variants another iso code latj should be requested for the latin cursive joining style and in my opinion for latin we should request to unicode the allocation of an explicit cursive joiner character like the one that exists in unicode for arabic it could be used as well for devanagari and all other ltr scripts the cursive joiner defined for arabic could be used for any other rtl script if they need it for their own cursive style it would be convenient to allow better presentation of joining scripts with justified text where a renderer cannot simply increase the gap between letters without inserting visible joiner glyphs however the joiner glyphs must be mapped in fonts and there s currently no simple way to do that cleanly like we can do that for arabic if there s a joiner glyph defined in the font its default metric should be the average width of one letter in the script so this is a visible character and by default in non cursive styles it should be a stroke on the default baseline not above it not below it in cursive styles the stroke will be adapted to the joining location of letters on both sides and the font will remap some glyph variants for various heights the renderer will then insert as many glyphs as needed to fil lthe gap this glyph can also be implicitly divided by the renderer in three parts left central right to compose what is needed opentype however could allow defining metrics for each letters so that they indicate which part on the left or right is a joiner to the next letter so that this part can be replaced or these lateral subglyphs could be references to a joining type the renderer will then compute the matrix of joining types to infer the accurate joiner to use between them while rendering only the central part of the glyph the other parts being drawn by the rules on joining types exactly like for arabic
| 1
|
7,330
| 24,648,554,123
|
IssuesEvent
|
2022-10-17 16:38:08
|
bcgov/api-services-portal
|
https://api.github.com/repos/bcgov/api-services-portal
|
closed
|
Cypress Test - Consumer detail - edit labels
|
automation
|
1. Manage/Edit labels spec
1.1 authenticates Mark (Access-Manager)
1.2 Navigate to Consumer page and filter the product
1.3 Click on the first consumer
1.4 Verify that labels can be deleted
1.5 Verify that labels can be updated
1.6 Verify that labels can be added
|
1.0
|
Cypress Test - Consumer detail - edit labels - 1. Manage/Edit labels spec
1.1 authenticates Mark (Access-Manager)
1.2 Navigate to Consumer page and filter the product
1.3 Click on the first consumer
1.4 Verify that labels can be deleted
1.5 Verify that labels can be updated
1.6 Verify that labels can be added
|
non_process
|
cypress test consumer detail edit labels manage edit labels spec authenticates mark access manager navigate to consumer page and filter the product click on the first consumer verify that labels can be deleted verify that labels can be updated verify that labels can be added
| 0
|
377,483
| 11,171,584,125
|
IssuesEvent
|
2019-12-28 21:00:55
|
openmsupply/mobile
|
https://api.github.com/repos/openmsupply/mobile
|
opened
|
Disabled the Next button until items have been chosen in a prescription
|
Bug: development Docs: not needed Effort: small Module: dispensary Priority: high
|
## Describe the bug
Currently can proceed through the 'checkout' with no items.
### To reproduce
Steps to reproduce the behavior:
1. Create a new prescription
2. don't add any items and click next
4. See error
### Expected behaviour
`NEXT` should be disabled until at least one item has been chosen.
### Proposed Solution
N/A
### Version and device info
Dispensing development
### Additional context
N/A
|
1.0
|
Disabled the Next button until items have been chosen in a prescription - ## Describe the bug
Currently can proceed through the 'checkout' with no items.
### To reproduce
Steps to reproduce the behavior:
1. Create a new prescription
2. don't add any items and click next
4. See error
### Expected behaviour
`NEXT` should be disabled until at least one item has been chosen.
### Proposed Solution
N/A
### Version and device info
Dispensing development
### Additional context
N/A
|
non_process
|
disabled the next button until items have been chosen in a prescription describe the bug currently can proceed through the checkout with no items to reproduce steps to reproduce the behavior create a new prescription don t add any items and click next see error expected behaviour next should be disabled until at least one item has been chosen proposed solution n a version and device info dispensing development additional context n a
| 0
|
20,203
| 26,779,255,538
|
IssuesEvent
|
2023-01-31 19:41:21
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
closed
|
promote library to GA
|
type: process api: mediatranslation
|
Package name: **@google-cloud/media-translation**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
1.0
|
promote library to GA - Package name: **@google-cloud/media-translation**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
process
|
promote library to ga package name google cloud media translation current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
| 1
|
8,648
| 11,789,750,207
|
IssuesEvent
|
2020-03-17 17:41:42
|
googleapis/synthtool
|
https://api.github.com/repos/googleapis/synthtool
|
closed
|
Java: missing template tests
|
type: process
|
- [x] Run template generation using a sample `.repo-metadata.json` file
- [ ] Ensure any generated `pom.xml` files are valid XML
|
1.0
|
Java: missing template tests - - [x] Run template generation using a sample `.repo-metadata.json` file
- [ ] Ensure any generated `pom.xml` files are valid XML
|
process
|
java missing template tests run template generation using a sample repo metadata json file ensure any generated pom xml files are valid xml
| 1
|
205,574
| 15,648,067,863
|
IssuesEvent
|
2021-03-23 04:52:21
|
RTXteam/RTX
|
https://api.github.com/repos/RTXteam/RTX
|
closed
|
ARAX returning TRAPI-noncompliant result?
|
ARAX demo need to test on production question
|
From the BDT Friday stand-up meeting (2021.03.12),
<img width="1441" alt="Screen Shot 2021-03-12 at 9 04 20 AM" src="https://user-images.githubusercontent.com/5562107/110973381-0adf7a00-8312-11eb-98f2-e90c9bfd8471.png">
I think Sarah said that the red "50" means results that are not TRAPI-compliant. Just bringing this up in case it is something fixable that is not currently on the docket to be fixed via another issue (apologies if this already duplicates an in-progress issue; just close it out in that case).
|
1.0
|
ARAX returning TRAPI-noncompliant result? - From the BDT Friday stand-up meeting (2021.03.12),
<img width="1441" alt="Screen Shot 2021-03-12 at 9 04 20 AM" src="https://user-images.githubusercontent.com/5562107/110973381-0adf7a00-8312-11eb-98f2-e90c9bfd8471.png">
I think Sarah said that the red "50" means results that are not TRAPI-compliant. Just bringing this up in case it is something fixable that is not currently on the docket to be fixed via another issue (apologies if this already duplicates an in-progress issue; just close it out in that case).
|
non_process
|
arax returning trapi noncompliant result from the bdt friday stand up meeting img width alt screen shot at am src i think sarah said that the red means results that are not trapi compliant just bringing this up in case it is something fixable that is not currently on the docket to be fixed via another issue apologies if this already duplicates an in progress issue just close it out in that case
| 0
|
17,095
| 22,609,855,234
|
IssuesEvent
|
2022-06-29 16:09:20
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
Fujitsu FR60/80 Support
|
Feature: Processor/Unsupported Status: Waiting on customer
|
Summary
----
I developed a [FR 60 Plugin](https://github.com/desrdev/ghidra-fr60) and I would like to upstream this work. I am unsure if the maintainers of this project are interested in accepting this functionality to mainline. If there is interest, would it be for generalizable functionality of the architecture or the core specific features as well.
I am also wondering what the best avenue is to ask questions regarding this work if I have any (github issues perhaps?).
Details
----
I am working on analyzing the DVR processor in the Sony PSX which uses the Fujitsu [MB91302A](https://www.fujitsu.com/downloads/MICRO/fma/pdfmcu/hm91301CM71-10114-3E.pdf) processor.
To aid with this I've built a Ghidra plugin which enables decompilation of the firmware and automatically setting up memory space and identifying & labeling registers and Fujitsu uTron kernel system calls.
This plugin is available on my github here: https://github.com/desrdev/ghidra-fr60
This was developed with a single goal in mind, and after completing the work I spotted the following projects targeting similar goals and architecture:
https://github.com/simeonpilgrim/ghidra_fujitsu_fr
https://github.com/ammaraskar/ghidra-fujitsu-risc
With the exception of `ghidra-fujitsu-risc`, the projects are released with compatible licenses to Ghidra. Both projects are targeting a single processor on a specific kind of device however and are not truly general plugins for the Fujitsu 32bit processors.
From a quick glance, all three of our implementations are fairly close to each other which suggests the CPU instruction architecture can be generalized across all 32bit Fujitsu processors.
Fujitsu no longer manufactures or advertises these components, however Infineon/Cypress do market these cores and they are still in active use today.
Given that the core is actively used and there are now 3 separate targets repeating work, I feel there is a case to be made to upstream some part of these projects to Ghidra mainline to allow reuse of the generalizable aspects of all three of these projects.
**I am open to doing this work and maintaining this new feature**, but I am unfamiliar with the process, style guide, and best practices when designing Ghidra cores or contributing to the Ghidra project in general.
Even in my own plugin there are features and functions I wish to add, but I am not sure of the idiomatic way to approach them.
To reiterate, I am willing to do the work required to get this mainline ready, and the plugin I developed is fully functional and I use it almost daily, but I will likely need some assistance to do so and I understand if the maintainers do not have the bandwidth/interest to take this on.
Thank you for your time.
|
1.0
|
Fujitsu FR60/80 Support - Summary
----
I developed a [FR 60 Plugin](https://github.com/desrdev/ghidra-fr60) and I would like to upstream this work. I am unsure if the maintainers of this project are interested in accepting this functionality to mainline. If there is interest, would it be for generalizable functionality of the architecture or the core specific features as well.
I am also wondering what the best avenue is to ask questions regarding this work if I have any (github issues perhaps?).
Details
----
I am working on analyzing the DVR processor in the Sony PSX which uses the Fujitsu [MB91302A](https://www.fujitsu.com/downloads/MICRO/fma/pdfmcu/hm91301CM71-10114-3E.pdf) processor.
To aid with this I've built a Ghidra plugin which enables decompilation of the firmware and automatically setting up memory space and identifying & labeling registers and Fujitsu uTron kernel system calls.
This plugin is available on my github here: https://github.com/desrdev/ghidra-fr60
This was developed with a single goal in mind, and after completing the work I spotted the following projects targeting similar goals and architecture:
https://github.com/simeonpilgrim/ghidra_fujitsu_fr
https://github.com/ammaraskar/ghidra-fujitsu-risc
With the exception of `ghidra-fujitsu-risc`, the projects are released with compatible licenses to Ghidra. Both projects are targeting a single processor on a specific kind of device however and are not truly general plugins for the Fujitsu 32bit processors.
From a quick glance, all three of our implementations are fairly close to each other which suggests the CPU instruction architecture can be generalized across all 32bit Fujitsu processors.
Fujitsu no longer manufactures or advertises these components, however Infineon/Cypress do market these cores and they are still in active use today.
Given that the core is actively used and there are now 3 separate targets repeating work, I feel there is a case to be made to upstream some part of these projects to Ghidra mainline to allow reuse of the generalizable aspects of all three of these projects.
**I am open to doing this work and maintaining this new feature**, but I am unfamiliar with the process, style guide, and best practices when designing Ghidra cores or contributing to the Ghidra project in general.
Even in my own plugin there are features and functions I wish to add, but I am not sure of the idiomatic way to approach them.
To reiterate, I am willing to do the work required to get this mainline ready, and the plugin I developed is fully functional and I use it almost daily, but I will likely need some assistance to do so and I understand if the maintainers do not have the bandwidth/interest to take this on.
Thank you for your time.
|
process
|
fujitsu support summary i developed a and i would like to upstream this work i am unsure if the maintainers of this project are interested in accepting this functionality to mainline if there is interest would it be for generalizable functionality of the architecture or the core specific features as well i am also wondering what the best avenue is to ask questions regarding this work if i have any github issues perhaps details i am working on analyzing the dvr processor in the sony psx which uses the fujitsu processor to aid with this i ve built a ghidra plugin which enables decompilation of the firmware and automatically setting up memory space and identifying labeling registers and fujitsu utron kernel system calls this plugin is available on my github here this was developed with a single goal in mind and after completing the work i spotted the following projects targeting similar goals and architecture with the exception of ghidra fujitsu risc the projects are released with compatible licenses to ghidra both projects are targeting a single processor on a specific kind of device however and are not truly general plugins for the fujitsu processors from a quick glance all three of our implementations are fairly close to each other which suggests the cpu instruction architecture can be generalized across all fujitsu processors fujitsu no longer manufactures or advertises these components however infineon cypress do market these cores and they are still in active use today given that the core is actively used and there are now separate targets repeating work i feel there is a case to be made to upstream some part of these projects to ghidra mainline to allow reuse of the generalizable aspects of all three of these projects i am open to doing this work and maintaining this new feature but i am unfamiliar with the process style guide and best practices when designing ghidra cores or contributing to the ghidra project in general even in my own plugin there are features and functions i wish to add but i am not sure of the idiomatic way to approach them to reiterate i am willing to do the work required to get this mainline ready and the plugin i developed is fully functional and i use it almost daily but i will likely need some assistance to do so and i understand if the maintainers do not have the bandwidth interest to take this on thank you for your time
| 1
|
337,561
| 30,249,536,383
|
IssuesEvent
|
2023-07-06 19:18:31
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix jax_numpy_logic.test_jax_greater
|
JAX Frontend Sub Task Failing Test
|
| | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5475303530"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5475588308"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5473078102"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5473974053"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix jax_numpy_logic.test_jax_greater - | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5475303530"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5475588308"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5473078102"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5473974053"><img src=https://img.shields.io/badge/-success-success></a>
|
non_process
|
fix jax numpy logic test jax greater paddle a href src torch a href src numpy a href src jax a href src tensorflow a href src
| 0
|
807,398
| 30,000,383,982
|
IssuesEvent
|
2023-06-26 08:56:06
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.reddit.com - design is broken
|
priority-critical browser-fenix engine-gecko android13
|
<!-- @browser: Firefox Mobile 116.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/116.0 Firefox/116.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/124045 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.reddit.com/r/midjourney/comments/14hxw5s/average_person_from_different_us_states/
**Browser / Version**: Firefox Mobile 116.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
Image browsing control with content does not load. Loads in chrome android
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/ba7359d2-b583-483f-a8ea-1321707418ad.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230624091338</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/6/57d77eb9-4ce7-47e3-9227-5a28f25d2291)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.reddit.com - design is broken - <!-- @browser: Firefox Mobile 116.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/116.0 Firefox/116.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/124045 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.reddit.com/r/midjourney/comments/14hxw5s/average_person_from_different_us_states/
**Browser / Version**: Firefox Mobile 116.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
Image browsing control with content does not load. Loads in chrome android
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/ba7359d2-b583-483f-a8ea-1321707418ad.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230624091338</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/6/57d77eb9-4ce7-47e3-9227-5a28f25d2291)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce image browsing control with content does not load loads in chrome android view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
444,020
| 12,805,139,295
|
IssuesEvent
|
2020-07-03 06:48:56
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Make staging server build use staging account system
|
Priority: Medium Status: Fixed
|
Account server is https://auth.develop.strangeloopgames.com/, need to use this instead of the normal account server for staging builds.
- Find or add a compile flag to the staging build process (talk to Eugene aka Jaskes about this)
- Make the server check for this during compilation
|
1.0
|
Make staging server build use staging account system - Account server is https://auth.develop.strangeloopgames.com/, need to use this instead of the normal account server for staging builds.
- Find or add a compile flag to the staging build process (talk to Eugene aka Jaskes about this)
- Make the server check for this during compilation
|
non_process
|
make staging server build use staging account system account server is need to use this instead of the normal account server for staging builds find or add a compile flag to the staging build process talk to eugene aka jaskes about this make the server check for this during compilation
| 0
|
65,503
| 14,727,876,951
|
IssuesEvent
|
2021-01-06 09:11:10
|
Seagate/cortx-s3server
|
https://api.github.com/repos/Seagate/cortx-s3server
|
closed
|
CVE-2015-7576 (Low) detected in actionpack-4.2.2.gem
|
needs-attention needs-triage security vulnerability
|
## CVE-2015-7576 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-4.2.2.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-4.2.2.gem">https://rubygems.org/gems/actionpack-4.2.2.gem</a></p>
<p>
Dependency Hierarchy:
- coffee-rails-4.1.0.gem (Root Library)
- railties-4.2.2.gem
- :x: **actionpack-4.2.2.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/fde64200b4f94603ae17220b98da6422a531445e">fde64200b4f94603ae17220b98da6422a531445e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The http_basic_authenticate_with method in actionpack/lib/action_controller/metal/http_authentication.rb in the Basic Authentication implementation in Action Controller in Ruby on Rails before 3.2.22.1, 4.0.x and 4.1.x before 4.1.14.1, 4.2.x before 4.2.5.1, and 5.x before 5.0.0.beta1.1 does not use a constant-time algorithm for verifying credentials, which makes it easier for remote attackers to bypass authentication by measuring timing differences.
<p>Publish Date: 2016-02-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7576>CVE-2015-7576</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-7576">https://nvd.nist.gov/vuln/detail/CVE-2015-7576</a></p>
<p>Release Date: 2016-02-16</p>
<p>Fix Resolution: 3.2.22.1,4.1.14.1,4.2.5.1,5.0.0.beta1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-7576 (Low) detected in actionpack-4.2.2.gem - ## CVE-2015-7576 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-4.2.2.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-4.2.2.gem">https://rubygems.org/gems/actionpack-4.2.2.gem</a></p>
<p>
Dependency Hierarchy:
- coffee-rails-4.1.0.gem (Root Library)
- railties-4.2.2.gem
- :x: **actionpack-4.2.2.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/fde64200b4f94603ae17220b98da6422a531445e">fde64200b4f94603ae17220b98da6422a531445e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The http_basic_authenticate_with method in actionpack/lib/action_controller/metal/http_authentication.rb in the Basic Authentication implementation in Action Controller in Ruby on Rails before 3.2.22.1, 4.0.x and 4.1.x before 4.1.14.1, 4.2.x before 4.2.5.1, and 5.x before 5.0.0.beta1.1 does not use a constant-time algorithm for verifying credentials, which makes it easier for remote attackers to bypass authentication by measuring timing differences.
<p>Publish Date: 2016-02-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7576>CVE-2015-7576</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-7576">https://nvd.nist.gov/vuln/detail/CVE-2015-7576</a></p>
<p>Release Date: 2016-02-16</p>
<p>Fix Resolution: 3.2.22.1,4.1.14.1,4.2.5.1,5.0.0.beta1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve low detected in actionpack gem cve low severity vulnerability vulnerable library actionpack gem web apps on rails simple battle tested conventions for building and testing mvc web applications works with any rack compatible server library home page a href dependency hierarchy coffee rails gem root library railties gem x actionpack gem vulnerable library found in head commit a href found in base branch main vulnerability details the http basic authenticate with method in actionpack lib action controller metal http authentication rb in the basic authentication implementation in action controller in ruby on rails before x and x before x before and x before does not use a constant time algorithm for verifying credentials which makes it easier for remote attackers to bypass authentication by measuring timing differences publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
1,117
| 3,592,095,378
|
IssuesEvent
|
2016-02-01 14:52:04
|
Project60/org.project60.sepa
|
https://api.github.com/repos/Project60/org.project60.sepa
|
closed
|
PaymentProcessor: duplicate period dropdown
|
bug CiviSEPA 1.1 CiviSEPA 1.2 payment processor
|
If you're using the SEPA payment processor on a donation page for recurring contributions with customized intervals along with other payment processors, the generated period dropdown field ("monthly", "quarterly", ...) that replaces the interval input fields will get duplicated if you change the selected payment processor multiple times.
|
1.0
|
PaymentProcessor: duplicate period dropdown - If you're using the SEPA payment processor on a donation page for recurring contributions with customized intervals along with other payment processors, the generated period dropdown field ("monthly", "quarterly", ...) that replaces the interval input fields will get duplicated if you change the selected payment processor multiple times.
|
process
|
paymentprocessor duplicate period dropdown if you re using the sepa payment processor on a donation page for recurring contributions with customized intervals along with other payment processors the generated period dropdown field monthly quarterly that replaces the interval input fields will get duplicated if you change the selected payment processor multiple times
| 1
|
11,303
| 14,106,787,513
|
IssuesEvent
|
2020-11-06 15:24:00
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Recursively process source queries
|
.Query Language (MBQL) Querying/Processor
|
Most of the middleware should not be aware of source queries, but we’d rather just be able to call the entire preprocessing middleware stack on the source query.
|
1.0
|
Recursively process source queries - Most of the middleware should not be aware of source queries, but we’d rather just be able to call the entire preprocessing middleware stack on the source query.
|
process
|
recursively process source queries most of the middleware should not be aware of source queries but we’d rather just be able to call the entire preprocessing middleware stack on the source query
| 1
|
93,801
| 19,339,040,075
|
IssuesEvent
|
2021-12-15 00:48:11
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
opened
|
insights: FE space under series labels
|
bug ux team/code-insights
|
See the video below. There is a lot of extra space underneath the data series labels.
https://user-images.githubusercontent.com/1855233/146102299-8ed0e3e2-84b4-40f9-97cf-9a8b9b62c330.mov
|
1.0
|
insights: FE space under series labels - See the video below. There is a lot of extra space underneath the data series labels.
https://user-images.githubusercontent.com/1855233/146102299-8ed0e3e2-84b4-40f9-97cf-9a8b9b62c330.mov
|
non_process
|
insights fe space under series labels see the video below there is a lot of extra space underneath the data series labels
| 0
|
13,647
| 16,358,668,018
|
IssuesEvent
|
2021-05-14 05:25:57
|
Vanuatu-National-Statistics-Office/vnso-RAP-tradeStats-materials
|
https://api.github.com/repos/Vanuatu-National-Statistics-Office/vnso-RAP-tradeStats-materials
|
closed
|
Increase in observations when merging in Principle Commodities
|
bug dashboard data processing monthly report
|
When merging Principle Commodities there is still a jump from 12540 to 12545 observations. Not sure why this is happening
https://github.com/Vanuatu-National-Statistics-Office/vnso-RAP-tradeStats-materials/issues/49
|
1.0
|
Increase in observations when merging in Principle Commodities - When merging Principle Commodities there is still a jump from 12540 to 12545 observations. Not sure why this is happening
https://github.com/Vanuatu-National-Statistics-Office/vnso-RAP-tradeStats-materials/issues/49
|
process
|
increase in observations when merging in principle commodities when merging principle commodities there is still a jump from to observations not sure why this is happening
| 1
|
118,660
| 11,985,558,139
|
IssuesEvent
|
2020-04-07 17:43:24
|
reapit/foundations
|
https://api.github.com/repos/reapit/foundations
|
closed
|
Create glossary of terminology for platform API documentation
|
documentation platform-team
|
We should create a glossary that explains the terminology that our endpoints and other documentation makes use of.
- Provide clear, concise human readable documentation to explain the concepts exposed by the interactive API explorer
- Also include concepts that we talk about in our documentation (eg. customer / client / user / negotiator)
- Update other documentation and API explorer to link and make reference to the glossary
|
1.0
|
Create glossary of terminology for platform API documentation - We should create a glossary that explains the terminology that our endpoints and other documentation makes use of.
- Provide clear, concise human readable documentation to explain the concepts exposed by the interactive API explorer
- Also include concepts that we talk about in our documentation (eg. customer / client / user / negotiator)
- Update other documentation and API explorer to link and make reference to the glossary
|
non_process
|
create glossary of terminology for platform api documentation we should create a glossary that explains the terminology that our endpoints and other documentation makes use of provide clear concise human readable documentation to explain the concepts exposed by the interactive api explorer also include concepts that we talk about in our documentation eg customer client user negotiator update other documentation and api explorer to link and make reference to the glossary
| 0
|
21,032
| 27,972,094,880
|
IssuesEvent
|
2023-03-25 05:47:29
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Add skip/src/dstnodata params to gdal2xyz (Request in QGIS)
|
Processing Alg 3.32
|
### Request for documentation
From pull request QGIS/qgis#52187
Author: @lpinner
QGIS version: 3.32
**Add skip/src/dstnodata params to gdal2xyz**
### PR Description:
## Description
Adds `-skipnodata` and `-src/dstnodata` parameters to GDAL provider gdal2xyz processing algorithm. These parameters were added to GDAL's gdal2xyz.py utilitues as of GDAL 3.3, though the `-src/dstnodata` params are [broken as of GDAL 3.6](https://github.com/OSGeo/gdal/issues/7410).
Suggested help in [QGIS-Documentation gdal2xyz_params branch](https://github.com/lpinner/QGIS-Documentation/blob/gdal2xyz_params/docs/user_manual/processing_algs/gdal/rasterconversion.rst) (PR: [#8097](https://github.com/qgis/QGIS-Documentation/pull/8097))
### Commits tagged with [need-docs] or [FEATURE]
|
1.0
|
Add skip/src/dstnodata params to gdal2xyz (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#52187
Author: @lpinner
QGIS version: 3.32
**Add skip/src/dstnodata params to gdal2xyz**
### PR Description:
## Description
Adds `-skipnodata` and `-src/dstnodata` parameters to GDAL provider gdal2xyz processing algorithm. These parameters were added to GDAL's gdal2xyz.py utilitues as of GDAL 3.3, though the `-src/dstnodata` params are [broken as of GDAL 3.6](https://github.com/OSGeo/gdal/issues/7410).
Suggested help in [QGIS-Documentation gdal2xyz_params branch](https://github.com/lpinner/QGIS-Documentation/blob/gdal2xyz_params/docs/user_manual/processing_algs/gdal/rasterconversion.rst) (PR: [#8097](https://github.com/qgis/QGIS-Documentation/pull/8097))
### Commits tagged with [need-docs] or [FEATURE]
|
process
|
add skip src dstnodata params to request in qgis request for documentation from pull request qgis qgis author lpinner qgis version add skip src dstnodata params to pr description description adds skipnodata and src dstnodata parameters to gdal provider processing algorithm these parameters were added to gdal s py utilitues as of gdal though the src dstnodata params are suggested help in pr commits tagged with or
| 1
|
21,996
| 30,494,875,284
|
IssuesEvent
|
2023-07-18 10:07:14
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Casting -> coercion
|
doc-enhancement cba Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc
|
I think Type coercion is meant instead of Type casting in the examples with Boolean to Integer coercion
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/expressions.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Casting -> coercion -
I think Type coercion is meant instead of Type casting in the examples with Boolean to Integer coercion
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/expressions.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
casting coercion i think type coercion is meant instead of type casting in the examples with boolean to integer coercion document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login juliakm microsoft alias jukullam
| 1
|
7,247
| 10,415,966,853
|
IssuesEvent
|
2019-09-14 09:09:38
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
GRASS v.build.polylines dosn't work in Qgis 3.4.10
|
Bug Feedback Processing
|
In reference to issue #30680, Grass dosn't work in Qgis 3.4 to 3.8
I'm tested various GRASS algorithms and dosn't works with Qgis 3.4.10 from the Qgis process window.
> The specified GRASS 7 folder "C:\PROGRA~1\QGIS3~1.4\bin\bin" does not contain a valid set of GRASS 7 modules. Please, go to the Processing settings dialog, and check that the GRASS 7 folder is correctly configured
checked in 3 computers from the same office
|
1.0
|
GRASS v.build.polylines dosn't work in Qgis 3.4.10 - In reference to issue #30680, Grass dosn't work in Qgis 3.4 to 3.8
I'm tested various GRASS algorithms and dosn't works with Qgis 3.4.10 from the Qgis process window.
> The specified GRASS 7 folder "C:\PROGRA~1\QGIS3~1.4\bin\bin" does not contain a valid set of GRASS 7 modules. Please, go to the Processing settings dialog, and check that the GRASS 7 folder is correctly configured
checked in 3 computers from the same office
|
process
|
grass v build polylines dosn t work in qgis in reference to issue grass dosn t work in qgis to i m tested various grass algorithms and dosn t works with qgis from the qgis process window the specified grass folder c progra bin bin does not contain a valid set of grass modules please go to the processing settings dialog and check that the grass folder is correctly configured checked in computers from the same office
| 1
|
83,407
| 24,055,513,170
|
IssuesEvent
|
2022-09-16 16:27:06
|
gradle/gradle
|
https://api.github.com/repos/gradle/gradle
|
closed
|
Nested composites with cyclic dependencies cause an error
|
a:bug in:composite-builds stale
|
Version: 4.10.nightly
Let's say you have the following dependencies between composite builds:
```
projectA --compile-->
projectB --testcompile-->
projectC --compile-->
projectA
```
This does not create a cycle due to the testcompile dependency between projectB and projectC.
Now let's say each of these projects is connected with composite build `includedBuild` such that
```
projectA
includeBuild '../projectB'
projectB
includeBuild '../projectC'
projectC
includeBuild '../projectA'
```
Now if you try to build `cd projectA; ./gradlew build`
This will cause the error:
```
Included build in /Users/nicholasdipiazza/projectA has the same root project name 'projectA' as the main build.
```
This is because you are building root project `projectA` but then projectC includes `projectA`.
This should just realize that you already included that build and ignore it.
The work around is to build this project from a different project root, then it works fine.
|
1.0
|
Nested composites with cyclic dependencies cause an error - Version: 4.10.nightly
Let's say you have the following dependencies between composite builds:
```
projectA --compile-->
projectB --testcompile-->
projectC --compile-->
projectA
```
This does not create a cycle due to the testcompile dependency between projectB and projectC.
Now let's say each of these projects is connected with composite build `includedBuild` such that
```
projectA
includeBuild '../projectB'
projectB
includeBuild '../projectC'
projectC
includeBuild '../projectA'
```
Now if you try to build `cd projectA; ./gradlew build`
This will cause the error:
```
Included build in /Users/nicholasdipiazza/projectA has the same root project name 'projectA' as the main build.
```
This is because you are building root project `projectA` but then projectC includes `projectA`.
This should just realize that you already included that build and ignore it.
The work around is to build this project from a different project root, then it works fine.
|
non_process
|
nested composites with cyclic dependencies cause an error version nightly let s say you have the following dependencies between composite builds projecta compile projectb testcompile projectc compile projecta this does not create a cycle due to the testcompile dependency between projectb and projectc now let s say each of these projects is connected with composite build includedbuild such that projecta includebuild projectb projectb includebuild projectc projectc includebuild projecta now if you try to build cd projecta gradlew build this will cause the error included build in users nicholasdipiazza projecta has the same root project name projecta as the main build this is because you are building root project projecta but then projectc includes projecta this should just realize that you already included that build and ignore it the work around is to build this project from a different project root then it works fine
| 0
|
233,456
| 17,865,414,881
|
IssuesEvent
|
2021-09-06 08:51:40
|
NorESMhub/NorESM
|
https://api.github.com/repos/NorESMhub/NorESM
|
opened
|
Archiving NorESM output data
|
Documentation
|
Hi,
We wish to document "Best practices for archiving", which involves both short term and long term archiving. Can you please help me with this?
**Short term archiving**
- What is the best practice to archive files from HPC machines to e.g. NIRD?
- rsync, scp, sshfs (make a mount folder)
- I use rsync, but I'm not sure if this is "best practice" or if the -aP option should be changed:
rsync -aP CASENAME username@login.nird.sigma2.no:/projects/$PROJECTNAME/noresm/cases/
- Data access on nird: what is out policy for sharing data? Read access for all? For the project members?
**Long term archiving**
- What should be archived, where and how?
- Do we have a recipe for long term archiving?
- Is there a script we can run to archive?
A description can either be written her, in the file doc/output/archive_out.rst or you can send me an email.
Hilsen Ada
|
1.0
|
Archiving NorESM output data - Hi,
We wish to document "Best practices for archiving", which involves both short term and long term archiving. Can you please help me with this?
**Short term archiving**
- What is the best practice to archive files from HPC machines to e.g. NIRD?
- rsync, scp, sshfs (make a mount folder)
- I use rsync, but I'm not sure if this is "best practice" or if the -aP option should be changed:
rsync -aP CASENAME username@login.nird.sigma2.no:/projects/$PROJECTNAME/noresm/cases/
- Data access on nird: what is out policy for sharing data? Read access for all? For the project members?
**Long term archiving**
- What should be archived, where and how?
- Do we have a recipe for long term archiving?
- Is there a script we can run to archive?
A description can either be written her, in the file doc/output/archive_out.rst or you can send me an email.
Hilsen Ada
|
non_process
|
archiving noresm output data hi we wish to document best practices for archiving which involves both short term and long term archiving can you please help me with this short term archiving what is the best practice to archive files from hpc machines to e g nird rsync scp sshfs make a mount folder i use rsync but i m not sure if this is best practice or if the ap option should be changed rsync ap casename username login nird no projects projectname noresm cases data access on nird what is out policy for sharing data read access for all for the project members long term archiving what should be archived where and how do we have a recipe for long term archiving is there a script we can run to archive a description can either be written her in the file doc output archive out rst or you can send me an email hilsen ada
| 0
|
13,568
| 16,105,628,489
|
IssuesEvent
|
2021-04-27 14:38:14
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
`process.exit` results in a segmentation fault in v12.x
|
confirmed-bug fs process
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: v12.21.0
* **Platform**: Linux 5.8.0-38-generic #43~20.04.1-Ubuntu SMP Tue Jan 12 16:39:47 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
* **Subsystem**: process.exit
### What steps will reproduce the bug?
Setup a node instance,
```
» node
```
and run the following javascript code.
```
fs.write(7,fs.promises.chown,(err, written, buffer)=>{})
process.exit()
```
Then the node instance occurs a segmentation fault.
<!--
Enter details about your bug, preferably a simple code snippet that can be
run using `node` directly without installing third-party dependencies.
-->
### How often does it reproduce? Is there a required condition?
This problem can always be triggered following the steps above.
### What is the expected behavior?
If any error occurs, an exception or other similar error-reporting stuff should be thrown. There is no reason to abort the whole node process.
<!--
If possible please provide textual output instead of screenshots.
-->
### What do you see instead?
```
» node
Welcome to Node.js v12.21.0.
Type ".help" for more information.
> fs.write(7,fs.promises.chown,(err, written, buffer)=>{})
undefined
> process.exit()
[1] 160706 segmentation fault (core dumped) /home/zys/Toolchains/node-v12.21.0/node
```
<!--
If possible please provide textual output instead of screenshots.
-->
### Additional information
<!--
Tell us anything else you think we should know.
-->
|
1.0
|
`process.exit` results in a segmentation fault in v12.x - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: v12.21.0
* **Platform**: Linux 5.8.0-38-generic #43~20.04.1-Ubuntu SMP Tue Jan 12 16:39:47 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
* **Subsystem**: process.exit
### What steps will reproduce the bug?
Setup a node instance,
```
» node
```
and run the following javascript code.
```
fs.write(7,fs.promises.chown,(err, written, buffer)=>{})
process.exit()
```
Then the node instance occurs a segmentation fault.
<!--
Enter details about your bug, preferably a simple code snippet that can be
run using `node` directly without installing third-party dependencies.
-->
### How often does it reproduce? Is there a required condition?
This problem can always be triggered following the steps above.
### What is the expected behavior?
If any error occurs, an exception or other similar error-reporting stuff should be thrown. There is no reason to abort the whole node process.
<!--
If possible please provide textual output instead of screenshots.
-->
### What do you see instead?
```
» node
Welcome to Node.js v12.21.0.
Type ".help" for more information.
> fs.write(7,fs.promises.chown,(err, written, buffer)=>{})
undefined
> process.exit()
[1] 160706 segmentation fault (core dumped) /home/zys/Toolchains/node-v12.21.0/node
```
<!--
If possible please provide textual output instead of screenshots.
-->
### Additional information
<!--
Tell us anything else you think we should know.
-->
|
process
|
process exit results in a segmentation fault in x thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or output of osversion foreach object versionstring if else in powershell console windows subsystem if known please specify affected core module name version platform linux generic ubuntu smp tue jan utc gnu linux subsystem process exit what steps will reproduce the bug setup a node instance » node and run the following javascript code fs write fs promises chown err written buffer process exit then the node instance occurs a segmentation fault enter details about your bug preferably a simple code snippet that can be run using node directly without installing third party dependencies how often does it reproduce is there a required condition this problem can always be triggered following the steps above what is the expected behavior if any error occurs an exception or other similar error reporting stuff should be thrown there is no reason to abort the whole node process if possible please provide textual output instead of screenshots what do you see instead » node welcome to node js type help for more information fs write fs promises chown err written buffer undefined process exit segmentation fault core dumped home zys toolchains node node if possible please provide textual output instead of screenshots additional information tell us anything else you think we should know
| 1
|
7,637
| 8,017,064,813
|
IssuesEvent
|
2018-07-25 14:58:56
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
Data Source for Azure Container Registry
|
good first issue new-data-source service/container-registry
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
There is currently no data source for ACR resources. In our case, we share a single ACR instance among several projects, and it would be nice to be able to reference it. The current workaround is to inject the ACR resource ID as a terraform variable.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_container_registry
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_container_registry" "test" {
name = "myregistry"
resource_group_name = "my-resource-group"
}
```
|
1.0
|
Data Source for Azure Container Registry - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
There is currently no data source for ACR resources. In our case, we share a single ACR instance among several projects, and it would be nice to be able to reference it. The current workaround is to inject the ACR resource ID as a terraform variable.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_container_registry
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_container_registry" "test" {
name = "myregistry"
resource_group_name = "my-resource-group"
}
```
|
non_process
|
data source for azure container registry community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description there is currently no data source for acr resources in our case we share a single acr instance among several projects and it would be nice to be able to reference it the current workaround is to inject the acr resource id as a terraform variable new or affected resource s azurerm container registry potential terraform configuration hcl data azurerm container registry test name myregistry resource group name my resource group
| 0
|
14,579
| 17,703,385,488
|
IssuesEvent
|
2021-08-25 02:54:54
|
kevingduck/gh
|
https://api.github.com/repos/kevingduck/gh
|
reopened
|
Big release!
|
New Epic Processed Epic EPIC-9876
|
## Multi-phase Release | 🔗 For linking related issues across release phases (i.e., an epic)
🚧 **NOTICE: This is an experiment. Pay no attention to that man behind the curtain! See https://github.com/github/product-operations/issues/254 for more information.** 🚧
### Each phase of a release (alpha, beta, etc.) should have its own issue created in this repository.
### Use this template to create an issue that links all phases associated with a release.
Fill out the template below to
1. Create a tracking issue with a tracking label (e.g., "Epic-9876")
2. Automatically apply tracking label to all associated issues (e.g., any existing release issues)
3. **automatically generate an issue for each release phase indicated below** (e.g., one for GA, one for Public Beta, and one for Private Beta). (?)
Each issue, including this one, will have a label that links them all together (e.g., "Epic-9876").
### Step 1 of 2: Short description of the project:
This is a big release that will go through all phases, alpha through GA.
### Step 2 of 2: Release phases Expected
**For each expected phase, add an estimated ship date in this table**. **It's OK if the date changes later.** Otherwise, leave the ship date as (YYYY-MM-DD).
If a release issue already exists for any phase, add the link to the table and our bot will automatically apply a tracking label to that issue.
If there's an Estimated Ship Date but no link to an existing issue, our bot will create the release issues for you automatically. (?)
| Phases | Estimated Ship Date | Link to Issue (if it exists) |
|------------------|---------------------|----------------------------- |
| 🌎 GA | 2022-02-10 | (NA)
| 🚀 Public Beta | 2022-01-30 | (NA)
| 🤫 Private Beta | 2021-12-20 | (NA)
| 🌱 Alpha | 2021-10-01 | (NA)
### 🔗 Release Issues Associated with this Issue
#### Add links to any existing issues associated with this release, and our bot will automatically apply the tracking label from this issue.
#### Example: "https://github.com/github/releases/issues/9999"
- https://github.com/kevingduck/gh/issues/60
### Once this issue is created, our bot will reply with a comment to confirm the information you provided. Once you approve it, you can reply and our bot will:
1. create a parent tracking issue and label and apply it to the issue
2. apply the label to any issues indicated in the table above
3. generate a new release issue in this repo for each phase with a ship date and no existing link to an issue. (?)
|
1.0
|
Big release! - ## Multi-phase Release | 🔗 For linking related issues across release phases (i.e., an epic)
🚧 **NOTICE: This is an experiment. Pay no attention to that man behind the curtain! See https://github.com/github/product-operations/issues/254 for more information.** 🚧
### Each phase of a release (alpha, beta, etc.) should have its own issue created in this repository.
### Use this template to create an issue that links all phases associated with a release.
Fill out the template below to
1. Create a tracking issue with a tracking label (e.g., "Epic-9876")
2. Automatically apply tracking label to all associated issues (e.g., any existing release issues)
3. **automatically generate an issue for each release phase indicated below** (e.g., one for GA, one for Public Beta, and one for Private Beta). (?)
Each issue, including this one, will have a label that links them all together (e.g., "Epic-9876").
### Step 1 of 2: Short description of the project:
This is a big release that will go through all phases, alpha through GA.
### Step 2 of 2: Release phases Expected
**For each expected phase, add an estimated ship date in this table**. **It's OK if the date changes later.** Otherwise, leave the ship date as (YYYY-MM-DD).
If a release issue already exists for any phase, add the link to the table and our bot will automatically apply a tracking label to that issue.
If there's an Estimated Ship Date but no link to an existing issue, our bot will create the release issues for you automatically. (?)
| Phases | Estimated Ship Date | Link to Issue (if it exists) |
|------------------|---------------------|----------------------------- |
| 🌎 GA | 2022-02-10 | (NA)
| 🚀 Public Beta | 2022-01-30 | (NA)
| 🤫 Private Beta | 2021-12-20 | (NA)
| 🌱 Alpha | 2021-10-01 | (NA)
### 🔗 Release Issues Associated with this Issue
#### Add links to any existing issues associated with this release, and our bot will automatically apply the tracking label from this issue.
#### Example: "https://github.com/github/releases/issues/9999"
- https://github.com/kevingduck/gh/issues/60
### Once this issue is created, our bot will reply with a comment to confirm the information you provided. Once you approve it, you can reply and our bot will:
1. create a parent tracking issue and label and apply it to the issue
2. apply the label to any issues indicated in the table above
3. generate a new release issue in this repo for each phase with a ship date and no existing link to an issue. (?)
|
process
|
big release multi phase release 🔗 for linking related issues across release phases i e an epic 🚧 notice this is an experiment pay no attention to that man behind the curtain see for more information 🚧 each phase of a release alpha beta etc should have its own issue created in this repository use this template to create an issue that links all phases associated with a release fill out the template below to create a tracking issue with a tracking label e g epic automatically apply tracking label to all associated issues e g any existing release issues automatically generate an issue for each release phase indicated below e g one for ga one for public beta and one for private beta each issue including this one will have a label that links them all together e g epic step of short description of the project this is a big release that will go through all phases alpha through ga step of release phases expected for each expected phase add an estimated ship date in this table it s ok if the date changes later otherwise leave the ship date as yyyy mm dd if a release issue already exists for any phase add the link to the table and our bot will automatically apply a tracking label to that issue if there s an estimated ship date but no link to an existing issue our bot will create the release issues for you automatically phases estimated ship date link to issue if it exists 🌎 ga na 🚀 public beta na 🤫 private beta na 🌱 alpha na 🔗 release issues associated with this issue add links to any existing issues associated with this release and our bot will automatically apply the tracking label from this issue example once this issue is created our bot will reply with a comment to confirm the information you provided once you approve it you can reply and our bot will create a parent tracking issue and label and apply it to the issue apply the label to any issues indicated in the table above generate a new release issue in this repo for each phase with a ship date and no existing link to an issue
| 1
|
5,186
| 7,736,148,623
|
IssuesEvent
|
2018-05-27 23:06:38
|
lgmagalhaes88/cms-app
|
https://api.github.com/repos/lgmagalhaes88/cms-app
|
closed
|
Tela upload de arquivo
|
requirement template
|
Criar tela de upload de arquivo .txt com no formato: `nome_da_disciplina nome_do_professor período`.
Enviando-o para o servidor backend
Após o processamento e retorno das informações, é exibida um grid organizado pelos dias da semana e pelos períodos, informando as aulas do primeiro e segundo horário.
|
1.0
|
Tela upload de arquivo - Criar tela de upload de arquivo .txt com no formato: `nome_da_disciplina nome_do_professor período`.
Enviando-o para o servidor backend
Após o processamento e retorno das informações, é exibida um grid organizado pelos dias da semana e pelos períodos, informando as aulas do primeiro e segundo horário.
|
non_process
|
tela upload de arquivo criar tela de upload de arquivo txt com no formato nome da disciplina nome do professor período enviando o para o servidor backend após o processamento e retorno das informações é exibida um grid organizado pelos dias da semana e pelos períodos informando as aulas do primeiro e segundo horário
| 0
|
449,271
| 31,836,382,914
|
IssuesEvent
|
2023-09-14 13:46:01
|
vatesfr/xen-orchestra
|
https://api.github.com/repos/vatesfr/xen-orchestra
|
closed
|
Unable to stop backup
|
type: enhancement :sparkles: comp: documentation :book: Priority 3: can wait/workaround :yellow_circle: status::in backlog :white_check_mark:
|
**XOA or XO from the sources?**
XO from sources.
**Describe the bug**
A clear and concise description of what the bug is.
My backup job is stuck ( another problem ;-) )
I can't stop it !
<img width="265" alt="image" src="https://user-images.githubusercontent.com/4645967/167281207-eee77794-0efc-4799-8ab5-7508c0741716.png">
**To Reproduce**
Steps to reproduce the behavior:
1. Go to bacups
2. Have a stuck job ( ... )
3. You can't kill the operation
**Expected behavior**
A clear and concise description of what you expected to happen.
I would like to be able to click on stop backup
Instead, i have to restart xo
|
1.0
|
Unable to stop backup - **XOA or XO from the sources?**
XO from sources.
**Describe the bug**
A clear and concise description of what the bug is.
My backup job is stuck ( another problem ;-) )
I can't stop it !
<img width="265" alt="image" src="https://user-images.githubusercontent.com/4645967/167281207-eee77794-0efc-4799-8ab5-7508c0741716.png">
**To Reproduce**
Steps to reproduce the behavior:
1. Go to bacups
2. Have a stuck job ( ... )
3. You can't kill the operation
**Expected behavior**
A clear and concise description of what you expected to happen.
I would like to be able to click on stop backup
Instead, i have to restart xo
|
non_process
|
unable to stop backup xoa or xo from the sources xo from sources describe the bug a clear and concise description of what the bug is my backup job is stuck another problem i can t stop it img width alt image src to reproduce steps to reproduce the behavior go to bacups have a stuck job you can t kill the operation expected behavior a clear and concise description of what you expected to happen i would like to be able to click on stop backup instead i have to restart xo
| 0
|
102,272
| 8,823,836,572
|
IssuesEvent
|
2019-01-02 15:04:04
|
dwyl/hq
|
https://api.github.com/repos/dwyl/hq
|
closed
|
What is the company PAYE reference number? (for tax return purposes)
|
please-test question
|
I need to know:
What is the company PAYE reference number? (for tax return purposes)
I'm not sure if this information is sensitive so I don't mind if you don't respond here. Apparently they're usually on your payslips but I can't see ours on there.
|
1.0
|
What is the company PAYE reference number? (for tax return purposes) - I need to know:
What is the company PAYE reference number? (for tax return purposes)
I'm not sure if this information is sensitive so I don't mind if you don't respond here. Apparently they're usually on your payslips but I can't see ours on there.
|
non_process
|
what is the company paye reference number for tax return purposes i need to know what is the company paye reference number for tax return purposes i m not sure if this information is sensitive so i don t mind if you don t respond here apparently they re usually on your payslips but i can t see ours on there
| 0
|
14,819
| 18,156,878,427
|
IssuesEvent
|
2021-09-27 03:37:36
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
Bazel cannot generate google-cloud-asset-v1 because of external proto references
|
type: process
|
The google-cloud-asset-v1 library is still being generated via the docker release of gapic-generator-ruby. This is because asset-v1 references a bunch of protos, specifically:
* google/cloud/orgpolicy/v1/orgpolicy.proto
* google/cloud/osconfig/v1/inventory.proto
* google/identity/accesscontextmanager/type/device_resources.proto
* google/identity/accesscontextmanager/v1/access_level.proto
* google/identity/accesscontextmanager/v1/access_policy.proto
* google/identity/accesscontextmanager/v1/service_perimeter.proto
The docker-based synth script generates these protos inline directly in the library. The bazel config lists them merely as dependencies, meaning it expects the generated library to be linked dynamically to those protos (i.e. as dependencies). Since the bazel-generated library does not have those other protos as dependencies, it is unable to load. See e.g. #12538.
To solve this, for orgpolicy and osconfig, we could potentially handle this by adding dependencies on other clients. However, we do not have a library for accesscontextmanager, and it is blocked by internal issue b/187308446.
|
1.0
|
Bazel cannot generate google-cloud-asset-v1 because of external proto references - The google-cloud-asset-v1 library is still being generated via the docker release of gapic-generator-ruby. This is because asset-v1 references a bunch of protos, specifically:
* google/cloud/orgpolicy/v1/orgpolicy.proto
* google/cloud/osconfig/v1/inventory.proto
* google/identity/accesscontextmanager/type/device_resources.proto
* google/identity/accesscontextmanager/v1/access_level.proto
* google/identity/accesscontextmanager/v1/access_policy.proto
* google/identity/accesscontextmanager/v1/service_perimeter.proto
The docker-based synth script generates these protos inline directly in the library. The bazel config lists them merely as dependencies, meaning it expects the generated library to be linked dynamically to those protos (i.e. as dependencies). Since the bazel-generated library does not have those other protos as dependencies, it is unable to load. See e.g. #12538.
To solve this, for orgpolicy and osconfig, we could potentially handle this by adding dependencies on other clients. However, we do not have a library for accesscontextmanager, and it is blocked by internal issue b/187308446.
|
process
|
bazel cannot generate google cloud asset because of external proto references the google cloud asset library is still being generated via the docker release of gapic generator ruby this is because asset references a bunch of protos specifically google cloud orgpolicy orgpolicy proto google cloud osconfig inventory proto google identity accesscontextmanager type device resources proto google identity accesscontextmanager access level proto google identity accesscontextmanager access policy proto google identity accesscontextmanager service perimeter proto the docker based synth script generates these protos inline directly in the library the bazel config lists them merely as dependencies meaning it expects the generated library to be linked dynamically to those protos i e as dependencies since the bazel generated library does not have those other protos as dependencies it is unable to load see e g to solve this for orgpolicy and osconfig we could potentially handle this by adding dependencies on other clients however we do not have a library for accesscontextmanager and it is blocked by internal issue b
| 1
|
83,705
| 3,640,878,098
|
IssuesEvent
|
2016-02-13 06:27:45
|
jerradgenson/auteur
|
https://api.github.com/repos/jerradgenson/auteur
|
opened
|
RSS URLs reference index.html
|
enhancement low priority
|
RSS feed URLs should reference article directories instead of referencing their index.html files directly.
|
1.0
|
RSS URLs reference index.html - RSS feed URLs should reference article directories instead of referencing their index.html files directly.
|
non_process
|
rss urls reference index html rss feed urls should reference article directories instead of referencing their index html files directly
| 0
|
7,548
| 10,674,335,429
|
IssuesEvent
|
2019-10-21 09:14:08
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Too many term in immune signalling
|
multi-species process
|
all of these terms must represent a gene product upstream of the immune response signalling and so, essentially are all the same process.
GO:0035419 activation of MAPK activity involved in innate immune response | part_of
-- | --
GO:0035422 activation of MAPKKK activity involved in innate immune response | is_a
GO:0002220 innate immune response activating cell surface receptor signaling pathway | is_a
GO:0035421 activation of MAPKK activity involved in innate immune response | is_a
GO:0002221 pattern recognition receptor signaling pathway
I think we already got rid of most of these processes from other MAPK signalling?
|
1.0
|
Too many term in immune signalling -
all of these terms must represent a gene product upstream of the immune response signalling and so, essentially are all the same process.
GO:0035419 activation of MAPK activity involved in innate immune response | part_of
-- | --
GO:0035422 activation of MAPKKK activity involved in innate immune response | is_a
GO:0002220 innate immune response activating cell surface receptor signaling pathway | is_a
GO:0035421 activation of MAPKK activity involved in innate immune response | is_a
GO:0002221 pattern recognition receptor signaling pathway
I think we already got rid of most of these processes from other MAPK signalling?
|
process
|
too many term in immune signalling all of these terms must represent a gene product upstream of the immune response signalling and so essentially are all the same process go activation of mapk activity involved in innate immune response part of go activation of mapkkk activity involved in innate immune response is a go innate immune response activating cell surface receptor signaling pathway is a go activation of mapkk activity involved in innate immune response is a go pattern recognition receptor signaling pathway i think we already got rid of most of these processes from other mapk signalling
| 1
|
4,139
| 7,094,784,418
|
IssuesEvent
|
2018-01-13 08:32:06
|
triplea-game/triplea
|
https://api.github.com/repos/triplea-game/triplea
|
opened
|
Label Changes
|
category: dev & admin process discussion
|
Quick FYI and discussion for recent label updates. Some changes:
- rename p0 + p1 to "production - p0" and "production p1"
- regression renamed back to 'release blocker'
- type + category merged, just now 'category.*' labels, no 'type.*:' labels.
3rd is a bit more controversial, but it was not clear that having two dimensions for labeling was that useful. It was really not clear what defined those dimensions to begin with. With the labels all blended to 'category', it may be a bit more obvious how to group things cleanly. I think it's okay for issues to get multiple category labels (but too much overlap on too many issues may mean the categories are not very well defined).
|
1.0
|
Label Changes - Quick FYI and discussion for recent label updates. Some changes:
- rename p0 + p1 to "production - p0" and "production p1"
- regression renamed back to 'release blocker'
- type + category merged, just now 'category.*' labels, no 'type.*:' labels.
3rd is a bit more controversial, but it was not clear that having two dimensions for labeling was that useful. It was really not clear what defined those dimensions to begin with. With the labels all blended to 'category', it may be a bit more obvious how to group things cleanly. I think it's okay for issues to get multiple category labels (but too much overlap on too many issues may mean the categories are not very well defined).
|
process
|
label changes quick fyi and discussion for recent label updates some changes rename to production and production regression renamed back to release blocker type category merged just now category labels no type labels is a bit more controversial but it was not clear that having two dimensions for labeling was that useful it was really not clear what defined those dimensions to begin with with the labels all blended to category it may be a bit more obvious how to group things cleanly i think it s okay for issues to get multiple category labels but too much overlap on too many issues may mean the categories are not very well defined
| 1
|
19,916
| 26,378,759,772
|
IssuesEvent
|
2023-01-12 06:27:39
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Remote download minimal will download all input files for symlink
|
more data needed type: support / not a bug (process) team-Remote-Exec
|
### Description of the bug:
Symlink inputs(files they point to) are downloaded to host when remote_download_minimal is set.
Because C++ shared libraries are all symlinked to a solib directory, this causes all shared libraries to be downloaded.
I would expect them not to be downloaded. During execution, the actual file may be referenced by their digest.
``` at java.util.concurrent.locks.LockSupport.park(java.base@11.0.6/Unknown Source)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:537)
at com.google.devtools.build.lib.remote.util.Utils.getFromFuture(Utils.java:106)
at com.google.devtools.build.lib.remote.util.Utils.getFromFuture(Utils.java:94)
at com.google.devtools.build.lib.remote.RemoteActionInputFetcher.downloadFile(RemoteActionInputFetcher.java:149)
at com.google.devtools.build.lib.remote.RemoteActionFileSystem.downloadFileIfRemote(RemoteActionFileSystem.java:340)
at com.google.devtools.build.lib.remote.RemoteActionFileSystem.createSymbolicLink(RemoteActionFileSystem.java:194)
at com.google.devtools.build.lib.vfs.Path.createSymbolicLink(Path.java:489)
at com.google.devtools.build.lib.rules.cpp.SolibSymlinkAction.execute(SolibSymlinkAction.java:80)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169)
at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590)
at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382)
at java.util.concurrent.ThreadPoolExecutor.runWorker(java.base@11.0.6/Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(java.base@11.0.6/Unknown Source)
```
Is a stacktrace for the download.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Build C++ shared library with `--remote_download_minimal`
### Which operating system are you running Bazel on?
linux
### What is the output of `bazel info release`?
5.2.0
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
https://github.com/bazelbuild/bazel/issues/11119 seems indicate the symlink could be cached.
### Any other information, logs, or outputs that you want to share?
_No response_
|
1.0
|
Remote download minimal will download all input files for symlink - ### Description of the bug:
Symlink inputs(files they point to) are downloaded to host when remote_download_minimal is set.
Because C++ shared libraries are all symlinked to a solib directory, this causes all shared libraries to be downloaded.
I would expect them not to be downloaded. During execution, the actual file may be referenced by their digest.
``` at java.util.concurrent.locks.LockSupport.park(java.base@11.0.6/Unknown Source)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:537)
at com.google.devtools.build.lib.remote.util.Utils.getFromFuture(Utils.java:106)
at com.google.devtools.build.lib.remote.util.Utils.getFromFuture(Utils.java:94)
at com.google.devtools.build.lib.remote.RemoteActionInputFetcher.downloadFile(RemoteActionInputFetcher.java:149)
at com.google.devtools.build.lib.remote.RemoteActionFileSystem.downloadFileIfRemote(RemoteActionFileSystem.java:340)
at com.google.devtools.build.lib.remote.RemoteActionFileSystem.createSymbolicLink(RemoteActionFileSystem.java:194)
at com.google.devtools.build.lib.vfs.Path.createSymbolicLink(Path.java:489)
at com.google.devtools.build.lib.rules.cpp.SolibSymlinkAction.execute(SolibSymlinkAction.java:80)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169)
at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590)
at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382)
at java.util.concurrent.ThreadPoolExecutor.runWorker(java.base@11.0.6/Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(java.base@11.0.6/Unknown Source)
```
Is a stacktrace for the download.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Build C++ shared library with `--remote_download_minimal`
### Which operating system are you running Bazel on?
linux
### What is the output of `bazel info release`?
5.2.0
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
https://github.com/bazelbuild/bazel/issues/11119 seems indicate the symlink could be cached.
### Any other information, logs, or outputs that you want to share?
_No response_
|
process
|
remote download minimal will download all input files for symlink description of the bug symlink inputs files they point to are downloaded to host when remote download minimal is set because c shared libraries are all symlinked to a solib directory this causes all shared libraries to be downloaded i would expect them not to be downloaded during execution the actual file may be referenced by their digest at java util concurrent locks locksupport park java base unknown source at com google common util concurrent abstractfuture get abstractfuture java at com google devtools build lib remote util utils getfromfuture utils java at com google devtools build lib remote util utils getfromfuture utils java at com google devtools build lib remote remoteactioninputfetcher downloadfile remoteactioninputfetcher java at com google devtools build lib remote remoteactionfilesystem downloadfileifremote remoteactionfilesystem java at com google devtools build lib remote remoteactionfilesystem createsymboliclink remoteactionfilesystem java at com google devtools build lib vfs path createsymboliclink path java at com google devtools build lib rules cpp solibsymlinkaction execute solibsymlinkaction java at com google devtools build lib skyframe skyframeactionexecutor execute skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner continueaction skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner run skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionstate runstatemachine actionexecutionstate java at com google devtools build lib skyframe actionexecutionstate getresultordependonfuture actionexecutionstate java at com google devtools build lib skyframe skyframeactionexecutor executeaction skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionfunction checkcacheandexecuteifneeded actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction computeinternal actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction compute actionexecutionfunction java at com google devtools build skyframe abstractparallelevaluator evaluate run abstractparallelevaluator java at com google devtools build lib concurrent abstractqueuevisitor wrappedrunnable run abstractqueuevisitor java at java util concurrent threadpoolexecutor runworker java base unknown source at java util concurrent threadpoolexecutor worker run java base unknown source is a stacktrace for the download what s the simplest easiest way to reproduce this bug please provide a minimal example if possible build c shared library with remote download minimal which operating system are you running bazel on linux what is the output of bazel info release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web seems indicate the symlink could be cached any other information logs or outputs that you want to share no response
| 1
|
5,216
| 8,007,146,601
|
IssuesEvent
|
2018-07-24 00:36:08
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
opened
|
Release 'datastore 1.7.0'
|
api: datastore packaging type: process
|
Major changes are:
- Add support for Python 3.7.
- Drop support for Python 3.4.
- Bugfix: query offsets (#4675).
|
1.0
|
Release 'datastore 1.7.0' - Major changes are:
- Add support for Python 3.7.
- Drop support for Python 3.4.
- Bugfix: query offsets (#4675).
|
process
|
release datastore major changes are add support for python drop support for python bugfix query offsets
| 1
|
366,137
| 25,570,534,986
|
IssuesEvent
|
2022-11-30 17:18:54
|
redpanda-data/documentation
|
https://api.github.com/repos/redpanda-data/documentation
|
opened
|
Create a page to explain what is Redpanda's versioning scheme
|
documentation
|
### Describe the Issue
During the execution of this [PR](https://github.com/redpanda-data/documentation/pull/859) we mention that users should only upgrade one major version at a time. But we never clearly inform what constitutes a major version for us.
So, ideally we should create a page to clearly explain to our users what is our versioning scheme. Essentially this page should have:
- Why do we number our versions like that.
- What do we define as a major version release.
- What is a minor version.
- Why we don't have specific docs for minor versions.
- And any other subjects relative to versions, like for example, if we respect common versioning schemes like [SemVer](https://semver.org/)
### New feature or content gap requests
- Gap
|
1.0
|
Create a page to explain what is Redpanda's versioning scheme - ### Describe the Issue
During the execution of this [PR](https://github.com/redpanda-data/documentation/pull/859) we mention that users should only upgrade one major version at a time. But we never clearly inform what constitutes a major version for us.
So, ideally we should create a page to clearly explain to our users what is our versioning scheme. Essentially this page should have:
- Why do we number our versions like that.
- What do we define as a major version release.
- What is a minor version.
- Why we don't have specific docs for minor versions.
- And any other subjects relative to versions, like for example, if we respect common versioning schemes like [SemVer](https://semver.org/)
### New feature or content gap requests
- Gap
|
non_process
|
create a page to explain what is redpanda s versioning scheme describe the issue during the execution of this we mention that users should only upgrade one major version at a time but we never clearly inform what constitutes a major version for us so ideally we should create a page to clearly explain to our users what is our versioning scheme essentially this page should have why do we number our versions like that what do we define as a major version release what is a minor version why we don t have specific docs for minor versions and any other subjects relative to versions like for example if we respect common versioning schemes like new feature or content gap requests gap
| 0
|
417
| 2,852,496,369
|
IssuesEvent
|
2015-06-01 13:56:29
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
closed
|
Fix or delete ignored tests
|
Data Storage enhancement Processing
|
I've disabled a bunch of tests that were either erroneous, depended on hardcoded data not included in the repo, were too slow or couldn't be easily fixed. Please take a look and either delete or re-enable them.
Full list of disabled tests:
- [x] `test/server/test/ServerGenomeReleaseTest.java`
- [x] `test/server/test/ServerChainTestClass.java`
- [x] `test/server/test/ServerLoginLogoutTest.java`
- [x] `test/server/test/ServerAnnotationTest.java`
- [x] `test/response/test/GetGenomeReleaseResponseTest.java`
- [x] `test/response/test/AddGenomeReleaseResponseTest.java`
- [x] `test/conversion/test/ProfileDataConverterTest.java`
- [x] `test/database/test/unittests/FileNameValidatorTest.java`
- [x] `test/database/test/unittests/DatabaseAccessorThreadTests.java`
- [x] `test/command/test/GetProcessStatusCommandTest.java`
- [x] `test/command/test/ProcessCommandTest.java`
- [x] `test/response/test/LoginResponseTest.java`
|
1.0
|
Fix or delete ignored tests - I've disabled a bunch of tests that were either erroneous, depended on hardcoded data not included in the repo, were too slow or couldn't be easily fixed. Please take a look and either delete or re-enable them.
Full list of disabled tests:
- [x] `test/server/test/ServerGenomeReleaseTest.java`
- [x] `test/server/test/ServerChainTestClass.java`
- [x] `test/server/test/ServerLoginLogoutTest.java`
- [x] `test/server/test/ServerAnnotationTest.java`
- [x] `test/response/test/GetGenomeReleaseResponseTest.java`
- [x] `test/response/test/AddGenomeReleaseResponseTest.java`
- [x] `test/conversion/test/ProfileDataConverterTest.java`
- [x] `test/database/test/unittests/FileNameValidatorTest.java`
- [x] `test/database/test/unittests/DatabaseAccessorThreadTests.java`
- [x] `test/command/test/GetProcessStatusCommandTest.java`
- [x] `test/command/test/ProcessCommandTest.java`
- [x] `test/response/test/LoginResponseTest.java`
|
process
|
fix or delete ignored tests i ve disabled a bunch of tests that were either erroneous depended on hardcoded data not included in the repo were too slow or couldn t be easily fixed please take a look and either delete or re enable them full list of disabled tests test server test servergenomereleasetest java test server test serverchaintestclass java test server test serverloginlogouttest java test server test serverannotationtest java test response test getgenomereleaseresponsetest java test response test addgenomereleaseresponsetest java test conversion test profiledataconvertertest java test database test unittests filenamevalidatortest java test database test unittests databaseaccessorthreadtests java test command test getprocessstatuscommandtest java test command test processcommandtest java test response test loginresponsetest java
| 1
|
37,148
| 9,968,128,604
|
IssuesEvent
|
2019-07-08 14:57:48
|
GoogleChrome/workbox
|
https://api.github.com/repos/GoogleChrome/workbox
|
opened
|
Minimum required version of node should be v8
|
Breaking Change workbox-build workbox-cli workbox-webpack-plugin
|
As per https://nodejs.org/en/about/releases/, the earliest version of node that's still in the maintenance window is v8.x.
In the Workbox v5 release, we should switch out minimum required version of node to match, and update our `babel-preset-env` transpilation accordingly for the build tools.
Additionally, we should set up Travis CI to explicitly test against both node v8 and the latest node release (see #2091).
|
1.0
|
Minimum required version of node should be v8 - As per https://nodejs.org/en/about/releases/, the earliest version of node that's still in the maintenance window is v8.x.
In the Workbox v5 release, we should switch out minimum required version of node to match, and update our `babel-preset-env` transpilation accordingly for the build tools.
Additionally, we should set up Travis CI to explicitly test against both node v8 and the latest node release (see #2091).
|
non_process
|
minimum required version of node should be as per the earliest version of node that s still in the maintenance window is x in the workbox release we should switch out minimum required version of node to match and update our babel preset env transpilation accordingly for the build tools additionally we should set up travis ci to explicitly test against both node and the latest node release see
| 0
|
271,429
| 23,603,979,815
|
IssuesEvent
|
2022-08-24 06:29:14
|
MrBrax/LiveStreamDVR
|
https://api.github.com/repos/MrBrax/LiveStreamDVR
|
closed
|
Force resubscribe
|
needs testing
|
When pressing subscribe on the channel settings page, even if the subscription was deleted from the "about" page, then nothing happens.
Log just says this:
```
web_1 | 2022-06-06 19:19:50.004 | helper <INFO> Skip subscription to 23161357:stream.online (lirik), in cache.
web_1 | 2022-06-06 19:19:50.005 | helper <INFO> Skip subscription to 23161357:stream.offline (lirik), in cache.
web_1 | 2022-06-06 19:19:50.005 | helper <INFO> Skip subscription to 23161357:channel.update (lirik), in cache.
```
In order to subscribe again, the keys mentioned in the log must be deleted manually in the KeyValue-page. This should be automatic whenever the subscribe button is pressed.
|
1.0
|
Force resubscribe - When pressing subscribe on the channel settings page, even if the subscription was deleted from the "about" page, then nothing happens.
Log just says this:
```
web_1 | 2022-06-06 19:19:50.004 | helper <INFO> Skip subscription to 23161357:stream.online (lirik), in cache.
web_1 | 2022-06-06 19:19:50.005 | helper <INFO> Skip subscription to 23161357:stream.offline (lirik), in cache.
web_1 | 2022-06-06 19:19:50.005 | helper <INFO> Skip subscription to 23161357:channel.update (lirik), in cache.
```
In order to subscribe again, the keys mentioned in the log must be deleted manually in the KeyValue-page. This should be automatic whenever the subscribe button is pressed.
|
non_process
|
force resubscribe when pressing subscribe on the channel settings page even if the subscription was deleted from the about page then nothing happens log just says this web helper skip subscription to stream online lirik in cache web helper skip subscription to stream offline lirik in cache web helper skip subscription to channel update lirik in cache in order to subscribe again the keys mentioned in the log must be deleted manually in the keyvalue page this should be automatic whenever the subscribe button is pressed
| 0
|
10,170
| 13,044,162,738
|
IssuesEvent
|
2020-07-29 03:47:35
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `GetParamString` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `GetParamString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `GetParamString` from TiDB -
## Description
Port the scalar function `GetParamString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function getparamstring from tidb description port the scalar function getparamstring from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
726,907
| 25,015,682,899
|
IssuesEvent
|
2022-11-03 18:35:35
|
space-wizards/space-station-14
|
https://api.github.com/repos/space-wizards/space-station-14
|
closed
|
Dragons are OMEGAfucked
|
Issue: Bug Priority: 2-Before Release Difficulty: 1-Easy
|
I spawned one on live, and devour action simply didn't work. This absolutely cucks the dragon as it's 50% of what it does.
Their damage to windows/walls is also very low, I know you're supposed to be able to devour those instead of attacking them, but maybe it should be bumped up just a tiny bit?
|
1.0
|
Dragons are OMEGAfucked - I spawned one on live, and devour action simply didn't work. This absolutely cucks the dragon as it's 50% of what it does.
Their damage to windows/walls is also very low, I know you're supposed to be able to devour those instead of attacking them, but maybe it should be bumped up just a tiny bit?
|
non_process
|
dragons are omegafucked i spawned one on live and devour action simply didn t work this absolutely cucks the dragon as it s of what it does their damage to windows walls is also very low i know you re supposed to be able to devour those instead of attacking them but maybe it should be bumped up just a tiny bit
| 0
|
3,929
| 6,848,433,191
|
IssuesEvent
|
2017-11-13 18:30:15
|
syndesisio/syndesis-ui
|
https://api.github.com/repos/syndesisio/syndesis-ui
|
closed
|
Update the config.json in development mode so it's obvious you're running against the development server
|
dev process Priority - High
|
We should change the app name or otherwise update the config.json so that it's quick and easy to see that you're looking at the development server, saves having to go make some trivial change to check.
|
1.0
|
Update the config.json in development mode so it's obvious you're running against the development server - We should change the app name or otherwise update the config.json so that it's quick and easy to see that you're looking at the development server, saves having to go make some trivial change to check.
|
process
|
update the config json in development mode so it s obvious you re running against the development server we should change the app name or otherwise update the config json so that it s quick and easy to see that you re looking at the development server saves having to go make some trivial change to check
| 1
|
13,490
| 16,018,572,081
|
IssuesEvent
|
2021-04-20 19:18:18
|
anlsys/aml
|
https://api.github.com/repos/anlsys/aml
|
opened
|
Integration Tests
|
process:proposal
|
In GitLab by @NicolasDenoyelle on Mar 15, 2021, 10:42
Create a pipeline stage with integration tests.
For instance, the test could consist in:
- Fetch applications using AML.
- Build AML
- Build applications and link with AML.
- Run applications.
[XSBench](https://github.com/ANL-CESAR/XSBench) is an application using AML that could be used as an integration test.
|
1.0
|
Integration Tests - In GitLab by @NicolasDenoyelle on Mar 15, 2021, 10:42
Create a pipeline stage with integration tests.
For instance, the test could consist in:
- Fetch applications using AML.
- Build AML
- Build applications and link with AML.
- Run applications.
[XSBench](https://github.com/ANL-CESAR/XSBench) is an application using AML that could be used as an integration test.
|
process
|
integration tests in gitlab by nicolasdenoyelle on mar create a pipeline stage with integration tests for instance the test could consist in fetch applications using aml build aml build applications and link with aml run applications is an application using aml that could be used as an integration test
| 1
|
19,507
| 25,818,942,330
|
IssuesEvent
|
2022-12-12 08:05:15
|
googleapis/ruby-style
|
https://api.github.com/repos/googleapis/ruby-style
|
reopened
|
Dependency Dashboard
|
type: process
|
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
This repository currently has no open or pending branches.
## Detected dependencies
<details><summary>bundler</summary>
<blockquote>
<details><summary>Gemfile</summary>
</details>
</blockquote>
</details>
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/ci.yml</summary>
- `actions/checkout v3`
- `ruby/setup-ruby v1`
</details>
<details><summary>.github/workflows/release-please-label.yml</summary>
- `actions/github-script v6`
</details>
<details><summary>.github/workflows/release-please.yml</summary>
- `actions/checkout v3`
- `ruby/setup-ruby v1`
- `actions/setup-node v3`
</details>
</blockquote>
</details>
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
This repository currently has no open or pending branches.
## Detected dependencies
<details><summary>bundler</summary>
<blockquote>
<details><summary>Gemfile</summary>
</details>
</blockquote>
</details>
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/ci.yml</summary>
- `actions/checkout v3`
- `ruby/setup-ruby v1`
</details>
<details><summary>.github/workflows/release-please-label.yml</summary>
- `actions/github-script v6`
</details>
<details><summary>.github/workflows/release-please.yml</summary>
- `actions/checkout v3`
- `ruby/setup-ruby v1`
- `actions/setup-node v3`
</details>
</blockquote>
</details>
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more this repository currently has no open or pending branches detected dependencies bundler gemfile github actions github workflows ci yml actions checkout ruby setup ruby github workflows release please label yml actions github script github workflows release please yml actions checkout ruby setup ruby actions setup node check this box to trigger a request for renovate to run again on this repository
| 1
|
14,497
| 17,604,292,631
|
IssuesEvent
|
2021-08-17 15:13:32
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][processing] Export layers information algorithm (Request in QGIS)
|
Processing Alg 3.18
|
### Request for documentation
From pull request QGIS/qgis#40462
Author: @nirvn
QGIS version: 3.18
**[FEATURE][processing] Export layers information algorithm**
### PR Description:
## Description
This PR adds a new algorithm to QGIS' processing toolbox named export layers information. The algorithm creates a polygon layer corresponding to extent of selected layer(s) with additional layer details - CRS, provider name, file path, layer name, subset filer, abstract and attribution - attached as attributes to each polygon feature.
Test cases are still needed, which I will add after the PR is reviewed in case additional attributes are requested during the review process.
### Commits tagged with [need-docs] or [FEATURE]
"[FEATURE][processing] Export layers information algorithm"
|
1.0
|
[FEATURE][processing] Export layers information algorithm (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#40462
Author: @nirvn
QGIS version: 3.18
**[FEATURE][processing] Export layers information algorithm**
### PR Description:
## Description
This PR adds a new algorithm to QGIS' processing toolbox named export layers information. The algorithm creates a polygon layer corresponding to extent of selected layer(s) with additional layer details - CRS, provider name, file path, layer name, subset filer, abstract and attribution - attached as attributes to each polygon feature.
Test cases are still needed, which I will add after the PR is reviewed in case additional attributes are requested during the review process.
### Commits tagged with [need-docs] or [FEATURE]
"[FEATURE][processing] Export layers information algorithm"
|
process
|
export layers information algorithm request in qgis request for documentation from pull request qgis qgis author nirvn qgis version export layers information algorithm pr description description this pr adds a new algorithm to qgis processing toolbox named export layers information the algorithm creates a polygon layer corresponding to extent of selected layer s with additional layer details crs provider name file path layer name subset filer abstract and attribution attached as attributes to each polygon feature test cases are still needed which i will add after the pr is reviewed in case additional attributes are requested during the review process commits tagged with or export layers information algorithm
| 1
|
40,741
| 10,141,071,553
|
IssuesEvent
|
2019-08-03 10:30:43
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
closed
|
install error: libparcel_coalescing.so.0.9.11 missing
|
category: CMake tag: wontfix type: defect
|
I tried to install HPX locally on a Kubuntu system using
```
cmake -DCMAKE_INSTALL_PREFIX=/home/mario/hpx_gcc ../hpx
make -j4
make install
```
and I get the following:
```
CMake Error at cmake_install.cmake:48 (FILE):
file INSTALL cannot find
"/home/mario/hpx_gcc/lib/hpx/libparcel_coalescing.so.0.9.11".
```
Indeed I see:
```
mario@WorkBook: ~/hpx_gcc
$ l /home/mario/hpx_gcc/lib/hpx/ [18:42:58]
total 2.2M
drwxrwxr-x 2 mario mario 4.0K May 19 18:42 .
drwxrwxr-x 5 mario mario 4.0K May 19 18:11 ..
lrwxrwxrwx 1 mario mario 14 May 19 17:03 libmemory.so -> libmemory.so.0
lrwxrwxrwx 1 mario mario 19 May 19 17:03 libmemory.so.0 -> libmemory.so.0.9.11
-rwxrwxr-x 1 mario mario 1.1M May 19 17:03 libmemory.so.0.9.11
lrwxrwxrwx 1 mario mario 25 May 19 18:42 libparcel_coalescing.so -> libparcel_coalescing.so.0
lrwxrwxrwx 1 mario mario 30 May 19 18:42 libparcel_coalescing.so.0 -> libparcel_coalescing.so.0.9.11
lrwxrwxrwx 1 mario mario 12 May 19 17:07 libsine.so -> libsine.so.0
lrwxrwxrwx 1 mario mario 17 May 19 17:07 libsine.so.0 -> libsine.so.0.9.11
-rwxrwxr-x 1 mario mario 1.2M May 19 17:07 libsine.so.0.9.11
```
so `libparcel_coalescing.so.0.9.11` appears to be missing?
make tests gives the following errors
```
The following tests FAILED:
66 - tests.unit.actions.succeed_compile_const_ref_argument (Failed)
83 - tests.unit.components.distributed.tcp.action_invoke_no_more_than (Timeout)
157 - tests.unit.threads.thread_affinity (Timeout)
Errors while running CTest
```
which might be a separate issue?
|
1.0
|
install error: libparcel_coalescing.so.0.9.11 missing - I tried to install HPX locally on a Kubuntu system using
```
cmake -DCMAKE_INSTALL_PREFIX=/home/mario/hpx_gcc ../hpx
make -j4
make install
```
and I get the following:
```
CMake Error at cmake_install.cmake:48 (FILE):
file INSTALL cannot find
"/home/mario/hpx_gcc/lib/hpx/libparcel_coalescing.so.0.9.11".
```
Indeed I see:
```
mario@WorkBook: ~/hpx_gcc
$ l /home/mario/hpx_gcc/lib/hpx/ [18:42:58]
total 2.2M
drwxrwxr-x 2 mario mario 4.0K May 19 18:42 .
drwxrwxr-x 5 mario mario 4.0K May 19 18:11 ..
lrwxrwxrwx 1 mario mario 14 May 19 17:03 libmemory.so -> libmemory.so.0
lrwxrwxrwx 1 mario mario 19 May 19 17:03 libmemory.so.0 -> libmemory.so.0.9.11
-rwxrwxr-x 1 mario mario 1.1M May 19 17:03 libmemory.so.0.9.11
lrwxrwxrwx 1 mario mario 25 May 19 18:42 libparcel_coalescing.so -> libparcel_coalescing.so.0
lrwxrwxrwx 1 mario mario 30 May 19 18:42 libparcel_coalescing.so.0 -> libparcel_coalescing.so.0.9.11
lrwxrwxrwx 1 mario mario 12 May 19 17:07 libsine.so -> libsine.so.0
lrwxrwxrwx 1 mario mario 17 May 19 17:07 libsine.so.0 -> libsine.so.0.9.11
-rwxrwxr-x 1 mario mario 1.2M May 19 17:07 libsine.so.0.9.11
```
so `libparcel_coalescing.so.0.9.11` appears to be missing?
make tests gives the following errors
```
The following tests FAILED:
66 - tests.unit.actions.succeed_compile_const_ref_argument (Failed)
83 - tests.unit.components.distributed.tcp.action_invoke_no_more_than (Timeout)
157 - tests.unit.threads.thread_affinity (Timeout)
Errors while running CTest
```
which might be a separate issue?
|
non_process
|
install error libparcel coalescing so missing i tried to install hpx locally on a kubuntu system using cmake dcmake install prefix home mario hpx gcc hpx make make install and i get the following cmake error at cmake install cmake file file install cannot find home mario hpx gcc lib hpx libparcel coalescing so indeed i see mario workbook hpx gcc l home mario hpx gcc lib hpx total drwxrwxr x mario mario may drwxrwxr x mario mario may lrwxrwxrwx mario mario may libmemory so libmemory so lrwxrwxrwx mario mario may libmemory so libmemory so rwxrwxr x mario mario may libmemory so lrwxrwxrwx mario mario may libparcel coalescing so libparcel coalescing so lrwxrwxrwx mario mario may libparcel coalescing so libparcel coalescing so lrwxrwxrwx mario mario may libsine so libsine so lrwxrwxrwx mario mario may libsine so libsine so rwxrwxr x mario mario may libsine so so libparcel coalescing so appears to be missing make tests gives the following errors the following tests failed tests unit actions succeed compile const ref argument failed tests unit components distributed tcp action invoke no more than timeout tests unit threads thread affinity timeout errors while running ctest which might be a separate issue
| 0
|
1,780
| 6,715,574,036
|
IssuesEvent
|
2017-10-13 21:51:11
|
p4lang/p4-spec
|
https://api.github.com/repos/p4lang/p4-spec
|
closed
|
[PSA] Should PRE be defined to have multiple class-of-service queues per output port?
|
portable switch architecture
|
It is a relatively common feature for a queueing system to have multiple queues per output port, not merely one queue per port. Perhaps the PSA could have a constant that is part of the architecture spec, that could vary across implementations, that gives the maximum number of queues per output port, and a way to specify not only which output port a packet is enqueued for after ingress, before egress, but also which class-of-service queue for that output port?
|
1.0
|
[PSA] Should PRE be defined to have multiple class-of-service queues per output port? - It is a relatively common feature for a queueing system to have multiple queues per output port, not merely one queue per port. Perhaps the PSA could have a constant that is part of the architecture spec, that could vary across implementations, that gives the maximum number of queues per output port, and a way to specify not only which output port a packet is enqueued for after ingress, before egress, but also which class-of-service queue for that output port?
|
non_process
|
should pre be defined to have multiple class of service queues per output port it is a relatively common feature for a queueing system to have multiple queues per output port not merely one queue per port perhaps the psa could have a constant that is part of the architecture spec that could vary across implementations that gives the maximum number of queues per output port and a way to specify not only which output port a packet is enqueued for after ingress before egress but also which class of service queue for that output port
| 0
|
13,652
| 16,360,879,829
|
IssuesEvent
|
2021-05-14 09:16:26
|
CGAL/cgal
|
https://api.github.com/repos/CGAL/cgal
|
closed
|
SOR CGAL PointSet
|
Pkg::Point_set_processing_3 question
|
Hi,
I am trying to follow SOR outlier removal tutorial from here:
https://doc.cgal.org/latest/Point_set_processing_3/Point_set_processing_3_2remove_outliers_example_8cpp-example.html
But it works with Point vector.
Is there any possibility to use a PointSet with color also?
If yes is it possible to show a code example?
When I try to use pointset instead of point vector I get following error:

|
1.0
|
SOR CGAL PointSet - Hi,
I am trying to follow SOR outlier removal tutorial from here:
https://doc.cgal.org/latest/Point_set_processing_3/Point_set_processing_3_2remove_outliers_example_8cpp-example.html
But it works with Point vector.
Is there any possibility to use a PointSet with color also?
If yes is it possible to show a code example?
When I try to use pointset instead of point vector I get following error:

|
process
|
sor cgal pointset hi i am trying to follow sor outlier removal tutorial from here but it works with point vector is there any possibility to use a pointset with color also if yes is it possible to show a code example when i try to use pointset instead of point vector i get following error
| 1
|
734,067
| 25,337,302,987
|
IssuesEvent
|
2022-11-18 17:59:06
|
apache/lucene
|
https://api.github.com/repos/apache/lucene
|
closed
|
Add an equivalent of ant's stage-maven-artifacts for the release wizard [LUCENE-10293]
|
type:task legacy-jira-priority:Minor
|
Currently the RC Maven artifacts cannot be easily moved to Nexus. Devise a way to do it from gradle (either as a bundle ZIP or directly, the way it was done before).
---
Migrated from [LUCENE-10293](https://issues.apache.org/jira/browse/LUCENE-10293) by Dawid Weiss (@dweiss), updated Apr 14 2022
Linked issues:
- #10910
|
1.0
|
Add an equivalent of ant's stage-maven-artifacts for the release wizard [LUCENE-10293] - Currently the RC Maven artifacts cannot be easily moved to Nexus. Devise a way to do it from gradle (either as a bundle ZIP or directly, the way it was done before).
---
Migrated from [LUCENE-10293](https://issues.apache.org/jira/browse/LUCENE-10293) by Dawid Weiss (@dweiss), updated Apr 14 2022
Linked issues:
- #10910
|
non_process
|
add an equivalent of ant s stage maven artifacts for the release wizard currently the rc maven artifacts cannot be easily moved to nexus devise a way to do it from gradle either as a bundle zip or directly the way it was done before migrated from by dawid weiss dweiss updated apr linked issues
| 0
|
180,097
| 30,390,196,937
|
IssuesEvent
|
2023-07-13 06:16:04
|
airswap/airswap-web
|
https://api.github.com/repos/airswap/airswap-web
|
closed
|
Add wallet confirmation screen when user interacts with metamask (or other wallet)
|
enhancement needs-design
|
We need to improve the metamask (and other wallets) flow a bit. Right now the “Sign” and “Take” button will go in a loading state when opening metamask. I think for some users it didn’t open (or they didn’t notice it?) and that’s why they were stuck. So we need a “If your wallet does not open something went wrong” message. Same as marketplace.
Marketplace right now:

|
1.0
|
Add wallet confirmation screen when user interacts with metamask (or other wallet) - We need to improve the metamask (and other wallets) flow a bit. Right now the “Sign” and “Take” button will go in a loading state when opening metamask. I think for some users it didn’t open (or they didn’t notice it?) and that’s why they were stuck. So we need a “If your wallet does not open something went wrong” message. Same as marketplace.
Marketplace right now:

|
non_process
|
add wallet confirmation screen when user interacts with metamask or other wallet we need to improve the metamask and other wallets flow a bit right now the “sign” and “take” button will go in a loading state when opening metamask i think for some users it didn’t open or they didn’t notice it and that’s why they were stuck so we need a “if your wallet does not open something went wrong” message same as marketplace marketplace right now
| 0
|
90,994
| 15,856,355,169
|
IssuesEvent
|
2021-04-08 02:08:46
|
faizulho/storefront
|
https://api.github.com/repos/faizulho/storefront
|
opened
|
CVE-2020-28500 (Medium) detected in lodash-4.17.14.tgz
|
security vulnerability
|
## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.14.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz</a></p>
<p>Path to dependency file: storefront/package.json</p>
<p>Path to vulnerable library: storefront/node_modules/@commitlint/ensure/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- cli-8.2.0.tgz (Root Library)
- :x: **lodash-4.17.14.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash-4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-28500 (Medium) detected in lodash-4.17.14.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.14.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz</a></p>
<p>Path to dependency file: storefront/package.json</p>
<p>Path to vulnerable library: storefront/node_modules/@commitlint/ensure/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- cli-8.2.0.tgz (Root Library)
- :x: **lodash-4.17.14.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash-4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file storefront package json path to vulnerable library storefront node modules commitlint ensure node modules lodash package json dependency hierarchy cli tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
| 0
|
15,012
| 18,723,926,419
|
IssuesEvent
|
2021-11-03 14:34:30
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - relationshipOfResource
|
Term - change Class - ResourceRelationship non-normative Process - complete
|
## Change term
* Submitter: Peter Desmet @peterdesmet
* Justification (why is this change necessary?): The definition is unclear, even misleading. The way it is, directionality of the relationship can be interpreted both ways. The relationship must be from subject to object.
* Proponents (who needs this change): Everyone who uses the term.
Current Term definition: https://dwc.tdwg.org/terms/#dwc:relationshipOfResource
Proposed new attributes of the term:
* Term name (in lowerCamelCase): relationshipOfResource
* Organized in Class (e.g. Location, Taxon): ResourceRelationship
* Definition of the term: **The relationship of the subject (identified by resourceID) to the object (identified by relatedResourceID).**
* Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary.
* Examples: `sameAs`, `duplicate of`, `mother of`, `offspring of`, `sibling of`, **`parasite of`, `host of`**, `valid synonym of`, `located within`, **`pollinator of members of taxon`, `pollinated specific plant`, `pollinated by members of taxon`, `on slab with`**
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/relationshipOfResource-2018-09-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Associations/UnitAssociation/AssociationType
Original comment:
## The problem
The definition of [relationshipOfResource](http://rs.tdwg.org/dwc/terms/relationshipOfResource) is:
> The relationship of the resource identified by relatedResourceID to the subject (optionally identified by the resourceID). Recommended best practice is to use a controlled vocabulary.
But the intention was always to go from (@tucotuco concurs):
> resourceID (subject A) to relatedResourceID (object B)
Which is logical. The definition however goes in the other direction:
> The relationship of the resource identified by relatedResourceID (**B**) to the subject (**A**) (optionally identified by the resourceID). Recommended best practice is to use a controlled vocabulary.
That makes it very unintuitive, and the listed examples don't help (because e.g. `host to` could be interpreted in either direction). See for example how we struggled with it here: https://github.com/trias-project/uredinales-belgium-checklist/issues/8#issue-302633347:
> The resource relationship between rust fungi A and host B could be expressed as:
>
> id | resourceID | relatedResourceID | relationshipOfResource | relationshipAccordingTo
> --- | --- | --- | --- | ---
> A | (A) | B | host plant of | = bibliographicReference or source
> A | B | A | parasite of | = bibliographicReference or source
>
> It is unfortunate that the most logical relationship is `host plant of`, which is less specific than `parasite of` but that is because the definition for relationshipOfResource.
## Updating the definition
Can we still amend the definition please? E.g to:
> The relationship of the **subject (optionally identified by resourceID) to the object (identified by relatedResourceID)**. Recommended best practice is to use a controlled vocabulary.
## Breaking things
Obviously, people have probably done there best trying to understand the definition and been using it. Changing the direction of the relationship can break things. But:
- For describing relationships where the direction doesn't matter, nothing breaks (like `duplicate of`)
- How many datasets have been using the [Darwin Core Resource Relationship](http://rs.gbif.org/extension/dwc/resource_relation_2018_01_18.xml) (for directional relationships)? @mattblissett @timrobertson100 is there a way you can find out at GBIF / @pieterprovoost idem for OBIS?
I'd rather have this fixed now, assuming adoption is still low.
|
1.0
|
Change term - relationshipOfResource - ## Change term
* Submitter: Peter Desmet @peterdesmet
* Justification (why is this change necessary?): The definition is unclear, even misleading. The way it is, directionality of the relationship can be interpreted both ways. The relationship must be from subject to object.
* Proponents (who needs this change): Everyone who uses the term.
Current Term definition: https://dwc.tdwg.org/terms/#dwc:relationshipOfResource
Proposed new attributes of the term:
* Term name (in lowerCamelCase): relationshipOfResource
* Organized in Class (e.g. Location, Taxon): ResourceRelationship
* Definition of the term: **The relationship of the subject (identified by resourceID) to the object (identified by relatedResourceID).**
* Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary.
* Examples: `sameAs`, `duplicate of`, `mother of`, `offspring of`, `sibling of`, **`parasite of`, `host of`**, `valid synonym of`, `located within`, **`pollinator of members of taxon`, `pollinated specific plant`, `pollinated by members of taxon`, `on slab with`**
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/relationshipOfResource-2018-09-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Associations/UnitAssociation/AssociationType
Original comment:
## The problem
The definition of [relationshipOfResource](http://rs.tdwg.org/dwc/terms/relationshipOfResource) is:
> The relationship of the resource identified by relatedResourceID to the subject (optionally identified by the resourceID). Recommended best practice is to use a controlled vocabulary.
But the intention was always to go from (@tucotuco concurs):
> resourceID (subject A) to relatedResourceID (object B)
Which is logical. The definition however goes in the other direction:
> The relationship of the resource identified by relatedResourceID (**B**) to the subject (**A**) (optionally identified by the resourceID). Recommended best practice is to use a controlled vocabulary.
That makes it very unintuitive, and the listed examples don't help (because e.g. `host to` could be interpreted in either direction). See for example how we struggled with it here: https://github.com/trias-project/uredinales-belgium-checklist/issues/8#issue-302633347:
> The resource relationship between rust fungi A and host B could be expressed as:
>
> id | resourceID | relatedResourceID | relationshipOfResource | relationshipAccordingTo
> --- | --- | --- | --- | ---
> A | (A) | B | host plant of | = bibliographicReference or source
> A | B | A | parasite of | = bibliographicReference or source
>
> It is unfortunate that the most logical relationship is `host plant of`, which is less specific than `parasite of` but that is because the definition for relationshipOfResource.
## Updating the definition
Can we still amend the definition please? E.g to:
> The relationship of the **subject (optionally identified by resourceID) to the object (identified by relatedResourceID)**. Recommended best practice is to use a controlled vocabulary.
## Breaking things
Obviously, people have probably done there best trying to understand the definition and been using it. Changing the direction of the relationship can break things. But:
- For describing relationships where the direction doesn't matter, nothing breaks (like `duplicate of`)
- How many datasets have been using the [Darwin Core Resource Relationship](http://rs.gbif.org/extension/dwc/resource_relation_2018_01_18.xml) (for directional relationships)? @mattblissett @timrobertson100 is there a way you can find out at GBIF / @pieterprovoost idem for OBIS?
I'd rather have this fixed now, assuming adoption is still low.
|
process
|
change term relationshipofresource change term submitter peter desmet peterdesmet justification why is this change necessary the definition is unclear even misleading the way it is directionality of the relationship can be interpreted both ways the relationship must be from subject to object proponents who needs this change everyone who uses the term current term definition proposed new attributes of the term term name in lowercamelcase relationshipofresource organized in class e g location taxon resourcerelationship definition of the term the relationship of the subject identified by resourceid to the object identified by relatedresourceid usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary examples sameas duplicate of mother of offspring of sibling of parasite of host of valid synonym of located within pollinator of members of taxon pollinated specific plant pollinated by members of taxon on slab with refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable datasets dataset units unit associations unitassociation associationtype original comment the problem the definition of is the relationship of the resource identified by relatedresourceid to the subject optionally identified by the resourceid recommended best practice is to use a controlled vocabulary but the intention was always to go from tucotuco concurs resourceid subject a to relatedresourceid object b which is logical the definition however goes in the other direction the relationship of the resource identified by relatedresourceid b to the subject a optionally identified by the resourceid recommended best practice is to use a controlled vocabulary that makes it very unintuitive and the listed examples don t help because e g host to could be interpreted in either direction see for example how we struggled with it here the resource relationship between rust fungi a and host b could be expressed as id resourceid relatedresourceid relationshipofresource relationshipaccordingto a a b host plant of bibliographicreference or source a b a parasite of bibliographicreference or source it is unfortunate that the most logical relationship is host plant of which is less specific than parasite of but that is because the definition for relationshipofresource updating the definition can we still amend the definition please e g to the relationship of the subject optionally identified by resourceid to the object identified by relatedresourceid recommended best practice is to use a controlled vocabulary breaking things obviously people have probably done there best trying to understand the definition and been using it changing the direction of the relationship can break things but for describing relationships where the direction doesn t matter nothing breaks like duplicate of how many datasets have been using the for directional relationships mattblissett is there a way you can find out at gbif pieterprovoost idem for obis i d rather have this fixed now assuming adoption is still low
| 1
|
11,207
| 13,957,705,381
|
IssuesEvent
|
2020-10-24 08:14:30
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
MT-MITA: Harvest
|
Geoportal Harvesting process MT - Malta
|
Good Morning Angelo,
Kindly can you perform a harvest for the Maltese CSW as the last harvest was performed on the 26th March 2018. On a side note is there a problem with the harvesting or its just the portal is currently being busy? Kindly can you check whether the harvesting for Malta is seen set to daily from your side? Thanks for your help.
Regards,
Rene Agius
|
1.0
|
MT-MITA: Harvest - Good Morning Angelo,
Kindly can you perform a harvest for the Maltese CSW as the last harvest was performed on the 26th March 2018. On a side note is there a problem with the harvesting or its just the portal is currently being busy? Kindly can you check whether the harvesting for Malta is seen set to daily from your side? Thanks for your help.
Regards,
Rene Agius
|
process
|
mt mita harvest good morning angelo kindly can you perform a harvest for the maltese csw as the last harvest was performed on the march on a side note is there a problem with the harvesting or its just the portal is currently being busy kindly can you check whether the harvesting for malta is seen set to daily from your side thanks for your help regards rene agius
| 1
|
506,053
| 14,657,297,047
|
IssuesEvent
|
2020-12-28 15:17:56
|
ac2cz/FoxTelem
|
https://api.github.com/repos/ac2cz/FoxTelem
|
opened
|
PPM adjustment for RTL
|
RTL RTL-SDR enhancement low priority
|
Is it possible to implement an option to enter the ppm-value of an RTL-SDR as they differ from piece to piece (have 3 of them, one has 56 ppm, next 68, the last one 41)?
|
1.0
|
PPM adjustment for RTL - Is it possible to implement an option to enter the ppm-value of an RTL-SDR as they differ from piece to piece (have 3 of them, one has 56 ppm, next 68, the last one 41)?
|
non_process
|
ppm adjustment for rtl is it possible to implement an option to enter the ppm value of an rtl sdr as they differ from piece to piece have of them one has ppm next the last one
| 0
|
9,192
| 12,228,996,307
|
IssuesEvent
|
2020-05-03 21:55:19
|
chfor183/data_science_articles
|
https://api.github.com/repos/chfor183/data_science_articles
|
opened
|
Integrated Tools
|
Data Preprocessing Machine Learning Probability and Statistic
|
## TL;DR
Alteryx
RapidMiner
IBM SPSS
Orange
Google Analytics
SAS (Statistical Analysis System)
Data Robot
BigML
Knime
STATA
H2O.ai
## Key Takeaways
- 1
- 2
## Useful Code Snippets
```
function test() {
console.log("notice the blank line before this function?");
}
```
## Articles/Ressources
|
1.0
|
Integrated Tools - ## TL;DR
Alteryx
RapidMiner
IBM SPSS
Orange
Google Analytics
SAS (Statistical Analysis System)
Data Robot
BigML
Knime
STATA
H2O.ai
## Key Takeaways
- 1
- 2
## Useful Code Snippets
```
function test() {
console.log("notice the blank line before this function?");
}
```
## Articles/Ressources
|
process
|
integrated tools tl dr alteryx rapidminer ibm spss orange google analytics sas statistical analysis system data robot bigml knime stata ai key takeaways useful code snippets function test console log notice the blank line before this function articles ressources
| 1
|
17,324
| 23,142,866,745
|
IssuesEvent
|
2022-07-28 20:23:11
|
USGS-R/drb-do-ml
|
https://api.github.com/repos/USGS-R/drb-do-ml
|
reopened
|
Inspect hidden states for process understanding
|
experiment process-guidance
|
An idea that has come up is to inspect the hidden states to see if they are behaving as we would expect some state or flux in the process would behave.
The two examples that have come up are:
1. a **biomass**: some representation of a biomass that accumulates in the summer, is lower in the winter, and maybe has a sharp decrease after a scouring event
2. **discharge**: we aren't giving the baseline model discharge as an input. Does the model have a discharge-like hidden state?
We can look into answering these questions with the baseline LSTM model (#40).
|
1.0
|
Inspect hidden states for process understanding - An idea that has come up is to inspect the hidden states to see if they are behaving as we would expect some state or flux in the process would behave.
The two examples that have come up are:
1. a **biomass**: some representation of a biomass that accumulates in the summer, is lower in the winter, and maybe has a sharp decrease after a scouring event
2. **discharge**: we aren't giving the baseline model discharge as an input. Does the model have a discharge-like hidden state?
We can look into answering these questions with the baseline LSTM model (#40).
|
process
|
inspect hidden states for process understanding an idea that has come up is to inspect the hidden states to see if they are behaving as we would expect some state or flux in the process would behave the two examples that have come up are a biomass some representation of a biomass that accumulates in the summer is lower in the winter and maybe has a sharp decrease after a scouring event discharge we aren t giving the baseline model discharge as an input does the model have a discharge like hidden state we can look into answering these questions with the baseline lstm model
| 1
|
6,437
| 2,588,038,355
|
IssuesEvent
|
2015-02-17 22:13:34
|
ilios/ilios
|
https://api.github.com/repos/ilios/ilios
|
closed
|
load user attributes in management console asynchronously
|
Deprecated Functionality low priority wontfix
|
refs #364.
see `ilios.management.user_accounts.buildUserAddAndRolesDOM`
rewrite this whole mess to load user attributes asynchronously.
|
1.0
|
load user attributes in management console asynchronously - refs #364.
see `ilios.management.user_accounts.buildUserAddAndRolesDOM`
rewrite this whole mess to load user attributes asynchronously.
|
non_process
|
load user attributes in management console asynchronously refs see ilios management user accounts builduseraddandrolesdom rewrite this whole mess to load user attributes asynchronously
| 0
|
188,524
| 15,164,537,027
|
IssuesEvent
|
2021-02-12 13:52:57
|
arturo-lang/arturo
|
https://api.github.com/repos/arturo-lang/arturo
|
closed
|
[Reflection\boolean?] add example for documentation
|
documentation easy library todo
|
[Reflection\boolean?] add example for documentation
https://github.com/arturo-lang/arturo/blob/a971add892fe3d675b3320f356cf2d96179e2a22/src/library/Reflection.nim#L207
```text
attrs = NoAttrs,
returns = {Boolean},
# TODO(Reflection\boolean?) add example for documentation
# labels: library,documentation,easy
example = """
""":
##########################################################
```
6a937201380f786d373dd31f62d5d248c07af4e6
|
1.0
|
[Reflection\boolean?] add example for documentation - [Reflection\boolean?] add example for documentation
https://github.com/arturo-lang/arturo/blob/a971add892fe3d675b3320f356cf2d96179e2a22/src/library/Reflection.nim#L207
```text
attrs = NoAttrs,
returns = {Boolean},
# TODO(Reflection\boolean?) add example for documentation
# labels: library,documentation,easy
example = """
""":
##########################################################
```
6a937201380f786d373dd31f62d5d248c07af4e6
|
non_process
|
add example for documentation add example for documentation text attrs noattrs returns boolean todo reflection boolean add example for documentation labels library documentation easy example
| 0
|
14,731
| 17,950,275,452
|
IssuesEvent
|
2021-09-12 15:41:24
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Implement analysis progress endpoint [3]
|
Data Processing Week 2 Sprint 4
|
Do the implementation of the analysis progress to know in which part of the data processing flow is.
|
1.0
|
Implement analysis progress endpoint [3] - Do the implementation of the analysis progress to know in which part of the data processing flow is.
|
process
|
implement analysis progress endpoint do the implementation of the analysis progress to know in which part of the data processing flow is
| 1
|
9,384
| 3,038,668,804
|
IssuesEvent
|
2015-08-07 00:28:20
|
servo/servo
|
https://api.github.com/repos/servo/servo
|
closed
|
Write a reftest for the background attribute on the body element
|
A-testing C-assigned E-easy
|
#5851 merged without any tests. We should add one - have a reference page that uses CSS to set the background image for a page, and a test page that sets the attribute on the body element instead. See the documentation at http://testthewebforward.org/docs/reftests.html and examples at http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ . You'll need to run `./mach test-wpt --manifest-update tests/wpt/mozilla/tests/css/[new_test.html]` the first time after creating the test in order to update the MANIFEST.json that contains information about known tests.
|
1.0
|
Write a reftest for the background attribute on the body element - #5851 merged without any tests. We should add one - have a reference page that uses CSS to set the background image for a page, and a test page that sets the attribute on the body element instead. See the documentation at http://testthewebforward.org/docs/reftests.html and examples at http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ . You'll need to run `./mach test-wpt --manifest-update tests/wpt/mozilla/tests/css/[new_test.html]` the first time after creating the test in order to update the MANIFEST.json that contains information about known tests.
|
non_process
|
write a reftest for the background attribute on the body element merged without any tests we should add one have a reference page that uses css to set the background image for a page and a test page that sets the attribute on the body element instead see the documentation at and examples at you ll need to run mach test wpt manifest update tests wpt mozilla tests css the first time after creating the test in order to update the manifest json that contains information about known tests
| 0
|
820,392
| 30,771,007,132
|
IssuesEvent
|
2023-07-30 22:38:26
|
apcountryman/picolibrary
|
https://api.github.com/repos/apcountryman/picolibrary
|
opened
|
Add WIZnet W5500 TCP over IP client socket
|
priority-normal status-awaiting_development type-feature
|
Add WIZnet W5500 TCP over IP client socket (`::picolibrary::WIZnet::W5500::IP::TCP::Client`) and associated mock (`::picolibrary::Testing::Automated::WIZnet::W5500::IP::TCP::Mock_Client`).
- [ ] The `Client` class should be defined in the `include/picolibrary/wiznet/w5500/ip/tcp.h`/`source/picolibrary/wiznet/w5500/ip/tcp.cc` header/source file pair
- [ ] The `Mock_Client` class should be defined in the `include/picolibrary/testing/automated/wiznet/w5500/ip/tcp.h`/`source/picolibrary/testing/automated/wiznet/w5500/ip/tcp.cc` header/source file pair
- [ ] The `Client` class should have automated tests
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_Driver_Access_Key` class
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_Socket_Allocation_Key` class
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_TCP_Port_Allocator_Access_Key` class
- [ ] The `Client` class should have the following template parameters:
- [ ] `typename Network_Stack`: The type of network stack the socket is associated with
- [ ] The `Client` class should have the following member types:
- [ ] `using Size = std::uint16_t;`: The unsigned integer type used to report transmit/receive buffer information
- [ ] `enum class State;`: Socket state
- [ ] The `State` enum class should have an underlying type of `std::uint_fast8_t`
- [ ] The `State` enum class should have the following enumerators:
- [ ] `UNINITIALIZED,`: Uninitialized
- [ ] `INITIALIZED,`: Initialized
- [ ] `BOUND,`: Bound
- [ ] `CONNECTING,`: Connecting
- [ ] `CONNECTED,`: Connected
- [ ] The `Client` class should support the following operations:
- [ ] `constexpr Client() noexcept = default;`
- [ ] `constexpr Client( Network_Stack & network_stack, Socket_ID socket_id ) noexcept;`
- [ ] `constexpr Client( State state, Network_Stack & network_stack, Socket_ID socket_id, bool is_transmitting = false ) noexcept;`
- [ ] This function should only be available if `PICOLIBRARY_ENABLE_AUTOMATED_TESTING` is defined
- [ ] `constexpr Client( Client && source ) noexcept;`
- [ ] `~Client() noexcept;`
- [ ] `constexpr auto operator=( Client && expression ) noexcept -> Client &;`
- [ ] `constexpr auto state() const noexcept -> State;`: Get the socket's state
- [ ] `constexpr auto socket_id() const noexcept -> Socket_ID;`: Get the socket's socket ID
- [ ] `constexpr auto socket_interrupt_mask() const noexcept -> std::uint8_t;`: Get the socket's socket interrupt mask (mask used when checking the network stack's socket interrupt context)
- [ ] `void configure_no_delayed_ack_usage( No_Delayed_ACK_Usage no_delayed_ack_usage_configuration ) noexcept;`: Configure the socket's no delayed ACK usage (default to disabled)
- [ ] `auto no_delayed_ack_usage() const noexcept -> No_Delayed_ACK_Usage;`: Get the socket's no delayed ACK usage configuration
- [ ] `void configure_maximum_segment_size( std::uint16_t maximum_segment_size ) noexcept;`: Configure the socket's maximum segment size
- [ ] `auto maximum_segment_size() const noexcept -> std::uint16_t;`: Get the socket's maximum segment size
- [ ] `void configure_time_to_live( std::uint8_t time_to_live ) noexcept;`: Configure the socket's IPv4 time to live field value (defaults to 0x80)
- [ ] `auto time_to_live() const noexcept -> std::uint8_t;`: Get the socket's IPv4 time to live field value
- [ ] `void configure_keepalive_period( std::uint8_t keepalive_period ) noexcept;`: Configure the socket's keepalive packet transmission period (SN_KPALVTR register value, defaults to 0x00)
- [ ] `auto keepalive_period() const noexcept -> std::uint8_t;`: Get the socket's keepalive packet transmission period
- [ ] `void enable_interrupts( std::uint8_t mask ) noexcept;`: Enable interrupts
- [ ] `void disable_interrupts( std::uint8_t mask ) noexcept;`: Disable interrupts
- [ ] `auto enabled_interrupts() const noexcept -> std::uint8_t;`: Get a mask identifying the interrupts that are enabled
- [ ] `auto interrupt_context() const noexcept -> std::uint8_t;`: Get the interrupt context (SN_IR register value)
- [ ] `void clear_interrupts( std::uint8_t mask );`: Clear interrupts
- [ ] `void bind( ::picolibrary::IP::TCP::Endpoint const & endpoint = ::picolibrary::IP::TCP::Endpoint{} ) noexcept;`: Bind the socket to a local endpoint
- [ ] `auto connect( ::picolibrary::IP::TCP::Endpoint const & endpoint ) noexcept -> Result<void>;`: Connect to a remote endpoint
- [ ] `auto is_connected() const noexcept -> bool;`: Check if the socket is connected to a remote endpoint
- [ ] `auto local_endpoint() const noexcept -> ::picolibrary::IP::TCP::Endpoint;`: Get the connection's local endpoint
- [ ] `auto remote_endpoint() const noexcept -> ::picolibrary::IP::TCP::Endpoint;`: Get the connection's remote endpoint
- [ ] `auto outstanding() const noexcept -> Size;`: Get the amount of data that has yet to be transmitted to the remote endpoint
- [ ] `constexpr auto is_transmitting() const noexcept -> bool;`: Check if data transmission is in progress
- [ ] `auto transmit( std::uint8_t const * begin, std::uint8_t const * end ) noexcept -> Result<std::uint8_t const *>;`: Transmit data to the remote endpoint
- [ ] `auto transmit_keepalive() noexcept -> Result<void>;`: Manually transmit a keepalive packet
- [ ] `auto available() const noexcept -> Size;`: Get the amount of data that is immediately available to be received from the remote endpoint
- [ ] `auto receive( std::uint8_t * begin, std::uint8_t * end ) noexcept -> Result<std::uint8_t *>;`: Receive data from the remote endpoint
- [ ] `void shutdown() noexcept;`: Disable further data transmission and reception
- [ ] `constexpr void close() noexcept;`: Close the socket
- [ ] The `Mock_Client` class should have a movable mock handle
- [ ] The following member types should be added to the `::picolibrary::WIZnet::W5500::IP::Network_Stack` class:
- [ ] `using TCP_Client = TCP::Client;`: The type of TCP client socket that is used to interact with the network stack
- [ ] Support for the following operations should be added to the `::picolibrary::WIZnet::W5500::IP::Network_Stack` class
- [ ] `auto make_tcp_client() noexcept -> TCP_Client;`: Construct a TCP client socket
- [ ] `auto make_tcp_client( Socket_ID socket_id ) noexcept -> TCP_Client;`: Construct a TCP client socket that uses a specific W5500 socket
- [ ] Documentation
|
1.0
|
Add WIZnet W5500 TCP over IP client socket - Add WIZnet W5500 TCP over IP client socket (`::picolibrary::WIZnet::W5500::IP::TCP::Client`) and associated mock (`::picolibrary::Testing::Automated::WIZnet::W5500::IP::TCP::Mock_Client`).
- [ ] The `Client` class should be defined in the `include/picolibrary/wiznet/w5500/ip/tcp.h`/`source/picolibrary/wiznet/w5500/ip/tcp.cc` header/source file pair
- [ ] The `Mock_Client` class should be defined in the `include/picolibrary/testing/automated/wiznet/w5500/ip/tcp.h`/`source/picolibrary/testing/automated/wiznet/w5500/ip/tcp.cc` header/source file pair
- [ ] The `Client` class should have automated tests
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_Driver_Access_Key` class
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_Socket_Allocation_Key` class
- [ ] The `Client` class should be a friend of the `::picolibrary::WIZnet::W5500::IP::Network_Stack_TCP_Port_Allocator_Access_Key` class
- [ ] The `Client` class should have the following template parameters:
- [ ] `typename Network_Stack`: The type of network stack the socket is associated with
- [ ] The `Client` class should have the following member types:
- [ ] `using Size = std::uint16_t;`: The unsigned integer type used to report transmit/receive buffer information
- [ ] `enum class State;`: Socket state
- [ ] The `State` enum class should have an underlying type of `std::uint_fast8_t`
- [ ] The `State` enum class should have the following enumerators:
- [ ] `UNINITIALIZED,`: Uninitialized
- [ ] `INITIALIZED,`: Initialized
- [ ] `BOUND,`: Bound
- [ ] `CONNECTING,`: Connecting
- [ ] `CONNECTED,`: Connected
- [ ] The `Client` class should support the following operations:
- [ ] `constexpr Client() noexcept = default;`
- [ ] `constexpr Client( Network_Stack & network_stack, Socket_ID socket_id ) noexcept;`
- [ ] `constexpr Client( State state, Network_Stack & network_stack, Socket_ID socket_id, bool is_transmitting = false ) noexcept;`
- [ ] This function should only be available if `PICOLIBRARY_ENABLE_AUTOMATED_TESTING` is defined
- [ ] `constexpr Client( Client && source ) noexcept;`
- [ ] `~Client() noexcept;`
- [ ] `constexpr auto operator=( Client && expression ) noexcept -> Client &;`
- [ ] `constexpr auto state() const noexcept -> State;`: Get the socket's state
- [ ] `constexpr auto socket_id() const noexcept -> Socket_ID;`: Get the socket's socket ID
- [ ] `constexpr auto socket_interrupt_mask() const noexcept -> std::uint8_t;`: Get the socket's socket interrupt mask (mask used when checking the network stack's socket interrupt context)
- [ ] `void configure_no_delayed_ack_usage( No_Delayed_ACK_Usage no_delayed_ack_usage_configuration ) noexcept;`: Configure the socket's no delayed ACK usage (default to disabled)
- [ ] `auto no_delayed_ack_usage() const noexcept -> No_Delayed_ACK_Usage;`: Get the socket's no delayed ACK usage configuration
- [ ] `void configure_maximum_segment_size( std::uint16_t maximum_segment_size ) noexcept;`: Configure the socket's maximum segment size
- [ ] `auto maximum_segment_size() const noexcept -> std::uint16_t;`: Get the socket's maximum segment size
- [ ] `void configure_time_to_live( std::uint8_t time_to_live ) noexcept;`: Configure the socket's IPv4 time to live field value (defaults to 0x80)
- [ ] `auto time_to_live() const noexcept -> std::uint8_t;`: Get the socket's IPv4 time to live field value
- [ ] `void configure_keepalive_period( std::uint8_t keepalive_period ) noexcept;`: Configure the socket's keepalive packet transmission period (SN_KPALVTR register value, defaults to 0x00)
- [ ] `auto keepalive_period() const noexcept -> std::uint8_t;`: Get the socket's keepalive packet transmission period
- [ ] `void enable_interrupts( std::uint8_t mask ) noexcept;`: Enable interrupts
- [ ] `void disable_interrupts( std::uint8_t mask ) noexcept;`: Disable interrupts
- [ ] `auto enabled_interrupts() const noexcept -> std::uint8_t;`: Get a mask identifying the interrupts that are enabled
- [ ] `auto interrupt_context() const noexcept -> std::uint8_t;`: Get the interrupt context (SN_IR register value)
- [ ] `void clear_interrupts( std::uint8_t mask );`: Clear interrupts
- [ ] `void bind( ::picolibrary::IP::TCP::Endpoint const & endpoint = ::picolibrary::IP::TCP::Endpoint{} ) noexcept;`: Bind the socket to a local endpoint
- [ ] `auto connect( ::picolibrary::IP::TCP::Endpoint const & endpoint ) noexcept -> Result<void>;`: Connect to a remote endpoint
- [ ] `auto is_connected() const noexcept -> bool;`: Check if the socket is connected to a remote endpoint
- [ ] `auto local_endpoint() const noexcept -> ::picolibrary::IP::TCP::Endpoint;`: Get the connection's local endpoint
- [ ] `auto remote_endpoint() const noexcept -> ::picolibrary::IP::TCP::Endpoint;`: Get the connection's remote endpoint
- [ ] `auto outstanding() const noexcept -> Size;`: Get the amount of data that has yet to be transmitted to the remote endpoint
- [ ] `constexpr auto is_transmitting() const noexcept -> bool;`: Check if data transmission is in progress
- [ ] `auto transmit( std::uint8_t const * begin, std::uint8_t const * end ) noexcept -> Result<std::uint8_t const *>;`: Transmit data to the remote endpoint
- [ ] `auto transmit_keepalive() noexcept -> Result<void>;`: Manually transmit a keepalive packet
- [ ] `auto available() const noexcept -> Size;`: Get the amount of data that is immediately available to be received from the remote endpoint
- [ ] `auto receive( std::uint8_t * begin, std::uint8_t * end ) noexcept -> Result<std::uint8_t *>;`: Receive data from the remote endpoint
- [ ] `void shutdown() noexcept;`: Disable further data transmission and reception
- [ ] `constexpr void close() noexcept;`: Close the socket
- [ ] The `Mock_Client` class should have a movable mock handle
- [ ] The following member types should be added to the `::picolibrary::WIZnet::W5500::IP::Network_Stack` class:
- [ ] `using TCP_Client = TCP::Client;`: The type of TCP client socket that is used to interact with the network stack
- [ ] Support for the following operations should be added to the `::picolibrary::WIZnet::W5500::IP::Network_Stack` class
- [ ] `auto make_tcp_client() noexcept -> TCP_Client;`: Construct a TCP client socket
- [ ] `auto make_tcp_client( Socket_ID socket_id ) noexcept -> TCP_Client;`: Construct a TCP client socket that uses a specific W5500 socket
- [ ] Documentation
|
non_process
|
add wiznet tcp over ip client socket add wiznet tcp over ip client socket picolibrary wiznet ip tcp client and associated mock picolibrary testing automated wiznet ip tcp mock client the client class should be defined in the include picolibrary wiznet ip tcp h source picolibrary wiznet ip tcp cc header source file pair the mock client class should be defined in the include picolibrary testing automated wiznet ip tcp h source picolibrary testing automated wiznet ip tcp cc header source file pair the client class should have automated tests the client class should be a friend of the picolibrary wiznet ip network stack driver access key class the client class should be a friend of the picolibrary wiznet ip network stack socket allocation key class the client class should be a friend of the picolibrary wiznet ip network stack tcp port allocator access key class the client class should have the following template parameters typename network stack the type of network stack the socket is associated with the client class should have the following member types using size std t the unsigned integer type used to report transmit receive buffer information enum class state socket state the state enum class should have an underlying type of std uint t the state enum class should have the following enumerators uninitialized uninitialized initialized initialized bound bound connecting connecting connected connected the client class should support the following operations constexpr client noexcept default constexpr client network stack network stack socket id socket id noexcept constexpr client state state network stack network stack socket id socket id bool is transmitting false noexcept this function should only be available if picolibrary enable automated testing is defined constexpr client client source noexcept client noexcept constexpr auto operator client expression noexcept client constexpr auto state const noexcept state get the socket s state constexpr auto socket id const noexcept socket id get the socket s socket id constexpr auto socket interrupt mask const noexcept std t get the socket s socket interrupt mask mask used when checking the network stack s socket interrupt context void configure no delayed ack usage no delayed ack usage no delayed ack usage configuration noexcept configure the socket s no delayed ack usage default to disabled auto no delayed ack usage const noexcept no delayed ack usage get the socket s no delayed ack usage configuration void configure maximum segment size std t maximum segment size noexcept configure the socket s maximum segment size auto maximum segment size const noexcept std t get the socket s maximum segment size void configure time to live std t time to live noexcept configure the socket s time to live field value defaults to auto time to live const noexcept std t get the socket s time to live field value void configure keepalive period std t keepalive period noexcept configure the socket s keepalive packet transmission period sn kpalvtr register value defaults to auto keepalive period const noexcept std t get the socket s keepalive packet transmission period void enable interrupts std t mask noexcept enable interrupts void disable interrupts std t mask noexcept disable interrupts auto enabled interrupts const noexcept std t get a mask identifying the interrupts that are enabled auto interrupt context const noexcept std t get the interrupt context sn ir register value void clear interrupts std t mask clear interrupts void bind picolibrary ip tcp endpoint const endpoint picolibrary ip tcp endpoint noexcept bind the socket to a local endpoint auto connect picolibrary ip tcp endpoint const endpoint noexcept result connect to a remote endpoint auto is connected const noexcept bool check if the socket is connected to a remote endpoint auto local endpoint const noexcept picolibrary ip tcp endpoint get the connection s local endpoint auto remote endpoint const noexcept picolibrary ip tcp endpoint get the connection s remote endpoint auto outstanding const noexcept size get the amount of data that has yet to be transmitted to the remote endpoint constexpr auto is transmitting const noexcept bool check if data transmission is in progress auto transmit std t const begin std t const end noexcept result transmit data to the remote endpoint auto transmit keepalive noexcept result manually transmit a keepalive packet auto available const noexcept size get the amount of data that is immediately available to be received from the remote endpoint auto receive std t begin std t end noexcept result receive data from the remote endpoint void shutdown noexcept disable further data transmission and reception constexpr void close noexcept close the socket the mock client class should have a movable mock handle the following member types should be added to the picolibrary wiznet ip network stack class using tcp client tcp client the type of tcp client socket that is used to interact with the network stack support for the following operations should be added to the picolibrary wiznet ip network stack class auto make tcp client noexcept tcp client construct a tcp client socket auto make tcp client socket id socket id noexcept tcp client construct a tcp client socket that uses a specific socket documentation
| 0
|
10,342
| 13,170,876,107
|
IssuesEvent
|
2020-08-11 15:46:40
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
.transaction([]) fails with Request Timeout Error
|
bug/2-confirmed kind/bug process/candidate team/support topic: previewFeatures topic: transactionApi
|
## Bug description
Using the new `.transaction([])`, I am seeing this on the client (JavaScript):
```
{ Error: Request Timeout Error
at PrismaClientFetcher._request (/app/node_modules/@prisma/client/runtime/index.js:1:213208) code: 'UND_ERR_REQUEST_TIMEOUT', meta: undefined }
```
## How to reproduce
Add multiple child records to an existing parent record.
```
const requests = [];
// loop that creates around 30 different requests, each with perhaps 500 to 1000 and puts them in the array like this:
requests.push(prisma.{entity}.update({
where: { id: {id that exists}},
data: {
{related entity name}: {
create: [
{ between 500 and 1000 individual entities },
]
},
})
await prisma.transaction(requests);
```
## Expected behavior
Should complete without throwing the exception.
## Prisma information
Sorry, I can not share this information.
But, I can tell you it's just doing basic inserts.
## Environment & setup
- OS: macOS 10.15.5
- Database: MySQL 5.7.x
- Prisma version: 2.2.0
- Node.js version: v14.5.0
|
1.0
|
.transaction([]) fails with Request Timeout Error - ## Bug description
Using the new `.transaction([])`, I am seeing this on the client (JavaScript):
```
{ Error: Request Timeout Error
at PrismaClientFetcher._request (/app/node_modules/@prisma/client/runtime/index.js:1:213208) code: 'UND_ERR_REQUEST_TIMEOUT', meta: undefined }
```
## How to reproduce
Add multiple child records to an existing parent record.
```
const requests = [];
// loop that creates around 30 different requests, each with perhaps 500 to 1000 and puts them in the array like this:
requests.push(prisma.{entity}.update({
where: { id: {id that exists}},
data: {
{related entity name}: {
create: [
{ between 500 and 1000 individual entities },
]
},
})
await prisma.transaction(requests);
```
## Expected behavior
Should complete without throwing the exception.
## Prisma information
Sorry, I can not share this information.
But, I can tell you it's just doing basic inserts.
## Environment & setup
- OS: macOS 10.15.5
- Database: MySQL 5.7.x
- Prisma version: 2.2.0
- Node.js version: v14.5.0
|
process
|
transaction fails with request timeout error bug description using the new transaction i am seeing this on the client javascript error request timeout error at prismaclientfetcher request app node modules prisma client runtime index js code und err request timeout meta undefined how to reproduce add multiple child records to an existing parent record const requests loop that creates around different requests each with perhaps to and puts them in the array like this requests push prisma entity update where id id that exists data related entity name create between and individual entities await prisma transaction requests expected behavior should complete without throwing the exception prisma information sorry i can not share this information but i can tell you it s just doing basic inserts environment setup os macos database mysql x prisma version node js version
| 1
|
22,663
| 31,895,971,157
|
IssuesEvent
|
2023-09-18 01:45:13
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - preparations
|
Term - change normative Task Group - Material Sample Process - complete Class - MaterialEntity
|
## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): The MaterialSample Task Group concluded that `dwc:preparations` should be organized under MaterialEntity rather than Occurrence, and that developing a MaterialEntity extension -- to rigorously address the things (parts), treatment (preparations), and storage regimes (preservations) -- would be a timely (and overdue) project, but it is out of scope for the MaterialSample project. In the short-term, some clarification of the syntax and examples would be useful.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?):
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_preparations
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): preparations
* Term label (English, not normative): Preparations
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): **MaterialEntity** ~~Occurrence~~
* Definition of the term (normative): A list (concatenated and separated) of preparations and preservation methods for a **dwc:MaterialEntity**~~specimen~~.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to separate the values in a list with space vertical bar space ( | ).
* Examples (not normative): fossil, cast, photograph, DNA extract, skin | skull | skeleton, whole animal (ETOH) | tissue (EDTA)
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/SpecimenUnit/Preparations/PreparationsText
|
1.0
|
Change term - preparations - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): The MaterialSample Task Group concluded that `dwc:preparations` should be organized under MaterialEntity rather than Occurrence, and that developing a MaterialEntity extension -- to rigorously address the things (parts), treatment (preparations), and storage regimes (preservations) -- would be a timely (and overdue) project, but it is out of scope for the MaterialSample project. In the short-term, some clarification of the syntax and examples would be useful.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?):
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_preparations
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): preparations
* Term label (English, not normative): Preparations
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): **MaterialEntity** ~~Occurrence~~
* Definition of the term (normative): A list (concatenated and separated) of preparations and preservation methods for a **dwc:MaterialEntity**~~specimen~~.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to separate the values in a list with space vertical bar space ( | ).
* Examples (not normative): fossil, cast, photograph, DNA extract, skin | skull | skeleton, whole animal (ETOH) | tissue (EDTA)
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/SpecimenUnit/Preparations/PreparationsText
|
process
|
change term preparations term change submitter efficacy justification why is this change necessary the materialsample task group concluded that dwc preparations should be organized under materialentity rather than occurrence and that developing a materialentity extension to rigorously address the things parts treatment preparations and storage regimes preservations would be a timely and overdue project but it is out of scope for the materialsample project in the short term some clarification of the syntax and examples would be useful demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes preparations term label english not normative preparations organized in class e g occurrence event location taxon materialentity occurrence definition of the term normative a list concatenated and separated of preparations and preservation methods for a dwc materialentity specimen usage comments recommendations regarding content etc not normative recommended best practice is to separate the values in a list with space vertical bar space examples not normative fossil cast photograph dna extract skin skull skeleton whole animal etoh tissue edta refines identifier of the broader term this term refines normative replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit specimenunit preparations preparationstext
| 1
|
8,753
| 11,873,639,569
|
IssuesEvent
|
2020-03-26 17:37:27
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
opened
|
Add `cli` as executable for @prisma/cli?
|
kind/discussion kind/improvement process/candidate topic: cli
|
## Problem
When using npx we need to do `npx -p @prisma/cli@alpha prisma2 --version`
## Solution
If we put another name for the executable in the package.json `cli` npx should work like:
`npx @prisma/cli@alpha --version` which is a lot easier (no parameters!)
## Additional context
We may want to test how it's doing with a global install.
|
1.0
|
Add `cli` as executable for @prisma/cli? - ## Problem
When using npx we need to do `npx -p @prisma/cli@alpha prisma2 --version`
## Solution
If we put another name for the executable in the package.json `cli` npx should work like:
`npx @prisma/cli@alpha --version` which is a lot easier (no parameters!)
## Additional context
We may want to test how it's doing with a global install.
|
process
|
add cli as executable for prisma cli problem when using npx we need to do npx p prisma cli alpha version solution if we put another name for the executable in the package json cli npx should work like npx prisma cli alpha version which is a lot easier no parameters additional context we may want to test how it s doing with a global install
| 1
|
10,740
| 13,535,169,783
|
IssuesEvent
|
2020-09-16 07:08:34
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
[Bug report]动态组件遍历后,会生成很多个 wx:if=""
|
processing
|
**问题描述**
升级版本到 2.6.16之后,使用 动态组件 <compoent is="">来遍历呈现组件的时候,构建后每个组件上会产生很多 wx:if,然后提示警告:duplicate attribute: wx:if
写法如下:
```html
<component wx:for="{{componentsList}}" class="component" is="{{ item.name }}" componentData="{{ item.data }}" wx:key="uuid"></component>
```
生成以后的代码如下:
```html
<c-title componentData="{{ item.data }}" bindrefresh="refresh" wx:key="uuid" class="component" sid="{{sid}}" bindlogin="__invoke" data-eventconfigs='{{ {login:[["__model","sid","__mpx_event__",["value"],undefined]]} }}' mpxPageStatus="{{mpxPageStatus}}" wx:if='{{item.name === "copyright"}}' wx:if='{{item.name === "c-title"}}' wx:if='{{item.name === "c-splitter"}}' wx:if='{{item.name === "c-shop-header"}}' wx:if='{{item.name === "c-presell"}}' wx:if='{{item.name === "c-offline-store"}}' wx:if='{{item.name === "c-nav"}}' wx:if='{{item.name === "c-goods-list"}}' wx:if='{{item.name === "c-coupon"}}' wx:if='{{item.name === "c-contact"}}' wx:if='{{item.name === "c-article"}}' wx:if='{{item.name === "c-advertisement"}}' wx:if='{{item.name === "c-seckill"}}' wx:if='{{item.name === "c-tuan"}}' wx:if='{{item.name === "pagestatus"}}'></c-title>
```
把我所有组件都wx:if了一遍
|
1.0
|
[Bug report]动态组件遍历后,会生成很多个 wx:if="" - **问题描述**
升级版本到 2.6.16之后,使用 动态组件 <compoent is="">来遍历呈现组件的时候,构建后每个组件上会产生很多 wx:if,然后提示警告:duplicate attribute: wx:if
写法如下:
```html
<component wx:for="{{componentsList}}" class="component" is="{{ item.name }}" componentData="{{ item.data }}" wx:key="uuid"></component>
```
生成以后的代码如下:
```html
<c-title componentData="{{ item.data }}" bindrefresh="refresh" wx:key="uuid" class="component" sid="{{sid}}" bindlogin="__invoke" data-eventconfigs='{{ {login:[["__model","sid","__mpx_event__",["value"],undefined]]} }}' mpxPageStatus="{{mpxPageStatus}}" wx:if='{{item.name === "copyright"}}' wx:if='{{item.name === "c-title"}}' wx:if='{{item.name === "c-splitter"}}' wx:if='{{item.name === "c-shop-header"}}' wx:if='{{item.name === "c-presell"}}' wx:if='{{item.name === "c-offline-store"}}' wx:if='{{item.name === "c-nav"}}' wx:if='{{item.name === "c-goods-list"}}' wx:if='{{item.name === "c-coupon"}}' wx:if='{{item.name === "c-contact"}}' wx:if='{{item.name === "c-article"}}' wx:if='{{item.name === "c-advertisement"}}' wx:if='{{item.name === "c-seckill"}}' wx:if='{{item.name === "c-tuan"}}' wx:if='{{item.name === "pagestatus"}}'></c-title>
```
把我所有组件都wx:if了一遍
|
process
|
动态组件遍历后,会生成很多个 wx if 问题描述 升级版本到 ,使用 动态组件 来遍历呈现组件的时候,构建后每个组件上会产生很多 wx if,然后提示警告:duplicate attribute wx if 写法如下: html 生成以后的代码如下: html 把我所有组件都wx if了一遍
| 1
|
667,687
| 22,496,683,567
|
IssuesEvent
|
2022-06-23 08:13:01
|
fritz-marshal/fritz-beta-feedback
|
https://api.github.com/repos/fritz-marshal/fritz-beta-feedback
|
closed
|
Display full accuracy H0 for cosmology
|
enhancement triage: low priority
|
Hello,
We have noticed that there is a small inconsistency in the cosmological parameters in the About page on Fritz. The current parameters on the page are `FlatLambdaCDM(name=“Planck18_arXiv_v2”, H0=67.7 km / (Mpc s), Om0=0.31, Tcmb0=2.725 K, Neff=3.05, m_nu=[0. 0. 0.06] eV, Ob0=0.049).`
However, if you look at the astropy [`parameters.py`,](https://github.com/astropy/astropy/blob/a3263b683cf91747c062dafe31af994b2848f696/astropy/cosmology/parameters.py#L80) the Planck18_v2 has H0=67.66 km/(Mpc s).
For any source (on the source page, for example: ZTF21abfugtx) it is clear that the luminosity distance being used is **H0=67.66** and not H0=67.7 (as stated on the Fritz about page).
Would it please be possible to update the About page to reflect that H0=67.66 km/(Mps s) that is currently being used to estimate the distance?
Current version: v0.9.dev0+git20210617.fbd9aec+fritz.c76e20c
-Tawny and Andy
|
1.0
|
Display full accuracy H0 for cosmology - Hello,
We have noticed that there is a small inconsistency in the cosmological parameters in the About page on Fritz. The current parameters on the page are `FlatLambdaCDM(name=“Planck18_arXiv_v2”, H0=67.7 km / (Mpc s), Om0=0.31, Tcmb0=2.725 K, Neff=3.05, m_nu=[0. 0. 0.06] eV, Ob0=0.049).`
However, if you look at the astropy [`parameters.py`,](https://github.com/astropy/astropy/blob/a3263b683cf91747c062dafe31af994b2848f696/astropy/cosmology/parameters.py#L80) the Planck18_v2 has H0=67.66 km/(Mpc s).
For any source (on the source page, for example: ZTF21abfugtx) it is clear that the luminosity distance being used is **H0=67.66** and not H0=67.7 (as stated on the Fritz about page).
Would it please be possible to update the About page to reflect that H0=67.66 km/(Mps s) that is currently being used to estimate the distance?
Current version: v0.9.dev0+git20210617.fbd9aec+fritz.c76e20c
-Tawny and Andy
|
non_process
|
display full accuracy for cosmology hello we have noticed that there is a small inconsistency in the cosmological parameters in the about page on fritz the current parameters on the page are flatlambdacdm name “ arxiv ” km mpc s k neff m nu ev however if you look at the astropy the has km mpc s for any source on the source page for example it is clear that the luminosity distance being used is and not as stated on the fritz about page would it please be possible to update the about page to reflect that km mps s that is currently being used to estimate the distance current version fritz tawny and andy
| 0
|
8,802
| 11,908,264,258
|
IssuesEvent
|
2020-03-31 00:26:55
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QgsProcessingParameterMultipleLayers widget does not keep layer order
|
Bug Processing
|
Author Name: **Olivier Dalang** (@olivierdalang)
Original Redmine Issue: [19835](https://issues.qgis.org/issues/19835)
Affected QGIS version: 3.3(master)
Redmine category:processing/gui
---
Hi !
The QgsProcessingParameterMultipleLayers does not keep layer ordering when it is reopened.
This is very annoying for algorithms that are sensitive to the layer ordering and is very error prone as the difference is quite subtle.
*Steps to reproduce*
- Open "merge vector" algorithm (or any other having a multiple layers input)
- Add at least two layers as the multiple layer input
- Change the order of the layer
- Validate (at this point, the reorderd list of layers is set as parameter)
- Reopen the multiple layer input widget
- The layer order is reset to initial state
|
1.0
|
QgsProcessingParameterMultipleLayers widget does not keep layer order - Author Name: **Olivier Dalang** (@olivierdalang)
Original Redmine Issue: [19835](https://issues.qgis.org/issues/19835)
Affected QGIS version: 3.3(master)
Redmine category:processing/gui
---
Hi !
The QgsProcessingParameterMultipleLayers does not keep layer ordering when it is reopened.
This is very annoying for algorithms that are sensitive to the layer ordering and is very error prone as the difference is quite subtle.
*Steps to reproduce*
- Open "merge vector" algorithm (or any other having a multiple layers input)
- Add at least two layers as the multiple layer input
- Change the order of the layer
- Validate (at this point, the reorderd list of layers is set as parameter)
- Reopen the multiple layer input widget
- The layer order is reset to initial state
|
process
|
qgsprocessingparametermultiplelayers widget does not keep layer order author name olivier dalang olivierdalang original redmine issue affected qgis version master redmine category processing gui hi the qgsprocessingparametermultiplelayers does not keep layer ordering when it is reopened this is very annoying for algorithms that are sensitive to the layer ordering and is very error prone as the difference is quite subtle steps to reproduce open merge vector algorithm or any other having a multiple layers input add at least two layers as the multiple layer input change the order of the layer validate at this point the reorderd list of layers is set as parameter reopen the multiple layer input widget the layer order is reset to initial state
| 1
|
6,620
| 5,544,478,156
|
IssuesEvent
|
2017-03-22 19:14:13
|
HvyIndustries/crane
|
https://api.github.com/repos/HvyIndustries/crane
|
closed
|
Improve performance of caching system and parsing
|
improvement performance
|
## Improving writes on caching system :
Actually we flush all the cache into a single file. On big projects, that means every 5 sec after a change on a file, we have a flush over all the tree.
- 5 secs delay does not guarantee consistency of the tree in case of failure.
- By flushing all the tree, 5 secs may be a too short interval
- The tree may be large so I will use CPU & Disk intensively
We could fix this by :
- [ ] Separating ast symbols, one file file
- [ ] As flushing little files, maybe disabling zip support (no more need to use CPU for zip/unzip)
## Avoid parsing too early
Same trick as the cache flushing system, we could wait and trigger parsing every X seconds, instead at each document change, resulting in a decrease of CPU usage. Will be a good practice later for document validation.
|
True
|
Improve performance of caching system and parsing - ## Improving writes on caching system :
Actually we flush all the cache into a single file. On big projects, that means every 5 sec after a change on a file, we have a flush over all the tree.
- 5 secs delay does not guarantee consistency of the tree in case of failure.
- By flushing all the tree, 5 secs may be a too short interval
- The tree may be large so I will use CPU & Disk intensively
We could fix this by :
- [ ] Separating ast symbols, one file file
- [ ] As flushing little files, maybe disabling zip support (no more need to use CPU for zip/unzip)
## Avoid parsing too early
Same trick as the cache flushing system, we could wait and trigger parsing every X seconds, instead at each document change, resulting in a decrease of CPU usage. Will be a good practice later for document validation.
|
non_process
|
improve performance of caching system and parsing improving writes on caching system actually we flush all the cache into a single file on big projects that means every sec after a change on a file we have a flush over all the tree secs delay does not guarantee consistency of the tree in case of failure by flushing all the tree secs may be a too short interval the tree may be large so i will use cpu disk intensively we could fix this by separating ast symbols one file file as flushing little files maybe disabling zip support no more need to use cpu for zip unzip avoid parsing too early same trick as the cache flushing system we could wait and trigger parsing every x seconds instead at each document change resulting in a decrease of cpu usage will be a good practice later for document validation
| 0
|
195,628
| 14,741,837,945
|
IssuesEvent
|
2021-01-07 11:15:20
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
GetResultTests.testToAndFromXContentEmbedded failure
|
:Core/Infra/Core >test-failure Team:Core/Infra
|
**Build scan**: N/A, local build.
**Repro line**:
./gradlew ':server:test' --tests "org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded" -Dtests.seed=E5FC519FC1275FEB -Dtests.security.manager=true -Dtests.locale=ar-AE -Dtests.timezone=Africa/Blantyre -Druntime.java=15
**Reproduces locally?**:
Yes
**Applicable branches**:
master
**Failure excerpt**:
<details><summary>Failure Details</summary>
<p>
```
Suite: Test class org.elasticsearch.index.get.GetResultTests
1> [2021-01-06T03:49:14,797][INFO ][o.e.i.g.GetResultTests ] [testToAndFromXContentEmbedded] before test
1> [2021-01-06T03:49:15,006][INFO ][o.e.i.g.GetResultTests ] [testToAndFromXContentEmbedded] after test
2> REPRODUCE WITH: ./gradlew ':server:test' --tests "org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded" -Dtests.seed=E5FC519FC1275FEB -Dtests.security.manager=true -Dtests.locale=ar-AE -Dtests.timezone=Africa/Blantyre -Druntime.java=15
2> java.lang.AssertionError: expected:<{
"_index" : null,
"_id" : null,
"_seq_no" : 193165536962485811,
"_primary_term" : 71,
"_nested_path" : "Rmhd",
"_routing" : "VYZsPS",
"found" : true,
"_source" : {
"cDdkOed" : {
"kIOmfBuye" : -4355782572476556875
},
"CHPTRkb" : 0.1974391328248285,
"ngifuk" : "SIisj",
"YjVWYCSNxE" : {
"fErBpnn" : [
{
"pgdJSUie" : "KYm",
"LAMHSFzle" : 0.5538381620236812
},
{
"xJWtzXYOzq" : 2501515057617439655,
"uILBlbRl" : "CYOBNHnQgQ",
"kPYyInFAX" : 0.7238725410136407,
"kegvNLXO" : -6906501997487968558,
"QRRCPof" : 0.06681025923342088
},
{
"EWvLpm" : "ZYBSyKZl",
"VKbUdec" : 0.8878569548237388,
"mkvlDUag" : 0.035264934026060124,
"cysyqBfZvG" : "fCnF",
"DWiIKgq" : "duJkfxQLUh"
},
{
"XpiKBYLTYV" : 0.8365711755598817,
"nEDSFe" : "NwGW",
"gPlQFoaTqV" : 0.6106422641750852,
"KApbrEZv" : -3187916961566692938,
"YsLfTgF" : "HptyaRpTW"
},
{
"EtCSBHjKHV" : 2032070882424922864,
"XjtZKWu" : "aluUN",
"cdsLri" : 646340918915008831,
"xrYZLhjvg" : "qupmRsWQoA"
}
]
}
},
"fields" : {
"dFszYoJZbp" : [
"7qCb6oSJ8bGYgtic1pRLFNWN5Yu+8baHptebCvGpmaXzk5it6ZqjUee7rfKViLRdBMyJ3qfzg42AyZfun4nziLm17qK+8724hs+t7o2E7oSv7riW6KWnH+Woqe6ihe+MmOSFo3A="
],
"jTooydPG" : [
0.4488774315814221,
0.23740623389882665,
0.6111549990330503,
0.6939449725143741,
0.4813573856150549
],
"POYUYSDipG" : [
"lIT"
],
"FQirH" : [
{
"pzrOO" : false,
"xpySH" : "PrsvrkaFxc",
"Nfuue" : -445787976
}
],
"ajOPP" : [
{
"PxdIx" : 515179785,
"jYEEp" : "aAmcBIYApR",
"BqGTY" : false
},
{
"JRerP" : "xjecjpuTHz",
"ONjWt" : 833395824,
"tcZYJ" : false
},
{
"Rfyki" : 886453855,
"fcNqw" : false,
"obrYj" : "FDxlDkRXFo"
},
{
"fzSZH" : 1344392580,
"vsHAr" : false,
"NkxTf" : "FKZZYrNqIn"
},
{
"dSMsD" : "zyTpYzfuqT",
"wzqdI" : 1140600156,
"CXNJw" : true
}
]
}
}> but was:<{
"_index" : null,
"_id" : null,
"_seq_no" : 193165536962485811,
"_primary_term" : 71,
"_routing" : "VYZsPS",
"_nested_path" : "Rmhd",
"found" : true,
"_source" : {
"cDdkOed" : {
"kIOmfBuye" : -4355782572476556875
},
"CHPTRkb" : 0.1974391328248285,
"ngifuk" : "SIisj",
"YjVWYCSNxE" : {
"fErBpnn" : [
{
"pgdJSUie" : "KYm",
"LAMHSFzle" : 0.5538381620236812
},
{
"xJWtzXYOzq" : 2501515057617439655,
"uILBlbRl" : "CYOBNHnQgQ",
"kPYyInFAX" : 0.7238725410136407,
"kegvNLXO" : -6906501997487968558,
"QRRCPof" : 0.06681025923342088
},
{
"EWvLpm" : "ZYBSyKZl",
"VKbUdec" : 0.8878569548237388,
"mkvlDUag" : 0.035264934026060124,
"cysyqBfZvG" : "fCnF",
"DWiIKgq" : "duJkfxQLUh"
},
{
"XpiKBYLTYV" : 0.8365711755598817,
"nEDSFe" : "NwGW",
"gPlQFoaTqV" : 0.6106422641750852,
"KApbrEZv" : -3187916961566692938,
"YsLfTgF" : "HptyaRpTW"
},
{
"EtCSBHjKHV" : 2032070882424922864,
"XjtZKWu" : "aluUN",
"cdsLri" : 646340918915008831,
"xrYZLhjvg" : "qupmRsWQoA"
}
]
}
},
"fields" : {
"FQirH" : [
{
"pzrOO" : false,
"xpySH" : "PrsvrkaFxc",
"Nfuue" : -445787976
}
],
"dFszYoJZbp" : [
"7qCb6oSJ8bGYgtic1pRLFNWN5Yu+8baHptebCvGpmaXzk5it6ZqjUee7rfKViLRdBMyJ3qfzg42AyZfun4nziLm17qK+8724hs+t7o2E7oSv7riW6KWnH+Woqe6ihe+MmOSFo3A="
],
"jTooydPG" : [
0.4488774315814221,
0.23740623389882665,
0.6111549990330503,
0.6939449725143741,
0.4813573856150549
],
"POYUYSDipG" : [
"lIT"
],
"ajOPP" : [
{
"PxdIx" : 515179785,
"jYEEp" : "aAmcBIYApR",
"BqGTY" : false
},
{
"JRerP" : "xjecjpuTHz",
"ONjWt" : 833395824,
"tcZYJ" : false
},
{
"Rfyki" : 886453855,
"fcNqw" : false,
"obrYj" : "FDxlDkRXFo"
},
{
"fzSZH" : 1344392580,
"vsHAr" : false,
"NkxTf" : "FKZZYrNqIn"
},
{
"dSMsD" : "zyTpYzfuqT",
"wzqdI" : 1140600156,
"CXNJw" : true
}
]
}
}>
at __randomizedtesting.SeedInfo.seed([E5FC519FC1275FEB:B5959AD1926D8B93]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded(GetResultTests.java:120)
```
</p>
</details>
|
1.0
|
GetResultTests.testToAndFromXContentEmbedded failure - **Build scan**: N/A, local build.
**Repro line**:
./gradlew ':server:test' --tests "org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded" -Dtests.seed=E5FC519FC1275FEB -Dtests.security.manager=true -Dtests.locale=ar-AE -Dtests.timezone=Africa/Blantyre -Druntime.java=15
**Reproduces locally?**:
Yes
**Applicable branches**:
master
**Failure excerpt**:
<details><summary>Failure Details</summary>
<p>
```
Suite: Test class org.elasticsearch.index.get.GetResultTests
1> [2021-01-06T03:49:14,797][INFO ][o.e.i.g.GetResultTests ] [testToAndFromXContentEmbedded] before test
1> [2021-01-06T03:49:15,006][INFO ][o.e.i.g.GetResultTests ] [testToAndFromXContentEmbedded] after test
2> REPRODUCE WITH: ./gradlew ':server:test' --tests "org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded" -Dtests.seed=E5FC519FC1275FEB -Dtests.security.manager=true -Dtests.locale=ar-AE -Dtests.timezone=Africa/Blantyre -Druntime.java=15
2> java.lang.AssertionError: expected:<{
"_index" : null,
"_id" : null,
"_seq_no" : 193165536962485811,
"_primary_term" : 71,
"_nested_path" : "Rmhd",
"_routing" : "VYZsPS",
"found" : true,
"_source" : {
"cDdkOed" : {
"kIOmfBuye" : -4355782572476556875
},
"CHPTRkb" : 0.1974391328248285,
"ngifuk" : "SIisj",
"YjVWYCSNxE" : {
"fErBpnn" : [
{
"pgdJSUie" : "KYm",
"LAMHSFzle" : 0.5538381620236812
},
{
"xJWtzXYOzq" : 2501515057617439655,
"uILBlbRl" : "CYOBNHnQgQ",
"kPYyInFAX" : 0.7238725410136407,
"kegvNLXO" : -6906501997487968558,
"QRRCPof" : 0.06681025923342088
},
{
"EWvLpm" : "ZYBSyKZl",
"VKbUdec" : 0.8878569548237388,
"mkvlDUag" : 0.035264934026060124,
"cysyqBfZvG" : "fCnF",
"DWiIKgq" : "duJkfxQLUh"
},
{
"XpiKBYLTYV" : 0.8365711755598817,
"nEDSFe" : "NwGW",
"gPlQFoaTqV" : 0.6106422641750852,
"KApbrEZv" : -3187916961566692938,
"YsLfTgF" : "HptyaRpTW"
},
{
"EtCSBHjKHV" : 2032070882424922864,
"XjtZKWu" : "aluUN",
"cdsLri" : 646340918915008831,
"xrYZLhjvg" : "qupmRsWQoA"
}
]
}
},
"fields" : {
"dFszYoJZbp" : [
"7qCb6oSJ8bGYgtic1pRLFNWN5Yu+8baHptebCvGpmaXzk5it6ZqjUee7rfKViLRdBMyJ3qfzg42AyZfun4nziLm17qK+8724hs+t7o2E7oSv7riW6KWnH+Woqe6ihe+MmOSFo3A="
],
"jTooydPG" : [
0.4488774315814221,
0.23740623389882665,
0.6111549990330503,
0.6939449725143741,
0.4813573856150549
],
"POYUYSDipG" : [
"lIT"
],
"FQirH" : [
{
"pzrOO" : false,
"xpySH" : "PrsvrkaFxc",
"Nfuue" : -445787976
}
],
"ajOPP" : [
{
"PxdIx" : 515179785,
"jYEEp" : "aAmcBIYApR",
"BqGTY" : false
},
{
"JRerP" : "xjecjpuTHz",
"ONjWt" : 833395824,
"tcZYJ" : false
},
{
"Rfyki" : 886453855,
"fcNqw" : false,
"obrYj" : "FDxlDkRXFo"
},
{
"fzSZH" : 1344392580,
"vsHAr" : false,
"NkxTf" : "FKZZYrNqIn"
},
{
"dSMsD" : "zyTpYzfuqT",
"wzqdI" : 1140600156,
"CXNJw" : true
}
]
}
}> but was:<{
"_index" : null,
"_id" : null,
"_seq_no" : 193165536962485811,
"_primary_term" : 71,
"_routing" : "VYZsPS",
"_nested_path" : "Rmhd",
"found" : true,
"_source" : {
"cDdkOed" : {
"kIOmfBuye" : -4355782572476556875
},
"CHPTRkb" : 0.1974391328248285,
"ngifuk" : "SIisj",
"YjVWYCSNxE" : {
"fErBpnn" : [
{
"pgdJSUie" : "KYm",
"LAMHSFzle" : 0.5538381620236812
},
{
"xJWtzXYOzq" : 2501515057617439655,
"uILBlbRl" : "CYOBNHnQgQ",
"kPYyInFAX" : 0.7238725410136407,
"kegvNLXO" : -6906501997487968558,
"QRRCPof" : 0.06681025923342088
},
{
"EWvLpm" : "ZYBSyKZl",
"VKbUdec" : 0.8878569548237388,
"mkvlDUag" : 0.035264934026060124,
"cysyqBfZvG" : "fCnF",
"DWiIKgq" : "duJkfxQLUh"
},
{
"XpiKBYLTYV" : 0.8365711755598817,
"nEDSFe" : "NwGW",
"gPlQFoaTqV" : 0.6106422641750852,
"KApbrEZv" : -3187916961566692938,
"YsLfTgF" : "HptyaRpTW"
},
{
"EtCSBHjKHV" : 2032070882424922864,
"XjtZKWu" : "aluUN",
"cdsLri" : 646340918915008831,
"xrYZLhjvg" : "qupmRsWQoA"
}
]
}
},
"fields" : {
"FQirH" : [
{
"pzrOO" : false,
"xpySH" : "PrsvrkaFxc",
"Nfuue" : -445787976
}
],
"dFszYoJZbp" : [
"7qCb6oSJ8bGYgtic1pRLFNWN5Yu+8baHptebCvGpmaXzk5it6ZqjUee7rfKViLRdBMyJ3qfzg42AyZfun4nziLm17qK+8724hs+t7o2E7oSv7riW6KWnH+Woqe6ihe+MmOSFo3A="
],
"jTooydPG" : [
0.4488774315814221,
0.23740623389882665,
0.6111549990330503,
0.6939449725143741,
0.4813573856150549
],
"POYUYSDipG" : [
"lIT"
],
"ajOPP" : [
{
"PxdIx" : 515179785,
"jYEEp" : "aAmcBIYApR",
"BqGTY" : false
},
{
"JRerP" : "xjecjpuTHz",
"ONjWt" : 833395824,
"tcZYJ" : false
},
{
"Rfyki" : 886453855,
"fcNqw" : false,
"obrYj" : "FDxlDkRXFo"
},
{
"fzSZH" : 1344392580,
"vsHAr" : false,
"NkxTf" : "FKZZYrNqIn"
},
{
"dSMsD" : "zyTpYzfuqT",
"wzqdI" : 1140600156,
"CXNJw" : true
}
]
}
}>
at __randomizedtesting.SeedInfo.seed([E5FC519FC1275FEB:B5959AD1926D8B93]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.elasticsearch.index.get.GetResultTests.testToAndFromXContentEmbedded(GetResultTests.java:120)
```
</p>
</details>
|
non_process
|
getresulttests testtoandfromxcontentembedded failure build scan n a local build repro line gradlew server test tests org elasticsearch index get getresulttests testtoandfromxcontentembedded dtests seed dtests security manager true dtests locale ar ae dtests timezone africa blantyre druntime java reproduces locally yes applicable branches master failure excerpt failure details suite test class org elasticsearch index get getresulttests before test after test reproduce with gradlew server test tests org elasticsearch index get getresulttests testtoandfromxcontentembedded dtests seed dtests security manager true dtests locale ar ae dtests timezone africa blantyre druntime java java lang assertionerror expected index null id null seq no primary term nested path rmhd routing vyzsps found true source cddkoed kiomfbuye chptrkb ngifuk siisj yjvwycsnxe ferbpnn pgdjsuie kym lamhsfzle xjwtzxyozq uilblbrl cyobnhnqgq kpyyinfax kegvnlxo qrrcpof ewvlpm zybsykzl vkbudec mkvlduag cysyqbfzvg fcnf dwiikgq dujkfxqluh xpikbyltyv nedsfe nwgw gplqfoatqv kapbrezv yslftgf hptyarptw etcsbhjkhv xjtzkwu aluun cdslri xryzlhjvg qupmrswqoa fields dfszyojzbp jtooydpg poyuysdipg lit fqirh pzroo false xpysh prsvrkafxc nfuue ajopp pxdix jyeep aamcbiyapr bqgty false jrerp xjecjputhz onjwt tczyj false rfyki fcnqw false obryj fdxldkrxfo fzszh vshar false nkxtf fkzzyrnqin dsmsd zytpyzfuqt wzqdi cxnjw true but was index null id null seq no primary term routing vyzsps nested path rmhd found true source cddkoed kiomfbuye chptrkb ngifuk siisj yjvwycsnxe ferbpnn pgdjsuie kym lamhsfzle xjwtzxyozq uilblbrl cyobnhnqgq kpyyinfax kegvnlxo qrrcpof ewvlpm zybsykzl vkbudec mkvlduag cysyqbfzvg fcnf dwiikgq dujkfxqluh xpikbyltyv nedsfe nwgw gplqfoatqv kapbrezv yslftgf hptyarptw etcsbhjkhv xjtzkwu aluun cdslri xryzlhjvg qupmrswqoa fields fqirh pzroo false xpysh prsvrkafxc nfuue dfszyojzbp jtooydpg poyuysdipg lit ajopp pxdix jyeep aamcbiyapr bqgty false jrerp xjecjputhz onjwt tczyj false rfyki fcnqw false obryj fdxldkrxfo fzszh vshar false nkxtf fkzzyrnqin dsmsd zytpyzfuqt wzqdi cxnjw true at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org elasticsearch index get getresulttests testtoandfromxcontentembedded getresulttests java
| 0
|
12,307
| 14,859,389,951
|
IssuesEvent
|
2021-01-18 18:23:54
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
REQUESTED FILES (URLS): Max T.S. is wrong (in html report)
|
bug duplicate log-processing
|
When processing a list of urls the value of Max T.S. in REQUESTED FILES (URLS) is wrong.

This is easily seen as the Max value is less than the average value (see screenshot).
Attached is an anonymized nginx log. The command line to read the log and create the report was:
```
cat anonymized.log \
goaccess \
--log-format '%v %h %^ %^ [%d:%t %^] "%r" %s %b %T "%u" "%^"' \
--date-format %d/%b/%Y \
--time-format %H:%M:%S \
-o report_anonymized.html
```
[anonymized.log](https://github.com/allinurl/goaccess/files/5782073/anonymized.log)
as a test you can extract the seconds from above log file using
```
cut -d' ' -f 12 anonymized.log | sed -e 's/\./,/g' > anonymized.csv
```
and calculate the values in excel:
sum and avg are correct, but max should be 4,491s instead of 59ms
|
1.0
|
REQUESTED FILES (URLS): Max T.S. is wrong (in html report) - When processing a list of urls the value of Max T.S. in REQUESTED FILES (URLS) is wrong.

This is easily seen as the Max value is less than the average value (see screenshot).
Attached is an anonymized nginx log. The command line to read the log and create the report was:
```
cat anonymized.log \
goaccess \
--log-format '%v %h %^ %^ [%d:%t %^] "%r" %s %b %T "%u" "%^"' \
--date-format %d/%b/%Y \
--time-format %H:%M:%S \
-o report_anonymized.html
```
[anonymized.log](https://github.com/allinurl/goaccess/files/5782073/anonymized.log)
as a test you can extract the seconds from above log file using
```
cut -d' ' -f 12 anonymized.log | sed -e 's/\./,/g' > anonymized.csv
```
and calculate the values in excel:
sum and avg are correct, but max should be 4,491s instead of 59ms
|
process
|
requested files urls max t s is wrong in html report when processing a list of urls the value of max t s in requested files urls is wrong this is easily seen as the max value is less than the average value see screenshot attached is an anonymized nginx log the command line to read the log and create the report was cat anonymized log goaccess log format v h r s b t u date format d b y time format h m s o report anonymized html as a test you can extract the seconds from above log file using cut d f anonymized log sed e s g anonymized csv and calculate the values in excel sum and avg are correct but max should be instead of
| 1
|
20,098
| 26,631,008,515
|
IssuesEvent
|
2023-01-24 17:50:10
|
0xPolygonMiden/miden-vm
|
https://api.github.com/repos/0xPolygonMiden/miden-vm
|
opened
|
Optimize `mtree_set` and `mtree_cwm`
|
processor
|
As part of https://github.com/0xPolygonMiden/miden-vm/issues/646 and https://github.com/0xPolygonMiden/miden-vm/issues/653 the instructions `mtree_set` and `mtree_cwm` had their output formats changed.
This was done via the assembler, which introduces a couple of new core vm opcode for each masm assembly instruction. This issue is for optimizing those.
Follow up to https://github.com/0xPolygonMiden/miden-vm/pull/652#pullrequestreview-1266746611
|
1.0
|
Optimize `mtree_set` and `mtree_cwm` - As part of https://github.com/0xPolygonMiden/miden-vm/issues/646 and https://github.com/0xPolygonMiden/miden-vm/issues/653 the instructions `mtree_set` and `mtree_cwm` had their output formats changed.
This was done via the assembler, which introduces a couple of new core vm opcode for each masm assembly instruction. This issue is for optimizing those.
Follow up to https://github.com/0xPolygonMiden/miden-vm/pull/652#pullrequestreview-1266746611
|
process
|
optimize mtree set and mtree cwm as part of and the instructions mtree set and mtree cwm had their output formats changed this was done via the assembler which introduces a couple of new core vm opcode for each masm assembly instruction this issue is for optimizing those follow up to
| 1
|
1,923
| 4,760,242,578
|
IssuesEvent
|
2016-10-25 02:09:44
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Oracle Queries fail - Unable to display Raw Data
|
Bug Database/Oracle Query Processor
|
- Browser: Chrome 53.0.2785.143 m
- OS: Windows 7
- Oracle 10.2 with Ojdbc7
- Metabase version 0.20.0
- Metabase hosting environment: Linux Debian 6
- Metabase internal database: H2
As a issue subject I have a problem during "new question" when my request is for raw data . Indipendently from table’s rows/columns count I get always the error attached.
I've not tested the issue over others DB.

|
1.0
|
Oracle Queries fail - Unable to display Raw Data -
- Browser: Chrome 53.0.2785.143 m
- OS: Windows 7
- Oracle 10.2 with Ojdbc7
- Metabase version 0.20.0
- Metabase hosting environment: Linux Debian 6
- Metabase internal database: H2
As a issue subject I have a problem during "new question" when my request is for raw data . Indipendently from table’s rows/columns count I get always the error attached.
I've not tested the issue over others DB.

|
process
|
oracle queries fail unable to display raw data browser chrome m os windows oracle with metabase version metabase hosting environment linux debian metabase internal database as a issue subject i have a problem during new question when my request is for raw data indipendently from table’s rows columns count i get always the error attached i ve not tested the issue over others db
| 1
|
262,767
| 22,957,905,651
|
IssuesEvent
|
2022-07-19 13:11:39
|
cobudget/cobudget
|
https://api.github.com/repos/cobudget/cobudget
|
closed
|
[FEATURE] Only allow group admins to create group round
|
needs testing
|
Free round creation causes too much confusion and accidental rounds as participants think that a round is a bucket.
We should limit round creation in groups to group admins.
|
1.0
|
[FEATURE] Only allow group admins to create group round - Free round creation causes too much confusion and accidental rounds as participants think that a round is a bucket.
We should limit round creation in groups to group admins.
|
non_process
|
only allow group admins to create group round free round creation causes too much confusion and accidental rounds as participants think that a round is a bucket we should limit round creation in groups to group admins
| 0
|
18,172
| 24,215,859,859
|
IssuesEvent
|
2022-09-26 06:39:57
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QgsProcessingParameterFile warning messages and macOS: error unknown
|
Processing Regression Bug macOS
|
### What is the bug or the crash?
The QgsProcessingParameterFile processing parameter can be used to collect file pathes for a QgsProcessingAlgorithms.
This is usefull to collect none-standard / none-geospatial file formats which are required to run an algorithm.
However,
1. calling QgsProcessingParameterFile functions like `.valueAsString(...)`, `.valueAsPythonString(...)` or `.valueAsJsonObject(...)` cause that other dataproviders (gdal, ogr, pdal, ...) try opening the file path collected by the QgsProcessingParameterFile . As this fails, warning messages are shown. This is missleading, because at this point nobody expects a data provider to open the file.


2. In case of macOS the hidden tests agains different data providers can lead to a crash with Exception: unknown. This can kill the algorithm execution, e.g. when running it from a AlgorithmDialog where the algorithms `.asMap(...)` function is called to save the history details, which itself will calls the parameters `.valueAsJsonObject(...)` .
```
/Applications/QGIS.app/Contents/MacOS/bin/python3 "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/pydevd.py" --multiproc --qt-support=auto --client 127.0.0.1 --port 49522 --file /Users/jakimowb/Repositories/enmap-box/tests/src/issues/test_issue_1330_QGIS.py
Connected to pydev debugger (build 211.7442.45)
Application state:
QGIS_PREFIX_PATH env var: /Applications/QGIS.app/contents/MacOS
Prefix: /Applications/QGIS.app/contents/MacOS
Plugin Path: /Applications/QGIS.app/contents/MacOS/../PlugIns/qgis
Package Data Path: /Applications/QGIS.app/contents/MacOS/../Resources
Active Theme Name:
Active Theme Path: /Applications/QGIS.app/contents/MacOS/../Resources/resources/themes//icons/
Default Theme Path: :/images/themes/default/
SVG Search Paths: /Applications/QGIS.app/contents/MacOS/../Resources/svg/
/tmp/QGIS-PythonTestConfigPathfe6_5gvl/profiles/default/svg/
User DB Path: /Applications/QGIS.app/contents/MacOS/../Resources/resources/qgis.db
Auth DB Path: /tmp/QGIS-PythonTestConfigPathfe6_5gvl/profiles/default/qgis-auth.db
OGR(1): Cannot open /private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt ().()
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR: Status 3: Unable to load mesh (null)
Traceback (most recent call last):
File "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/pydevd.py", line 1483, in _exec
pydev_imports.execfile(file, globals, locals) # execute the script
File "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/_pydev_imps/_pydev_execfile.py", line 18, in execfile
exec(compile(contents+"\n", file, 'exec'), glob, loc)
File "/Users/jakimowb/Repositories/enmap-box/tests/src/issues/test_issue_1330_QGIS.py", line 12, in <module>
print(f'asString={parameter.valueAsString(path, context)}')
Exception: unknown
```
### Steps to reproduce the issue
run this code in a macOS QGIS python shell:
```
import pathlib
context = QgsProcessingContext()
parameter = QgsProcessingParameterFile('file', fileFilter='Any file (*.*)')
path = QgsProcessingUtils.generateTempFilename('tempfile.txt')
with open(path, 'w') as f:
f.write('Dummy')
assert pathlib.Path(path).is_file()
assert parameter.checkValueIsAcceptable(path, context)
print(f'asString={parameter.valueAsString(path, context)}')
print(f'asPythonString={parameter.valueAsPythonString(path, context)}')
print(f'asJsonObject={parameter.valueAsJsonObject(path, context)}')
```
### Versions
QGIS for macOS 3.24
QGIS OSGeo4W 3.25 rev [95a55f9782](https://github.com/qgis/QGIS/commit/95a55f9782)
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
QgsProcessingParameterFile warning messages and macOS: error unknown - ### What is the bug or the crash?
The QgsProcessingParameterFile processing parameter can be used to collect file pathes for a QgsProcessingAlgorithms.
This is usefull to collect none-standard / none-geospatial file formats which are required to run an algorithm.
However,
1. calling QgsProcessingParameterFile functions like `.valueAsString(...)`, `.valueAsPythonString(...)` or `.valueAsJsonObject(...)` cause that other dataproviders (gdal, ogr, pdal, ...) try opening the file path collected by the QgsProcessingParameterFile . As this fails, warning messages are shown. This is missleading, because at this point nobody expects a data provider to open the file.


2. In case of macOS the hidden tests agains different data providers can lead to a crash with Exception: unknown. This can kill the algorithm execution, e.g. when running it from a AlgorithmDialog where the algorithms `.asMap(...)` function is called to save the history details, which itself will calls the parameters `.valueAsJsonObject(...)` .
```
/Applications/QGIS.app/Contents/MacOS/bin/python3 "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/pydevd.py" --multiproc --qt-support=auto --client 127.0.0.1 --port 49522 --file /Users/jakimowb/Repositories/enmap-box/tests/src/issues/test_issue_1330_QGIS.py
Connected to pydev debugger (build 211.7442.45)
Application state:
QGIS_PREFIX_PATH env var: /Applications/QGIS.app/contents/MacOS
Prefix: /Applications/QGIS.app/contents/MacOS
Plugin Path: /Applications/QGIS.app/contents/MacOS/../PlugIns/qgis
Package Data Path: /Applications/QGIS.app/contents/MacOS/../Resources
Active Theme Name:
Active Theme Path: /Applications/QGIS.app/contents/MacOS/../Resources/resources/themes//icons/
Default Theme Path: :/images/themes/default/
SVG Search Paths: /Applications/QGIS.app/contents/MacOS/../Resources/svg/
/tmp/QGIS-PythonTestConfigPathfe6_5gvl/profiles/default/svg/
User DB Path: /Applications/QGIS.app/contents/MacOS/../Resources/resources/qgis.db
Auth DB Path: /tmp/QGIS-PythonTestConfigPathfe6_5gvl/profiles/default/qgis-auth.db
OGR(1): Cannot open /private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt ().()
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR 4: `/private/var/folders/1d/23g39kb961387shpwfm6lh280000gn/T/processing_wqOGjl/f4ef3715d3e148c6a67fcf8e8ed255b4/tempfile.txt' not recognized as a supported file format.
ERROR: Status 3: Unable to load mesh (null)
Traceback (most recent call last):
File "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/pydevd.py", line 1483, in _exec
pydev_imports.execfile(file, globals, locals) # execute the script
File "/Applications/PyCharm CE.app/Contents/plugins/python-ce/helpers/pydev/_pydev_imps/_pydev_execfile.py", line 18, in execfile
exec(compile(contents+"\n", file, 'exec'), glob, loc)
File "/Users/jakimowb/Repositories/enmap-box/tests/src/issues/test_issue_1330_QGIS.py", line 12, in <module>
print(f'asString={parameter.valueAsString(path, context)}')
Exception: unknown
```
### Steps to reproduce the issue
run this code in a macOS QGIS python shell:
```
import pathlib
context = QgsProcessingContext()
parameter = QgsProcessingParameterFile('file', fileFilter='Any file (*.*)')
path = QgsProcessingUtils.generateTempFilename('tempfile.txt')
with open(path, 'w') as f:
f.write('Dummy')
assert pathlib.Path(path).is_file()
assert parameter.checkValueIsAcceptable(path, context)
print(f'asString={parameter.valueAsString(path, context)}')
print(f'asPythonString={parameter.valueAsPythonString(path, context)}')
print(f'asJsonObject={parameter.valueAsJsonObject(path, context)}')
```
### Versions
QGIS for macOS 3.24
QGIS OSGeo4W 3.25 rev [95a55f9782](https://github.com/qgis/QGIS/commit/95a55f9782)
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
qgsprocessingparameterfile warning messages and macos error unknown what is the bug or the crash the qgsprocessingparameterfile processing parameter can be used to collect file pathes for a qgsprocessingalgorithms this is usefull to collect none standard none geospatial file formats which are required to run an algorithm however calling qgsprocessingparameterfile functions like valueasstring valueaspythonstring or valueasjsonobject cause that other dataproviders gdal ogr pdal try opening the file path collected by the qgsprocessingparameterfile as this fails warning messages are shown this is missleading because at this point nobody expects a data provider to open the file in case of macos the hidden tests agains different data providers can lead to a crash with exception unknown this can kill the algorithm execution e g when running it from a algorithmdialog where the algorithms asmap function is called to save the history details which itself will calls the parameters valueasjsonobject applications qgis app contents macos bin applications pycharm ce app contents plugins python ce helpers pydev pydevd py multiproc qt support auto client port file users jakimowb repositories enmap box tests src issues test issue qgis py connected to pydev debugger build application state qgis prefix path env var applications qgis app contents macos prefix applications qgis app contents macos plugin path applications qgis app contents macos plugins qgis package data path applications qgis app contents macos resources active theme name active theme path applications qgis app contents macos resources resources themes icons default theme path images themes default svg search paths applications qgis app contents macos resources svg tmp qgis profiles default svg user db path applications qgis app contents macos resources resources qgis db auth db path tmp qgis profiles default qgis auth db ogr cannot open private var folders t processing wqogjl tempfile txt error private var folders t processing wqogjl tempfile txt not recognized as a supported file format error private var folders t processing wqogjl tempfile txt not recognized as a supported file format error private var folders t processing wqogjl tempfile txt not recognized as a supported file format error private var folders t processing wqogjl tempfile txt not recognized as a supported file format error status unable to load mesh null traceback most recent call last file applications pycharm ce app contents plugins python ce helpers pydev pydevd py line in exec pydev imports execfile file globals locals execute the script file applications pycharm ce app contents plugins python ce helpers pydev pydev imps pydev execfile py line in execfile exec compile contents n file exec glob loc file users jakimowb repositories enmap box tests src issues test issue qgis py line in print f asstring parameter valueasstring path context exception unknown steps to reproduce the issue run this code in a macos qgis python shell import pathlib context qgsprocessingcontext parameter qgsprocessingparameterfile file filefilter any file path qgsprocessingutils generatetempfilename tempfile txt with open path w as f f write dummy assert pathlib path path is file assert parameter checkvalueisacceptable path context print f asstring parameter valueasstring path context print f aspythonstring parameter valueaspythonstring path context print f asjsonobject parameter valueasjsonobject path context versions qgis for macos qgis rev supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
2,232
| 5,081,815,411
|
IssuesEvent
|
2016-12-29 12:30:00
|
our-city-app/oca-backend
|
https://api.github.com/repos/our-city-app/oca-backend
|
closed
|
Admin role for RegioManagerTeam
|
process_duplicate type_feature
|
We want to have the possibility to flag a RegioManagerTeam member as admin. Meaning that he can do everything that a shop admin can do, BUT limited to the apps of the RegioManagerTeam.
Right now, a shop admin can:
- manage:
- loyalty slides
- city vouchers
- locations
- app circles
- dashboard news
- outstanding orders
- outstanding charges
- prospects
- legal entities
- regiomanagers and teams
- q&a
- change the manager of an order
- cancel subscriptions
|
1.0
|
Admin role for RegioManagerTeam - We want to have the possibility to flag a RegioManagerTeam member as admin. Meaning that he can do everything that a shop admin can do, BUT limited to the apps of the RegioManagerTeam.
Right now, a shop admin can:
- manage:
- loyalty slides
- city vouchers
- locations
- app circles
- dashboard news
- outstanding orders
- outstanding charges
- prospects
- legal entities
- regiomanagers and teams
- q&a
- change the manager of an order
- cancel subscriptions
|
process
|
admin role for regiomanagerteam we want to have the possibility to flag a regiomanagerteam member as admin meaning that he can do everything that a shop admin can do but limited to the apps of the regiomanagerteam right now a shop admin can manage loyalty slides city vouchers locations app circles dashboard news outstanding orders outstanding charges prospects legal entities regiomanagers and teams q a change the manager of an order cancel subscriptions
| 1
|
21,796
| 30,308,170,579
|
IssuesEvent
|
2023-07-10 10:56:42
|
atc0005/check-process
|
https://api.github.com/repos/atc0005/check-process
|
closed
|
README missing performance data metrics table
|
bug documentation App: check_process output/perfdata
|
While reviewing the README I noticed that performance data metrics are not specified.
|
1.0
|
README missing performance data metrics table - While reviewing the README I noticed that performance data metrics are not specified.
|
process
|
readme missing performance data metrics table while reviewing the readme i noticed that performance data metrics are not specified
| 1
|
102,175
| 12,749,094,695
|
IssuesEvent
|
2020-06-26 21:44:48
|
elastic/logstash
|
https://api.github.com/repos/elastic/logstash
|
closed
|
Support list configs via env variables
|
design
|
The idea here came from: https://github.com/logstash-plugins/logstash-output-elasticsearch/issues/657 , where @eisig pointed out that you can't set a list of ES output hosts via an env var.
I'm not sure what the best approach is here, but its a good idea.
|
1.0
|
Support list configs via env variables - The idea here came from: https://github.com/logstash-plugins/logstash-output-elasticsearch/issues/657 , where @eisig pointed out that you can't set a list of ES output hosts via an env var.
I'm not sure what the best approach is here, but its a good idea.
|
non_process
|
support list configs via env variables the idea here came from where eisig pointed out that you can t set a list of es output hosts via an env var i m not sure what the best approach is here but its a good idea
| 0
|
31,259
| 6,476,307,764
|
IssuesEvent
|
2017-08-17 22:32:06
|
opencaching/opencaching-pl
|
https://api.github.com/repos/opencaching/opencaching-pl
|
closed
|
Call to undefined method powerTrailController::findPtOwners
|
Component_GeoPaths Priority_Medium Type_Defect
|
From OC PL error.log:
`[Wed Jun 21 05:34:02.724232 2017] [:error] [pid 3493] [client 62.87.254.252:40791] PHP Fatal error: Call to undefined method powerTrailController::findPtOwners() in /srv/ocpl/powerTrail/ajaxremoveUserFromPt.php on line 25, referer: https://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=597`
|
1.0
|
Call to undefined method powerTrailController::findPtOwners - From OC PL error.log:
`[Wed Jun 21 05:34:02.724232 2017] [:error] [pid 3493] [client 62.87.254.252:40791] PHP Fatal error: Call to undefined method powerTrailController::findPtOwners() in /srv/ocpl/powerTrail/ajaxremoveUserFromPt.php on line 25, referer: https://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=597`
|
non_process
|
call to undefined method powertrailcontroller findptowners from oc pl error log php fatal error call to undefined method powertrailcontroller findptowners in srv ocpl powertrail ajaxremoveuserfrompt php on line referer
| 0
|
2,156
| 5,006,163,675
|
IssuesEvent
|
2016-12-12 13:17:40
|
openvstorage/volumedriver
|
https://api.github.com/repos/openvstorage/volumedriver
|
closed
|
live migrate has non-deterministic outcome
|
priority_critical process_duplicate
|
As observed on OVH demo environment
Scenario: vm running fio, migrate to another location
Observed outcome: intermittent failure/success
VM had read-only fs, it got an IO error and for some reason HA kicked in
Potential issue: race between threshold on voldrv side and edge following, as suggested by @redlicha: the threshold on voldrv side could be set to 0, but this might not be what we want in this kind of demo?
So when it works it is by chance and not by design, the edge client checks the location periodically and will connect to the current voldrv ...
This needs to be investigated further as suggested to be able to give a proper demo of this feature
Might be related to a restarted proxy on the OVH environment, which could have triggered HA ...
|
1.0
|
live migrate has non-deterministic outcome - As observed on OVH demo environment
Scenario: vm running fio, migrate to another location
Observed outcome: intermittent failure/success
VM had read-only fs, it got an IO error and for some reason HA kicked in
Potential issue: race between threshold on voldrv side and edge following, as suggested by @redlicha: the threshold on voldrv side could be set to 0, but this might not be what we want in this kind of demo?
So when it works it is by chance and not by design, the edge client checks the location periodically and will connect to the current voldrv ...
This needs to be investigated further as suggested to be able to give a proper demo of this feature
Might be related to a restarted proxy on the OVH environment, which could have triggered HA ...
|
process
|
live migrate has non deterministic outcome as observed on ovh demo environment scenario vm running fio migrate to another location observed outcome intermittent failure success vm had read only fs it got an io error and for some reason ha kicked in potential issue race between threshold on voldrv side and edge following as suggested by redlicha the threshold on voldrv side could be set to but this might not be what we want in this kind of demo so when it works it is by chance and not by design the edge client checks the location periodically and will connect to the current voldrv this needs to be investigated further as suggested to be able to give a proper demo of this feature might be related to a restarted proxy on the ovh environment which could have triggered ha
| 1
|
176,412
| 21,411,032,387
|
IssuesEvent
|
2022-04-22 05:58:59
|
pazhanivel07/frameworks_base_Aosp10_r33
|
https://api.github.com/repos/pazhanivel07/frameworks_base_Aosp10_r33
|
opened
|
CVE-2021-0652 (High) detected in platform_frameworks_baseplatform-tools-29.0.6
|
security vulnerability
|
## CVE-2021-0652 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>platform_frameworks_baseplatform-tools-29.0.6</b></p></summary>
<p>
<p>Library home page: <a href=https://github.com/aosp-mirror/platform_frameworks_base.git>https://github.com/aosp-mirror/platform_frameworks_base.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_Aosp10_r33/commit/d0a412c03562493a433dc7e698ff88ab06a3468a">d0a412c03562493a433dc7e698ff88ab06a3468a</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/graphics/java/android/graphics/drawable/VectorDrawable.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In VectorDrawable::VectorDrawable of VectorDrawable.java, there is a possible way to introduce a memory corruption due to sharing of not thread-safe objects. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-8.1 Android-9 Android-10 Android-11Android ID: A-185178568
<p>Publish Date: 2021-10-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0652>CVE-2021-0652</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-10-01">https://source.android.com/security/bulletin/2021-10-01</a></p>
<p>Release Date: 2021-10-22</p>
<p>Fix Resolution: android-11.0.0_r46</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-0652 (High) detected in platform_frameworks_baseplatform-tools-29.0.6 - ## CVE-2021-0652 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>platform_frameworks_baseplatform-tools-29.0.6</b></p></summary>
<p>
<p>Library home page: <a href=https://github.com/aosp-mirror/platform_frameworks_base.git>https://github.com/aosp-mirror/platform_frameworks_base.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_Aosp10_r33/commit/d0a412c03562493a433dc7e698ff88ab06a3468a">d0a412c03562493a433dc7e698ff88ab06a3468a</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/graphics/java/android/graphics/drawable/VectorDrawable.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In VectorDrawable::VectorDrawable of VectorDrawable.java, there is a possible way to introduce a memory corruption due to sharing of not thread-safe objects. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-8.1 Android-9 Android-10 Android-11Android ID: A-185178568
<p>Publish Date: 2021-10-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0652>CVE-2021-0652</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-10-01">https://source.android.com/security/bulletin/2021-10-01</a></p>
<p>Release Date: 2021-10-22</p>
<p>Fix Resolution: android-11.0.0_r46</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in platform frameworks baseplatform tools cve high severity vulnerability vulnerable library platform frameworks baseplatform tools library home page a href found in head commit a href found in base branch main vulnerable source files graphics java android graphics drawable vectordrawable java vulnerability details in vectordrawable vectordrawable of vectordrawable java there is a possible way to introduce a memory corruption due to sharing of not thread safe objects this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with whitesource
| 0
|
2,454
| 5,238,880,487
|
IssuesEvent
|
2017-01-31 07:33:42
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Column filtering uses equal matching per default since 2.9.2
|
bug inprocess
|
The column filtering changed to eq matching per default in the latest build 2.10.0.
https://jsfiddle.net/69z2wepo/68697/
EDIT:
Problem seems to be introduced already before 2.10.0 release.
Edit:
Came in 2.9.2. -> 2.9.1 the last correct one.
|
1.0
|
Column filtering uses equal matching per default since 2.9.2 - The column filtering changed to eq matching per default in the latest build 2.10.0.
https://jsfiddle.net/69z2wepo/68697/
EDIT:
Problem seems to be introduced already before 2.10.0 release.
Edit:
Came in 2.9.2. -> 2.9.1 the last correct one.
|
process
|
column filtering uses equal matching per default since the column filtering changed to eq matching per default in the latest build edit problem seems to be introduced already before release edit came in the last correct one
| 1
|
17,819
| 23,744,064,255
|
IssuesEvent
|
2022-08-31 14:37:21
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
closed
|
[C++] Nightly Integration Testing Report for Firestore
|
type: process nightly-testing
|
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 6872ba47417796c8c24c6eab435534ad34b7a843
Last updated: Wed Aug 31 05:26 PDT 2022
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/2962975852)**
<hidden value="integration-test-status-comment"></hidden>
***
### ❌ [build against SDK] Integration test FAILED
Requested by @firebase-workflow-trigger[bot] on commit 2d90eba5d3d595fc69db3d74916e3f19f7506c6f
Last updated: Tue Aug 30 07:38 PDT 2022
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/2956172376)**
| Failures | Configs |
|----------|---------|
| missing_log | [TEST] [ERROR] [Android] [1/3 os: windows] [1/5 android_device: emulator_target]<br/> |
| firestore | [TEST] [ERROR] [Android] [2/3 os: macos ubuntu] [1/5 android_device: emulator_target]<br/>[TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_latest]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
|
1.0
|
[C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 6872ba47417796c8c24c6eab435534ad34b7a843
Last updated: Wed Aug 31 05:26 PDT 2022
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/2962975852)**
<hidden value="integration-test-status-comment"></hidden>
***
### ❌ [build against SDK] Integration test FAILED
Requested by @firebase-workflow-trigger[bot] on commit 2d90eba5d3d595fc69db3d74916e3f19f7506c6f
Last updated: Tue Aug 30 07:38 PDT 2022
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/2956172376)**
| Failures | Configs |
|----------|---------|
| missing_log | [TEST] [ERROR] [Android] [1/3 os: windows] [1/5 android_device: emulator_target]<br/> |
| firestore | [TEST] [ERROR] [Android] [2/3 os: macos ubuntu] [1/5 android_device: emulator_target]<br/>[TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_latest]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
|
process
|
nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated wed aug pdt ❌ nbsp integration test failed requested by firebase workflow trigger on commit last updated tue aug pdt failures configs missing log firestore failed tests nbsp nbsp crash timeout add flaky tests to
| 1
|
17,118
| 22,635,825,677
|
IssuesEvent
|
2022-06-30 18:52:32
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[feature][processing] A brand new Extract labels algorithm (Request in QGIS)
|
Processing Alg 3.24
|
### Request for documentation
From pull request QGIS/qgis#46657
Author: @nirvn
QGIS version: 3.24
**[feature][processing] A brand new Extract labels algorithm**
### PR Description:
## Description
This PR adds a brand new Extract labels algorithm to QGIS' processing toolbox. This algorithm extracts label information from a rendered map at a given extent and scale. If a map theme is provided, the rendered map will match the visibility and symbology of that theme. If left blank, the project's layer tree setup will be used.
Extracted label information include its position (served as point geometries), the associated layer name and feature ID, as well as label rotation, text, alignment, font family, size, weight, etc.
Screenshot of the algorithm dialog:

Screenshot showcasing an output of the algorithm:

The algorithm also generates a default symbol-less & label style for the output layer which allows for those saved layers to be drag and dropped into projects and have the label drawn exactly as they were exported.
https://user-images.githubusercontent.com/1728657/147822919-ddc2b698-7565-4d58-921f-b58e68aeec46.mp4
_Funded by [TEKSI](https://www.teksi.ch/)_
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] A brand new Extract labels algorithm"
|
1.0
|
[feature][processing] A brand new Extract labels algorithm (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#46657
Author: @nirvn
QGIS version: 3.24
**[feature][processing] A brand new Extract labels algorithm**
### PR Description:
## Description
This PR adds a brand new Extract labels algorithm to QGIS' processing toolbox. This algorithm extracts label information from a rendered map at a given extent and scale. If a map theme is provided, the rendered map will match the visibility and symbology of that theme. If left blank, the project's layer tree setup will be used.
Extracted label information include its position (served as point geometries), the associated layer name and feature ID, as well as label rotation, text, alignment, font family, size, weight, etc.
Screenshot of the algorithm dialog:

Screenshot showcasing an output of the algorithm:

The algorithm also generates a default symbol-less & label style for the output layer which allows for those saved layers to be drag and dropped into projects and have the label drawn exactly as they were exported.
https://user-images.githubusercontent.com/1728657/147822919-ddc2b698-7565-4d58-921f-b58e68aeec46.mp4
_Funded by [TEKSI](https://www.teksi.ch/)_
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] A brand new Extract labels algorithm"
|
process
|
a brand new extract labels algorithm request in qgis request for documentation from pull request qgis qgis author nirvn qgis version a brand new extract labels algorithm pr description description this pr adds a brand new extract labels algorithm to qgis processing toolbox this algorithm extracts label information from a rendered map at a given extent and scale if a map theme is provided the rendered map will match the visibility and symbology of that theme if left blank the project s layer tree setup will be used extracted label information include its position served as point geometries the associated layer name and feature id as well as label rotation text alignment font family size weight etc screenshot of the algorithm dialog screenshot showcasing an output of the algorithm the algorithm also generates a default symbol less label style for the output layer which allows for those saved layers to be drag and dropped into projects and have the label drawn exactly as they were exported funded by commits tagged with or a brand new extract labels algorithm
| 1
|
7,149
| 10,291,313,478
|
IssuesEvent
|
2019-08-27 15:13:17
|
heim-rs/heim
|
https://api.github.com/repos/heim-rs/heim
|
closed
|
Per-process cpu usage
|
A-process
|
I'm able to get a list of the pids/parent pids/etc for each process, but I was wondering if there was also a way to get the CPU % that each pid is using.
|
1.0
|
Per-process cpu usage - I'm able to get a list of the pids/parent pids/etc for each process, but I was wondering if there was also a way to get the CPU % that each pid is using.
|
process
|
per process cpu usage i m able to get a list of the pids parent pids etc for each process but i was wondering if there was also a way to get the cpu that each pid is using
| 1
|
241,983
| 18,506,929,669
|
IssuesEvent
|
2021-10-19 19:49:58
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Transparent textures are not sorting correctly
|
discussion topic:rendering documentation topic:2d
|
### Godot version
3.3.4
### System information
Windows 10, GLES3, NVIDIA GeForce RTX 2060, 30.0.14.7168
### Issue description
Hello!
I am working on a prototype that uses 2D pixel art tiles applied to simple square meshes in 3D gridmaps. Everything was working very well until I tried upgrading Godot from 3.3.2 to 3.3.3 or 3.3.4
I noticed that 3.3.3 included two PRs, which I have confirmed are causing my issue. On the 3.x branch, I manually reverted those two changes, compiled Godot, and tested my project, and no longer experienced the issue.
#50721
#43506
Since I am using pixel art, transparency is pretty important (especially in sprites, but also in the tiles themselves).
I know that those two PRs above are positioned as fixes themselves, so I am unsure if I was just relying on Godot incorrectly sorting my sprites/tiles, but the way Godot sorted my sprites/meshes in 3.3.2 seems to me to be correct.
Sorting in 3.3.2:

Sorting in 3.3.3+:

I have also noticed the same sorting issues when I compile Godot 4 (master branch) from source. If I have to stay on 3.3.2, I guess I could, but I was really hoping to start taking advantage of the incredible TIleMap editor improvements.
Please let me know if you have any questions!
Thanks!
### Steps to reproduce
1. Open project in Godot 3.3.2
2. Move around in the 3D editor or play the scene, and notice that the "Player" (Godot icon) is correctly drawn in front of cliffs.
3. Open project in Godot 3.3.3
4. Move around in the 3D editor or play the scene, and notice that the "Player" (Godot icon) is drawn behind cliffs when they should be drawn in front of them.
### Minimal reproduction project
[godot-transparency-sort-bug.zip](https://github.com/godotengine/godot/files/7368771/godot-transparency-sort-bug.zip)
|
1.0
|
Transparent textures are not sorting correctly - ### Godot version
3.3.4
### System information
Windows 10, GLES3, NVIDIA GeForce RTX 2060, 30.0.14.7168
### Issue description
Hello!
I am working on a prototype that uses 2D pixel art tiles applied to simple square meshes in 3D gridmaps. Everything was working very well until I tried upgrading Godot from 3.3.2 to 3.3.3 or 3.3.4
I noticed that 3.3.3 included two PRs, which I have confirmed are causing my issue. On the 3.x branch, I manually reverted those two changes, compiled Godot, and tested my project, and no longer experienced the issue.
#50721
#43506
Since I am using pixel art, transparency is pretty important (especially in sprites, but also in the tiles themselves).
I know that those two PRs above are positioned as fixes themselves, so I am unsure if I was just relying on Godot incorrectly sorting my sprites/tiles, but the way Godot sorted my sprites/meshes in 3.3.2 seems to me to be correct.
Sorting in 3.3.2:

Sorting in 3.3.3+:

I have also noticed the same sorting issues when I compile Godot 4 (master branch) from source. If I have to stay on 3.3.2, I guess I could, but I was really hoping to start taking advantage of the incredible TIleMap editor improvements.
Please let me know if you have any questions!
Thanks!
### Steps to reproduce
1. Open project in Godot 3.3.2
2. Move around in the 3D editor or play the scene, and notice that the "Player" (Godot icon) is correctly drawn in front of cliffs.
3. Open project in Godot 3.3.3
4. Move around in the 3D editor or play the scene, and notice that the "Player" (Godot icon) is drawn behind cliffs when they should be drawn in front of them.
### Minimal reproduction project
[godot-transparency-sort-bug.zip](https://github.com/godotengine/godot/files/7368771/godot-transparency-sort-bug.zip)
|
non_process
|
transparent textures are not sorting correctly godot version system information windows nvidia geforce rtx issue description hello i am working on a prototype that uses pixel art tiles applied to simple square meshes in gridmaps everything was working very well until i tried upgrading godot from to or i noticed that included two prs which i have confirmed are causing my issue on the x branch i manually reverted those two changes compiled godot and tested my project and no longer experienced the issue since i am using pixel art transparency is pretty important especially in sprites but also in the tiles themselves i know that those two prs above are positioned as fixes themselves so i am unsure if i was just relying on godot incorrectly sorting my sprites tiles but the way godot sorted my sprites meshes in seems to me to be correct sorting in sorting in i have also noticed the same sorting issues when i compile godot master branch from source if i have to stay on i guess i could but i was really hoping to start taking advantage of the incredible tilemap editor improvements please let me know if you have any questions thanks steps to reproduce open project in godot move around in the editor or play the scene and notice that the player godot icon is correctly drawn in front of cliffs open project in godot move around in the editor or play the scene and notice that the player godot icon is drawn behind cliffs when they should be drawn in front of them minimal reproduction project
| 0
|
2,637
| 5,412,729,766
|
IssuesEvent
|
2017-03-01 15:13:17
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
reopened
|
[subtitles] [en] MÉLENCHON : Réunion publique à Brest - #JLMBrest
|
Language: English Process: [0] Awaiting subtitles
|
## MÉLENCHON : Réunion publique à Brest - #JLMBrest
| Info
----- | -----
**Duration** | 2:13:18 :clock7:
**Language** | English
**Video** | [See it on YouTube](https://www.youtube.com/watch?v=WNwngAR4xg4) :arrow_upper_right:
**Subtitles** | [Edit them in YouTube](https://www.youtube.com/timedtext_editor?v=WNwngAR4xg4&tab=captions&bl=vmp&action_mde_edit_form=1&lang=en&ui=hd) :arrow_upper_right:
|
1.0
|
[subtitles] [en] MÉLENCHON : Réunion publique à Brest - #JLMBrest - ## MÉLENCHON : Réunion publique à Brest - #JLMBrest
| Info
----- | -----
**Duration** | 2:13:18 :clock7:
**Language** | English
**Video** | [See it on YouTube](https://www.youtube.com/watch?v=WNwngAR4xg4) :arrow_upper_right:
**Subtitles** | [Edit them in YouTube](https://www.youtube.com/timedtext_editor?v=WNwngAR4xg4&tab=captions&bl=vmp&action_mde_edit_form=1&lang=en&ui=hd) :arrow_upper_right:
|
process
|
mélenchon réunion publique à brest jlmbrest mélenchon réunion publique à brest jlmbrest nbsp info duration language english video arrow upper right subtitles arrow upper right
| 1
|
7,258
| 10,568,053,389
|
IssuesEvent
|
2019-10-06 10:06:01
|
lastunicorn/MedicX
|
https://api.github.com/repos/lastunicorn/MedicX
|
opened
|
Prevent application exit from "X" button if not saved
|
requirement
|
As a user,
If something is changed and not saved in the application (the star is displayed in the window's title) and I press the "X" button to close the application, I want to have a popup displayed and be presented with three options: Save and close, Do not save and close, Cancel
So that I am prevented to lose unsaved data.
|
1.0
|
Prevent application exit from "X" button if not saved - As a user,
If something is changed and not saved in the application (the star is displayed in the window's title) and I press the "X" button to close the application, I want to have a popup displayed and be presented with three options: Save and close, Do not save and close, Cancel
So that I am prevented to lose unsaved data.
|
non_process
|
prevent application exit from x button if not saved as a user if something is changed and not saved in the application the star is displayed in the window s title and i press the x button to close the application i want to have a popup displayed and be presented with three options save and close do not save and close cancel so that i am prevented to lose unsaved data
| 0
|
561,065
| 16,610,543,213
|
IssuesEvent
|
2021-06-02 10:55:56
|
mantidproject/mantid
|
https://api.github.com/repos/mantidproject/mantid
|
closed
|
Qt thread assertion when running script in debug
|
High Priority ISIS Team: Core
|
### Steps to reproduce
(1) Run this script having built in debug
```
# import mantid algorithms, numpy and matplotlib
from mantid.simpleapi import *
import matplotlib.pyplot as plt
import numpy as np
# SXD23767.raw is available in the TrainingCourseData from the downloads page
SXD23767 = Load(Filename='SXD23767.raw', LoadMonitors='Exclude')
# Set some UB with angles we can play with
SetUB(SXD23767, 1,1,2,90,90,120)
md_non_ortho = ConvertToDiffractionMDWorkspace(InputWorkspace='SXD23767', OutputDimensions='HKL')
```
(2) It crashes on loading the SXD data

### Platforms affected
Windows debug
|
1.0
|
Qt thread assertion when running script in debug - ### Steps to reproduce
(1) Run this script having built in debug
```
# import mantid algorithms, numpy and matplotlib
from mantid.simpleapi import *
import matplotlib.pyplot as plt
import numpy as np
# SXD23767.raw is available in the TrainingCourseData from the downloads page
SXD23767 = Load(Filename='SXD23767.raw', LoadMonitors='Exclude')
# Set some UB with angles we can play with
SetUB(SXD23767, 1,1,2,90,90,120)
md_non_ortho = ConvertToDiffractionMDWorkspace(InputWorkspace='SXD23767', OutputDimensions='HKL')
```
(2) It crashes on loading the SXD data

### Platforms affected
Windows debug
|
non_process
|
qt thread assertion when running script in debug steps to reproduce run this script having built in debug import mantid algorithms numpy and matplotlib from mantid simpleapi import import matplotlib pyplot as plt import numpy as np raw is available in the trainingcoursedata from the downloads page load filename raw loadmonitors exclude set some ub with angles we can play with setub md non ortho converttodiffractionmdworkspace inputworkspace outputdimensions hkl it crashes on loading the sxd data platforms affected windows debug
| 0
|
7,207
| 10,343,232,986
|
IssuesEvent
|
2019-09-04 08:30:56
|
heim-rs/heim
|
https://api.github.com/repos/heim-rs/heim
|
opened
|
process::Process:name is truncated for Linux
|
A-process C-enhancement O-linux
|
`/proc/{pid}/stat` returns truncated (capped to 15 chars) process name, `Process::cmdline` could be used to fetch full process name if necessary.
Blocked by #97
|
1.0
|
process::Process:name is truncated for Linux - `/proc/{pid}/stat` returns truncated (capped to 15 chars) process name, `Process::cmdline` could be used to fetch full process name if necessary.
Blocked by #97
|
process
|
process process name is truncated for linux proc pid stat returns truncated capped to chars process name process cmdline could be used to fetch full process name if necessary blocked by
| 1
|
13,840
| 16,602,300,701
|
IssuesEvent
|
2021-06-01 21:21:40
|
CodeForPhilly/paws-data-pipeline
|
https://api.github.com/repos/CodeForPhilly/paws-data-pipeline
|
opened
|
On server startup, check for incomplete execution
|
Async processes
|
As a step toward full async execution:
When server starts up, check for remnants of an incomplete execution (i.e., non-completed job record in DB). Assuming we can know it's dead[1], delete the in-progress record to allow a new run to be started.
We might want to check execution_status last_update time to make sure there's nothing running in a another process
(From #313)
<hr>
[1] That's the hard part, isn't it?
|
1.0
|
On server startup, check for incomplete execution - As a step toward full async execution:
When server starts up, check for remnants of an incomplete execution (i.e., non-completed job record in DB). Assuming we can know it's dead[1], delete the in-progress record to allow a new run to be started.
We might want to check execution_status last_update time to make sure there's nothing running in a another process
(From #313)
<hr>
[1] That's the hard part, isn't it?
|
process
|
on server startup check for incomplete execution as a step toward full async execution when server starts up check for remnants of an incomplete execution i e non completed job record in db assuming we can know it s dead delete the in progress record to allow a new run to be started we might want to check execution status last update time to make sure there s nothing running in a another process from that s the hard part isn t it
| 1
|
208,466
| 23,605,511,589
|
IssuesEvent
|
2022-08-24 07:55:31
|
ioana-nicolae/second
|
https://api.github.com/repos/ioana-nicolae/second
|
closed
|
CVE-2020-14060 (High) detected in jackson-databind-2.7.9.jar - autoclosed
|
security vulnerability
|
## CVE-2020-14060 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>branch3</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oadd.org.apache.xalan.lib.sql.JNDIConnectionPool (aka apache/drill).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14060>CVE-2020-14060</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14060">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14060</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: 2.8.0.rc1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
True
|
CVE-2020-14060 (High) detected in jackson-databind-2.7.9.jar - autoclosed - ## CVE-2020-14060 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>branch3</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oadd.org.apache.xalan.lib.sql.JNDIConnectionPool (aka apache/drill).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14060>CVE-2020-14060</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14060">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14060</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: 2.8.0.rc1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
non_process
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oadd org apache xalan lib sql jndiconnectionpool aka apache drill publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue
| 0
|
680,434
| 23,269,795,773
|
IssuesEvent
|
2022-08-04 21:24:14
|
layer5io/layer5
|
https://api.github.com/repos/layer5io/layer5
|
closed
|
Picture slider not showing on Community page
|
kind/bug help wanted priority/urgent area/community
|
#### Description
<img width="1440" alt="Screenshot 2022-08-04 at 5 29 45 AM" src="https://user-images.githubusercontent.com/85789734/182736299-28d19b3b-f19e-4381-ba5b-96f12f49d612.png">
The slider is not visible currently.
#### Expected Behavior
https://user-images.githubusercontent.com/85789734/182736336-9b034b27-0c75-42f8-ac3e-1e75d41d6cfc.mov
Restore the Slider on [layer5.io/community](https://layer5.io/community)
#### Environment:
- Host OS:
- Browser:
---
<img src="https://raw.githubusercontent.com/layer5io/layer5/master/.github/assets/images/layer5/5-light-small.svg" width="16px" align="left" /><h3> Contributor Resources and <a href="https://layer5.io/community/handbook">Handbook</a></h3>
The layer5.io website uses Gatsby, React, and GitHub Pages. Site content is found under the [`master` branch](https://github.com/layer5io/layer5/tree/master).
- 📚 See [contributing instructions](https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md)
- 🎨 Wireframes and designs for Layer5 site in [Figma](https://www.figma.com/file/5ZwEkSJwUPitURD59YHMEN/Layer5-Designs).
- 🙋🏾🙋🏼 Questions: [Discussion Forum](https://discuss.layer5.io) and [Community Slack](http://slack.layer5.io)
|
1.0
|
Picture slider not showing on Community page - #### Description
<img width="1440" alt="Screenshot 2022-08-04 at 5 29 45 AM" src="https://user-images.githubusercontent.com/85789734/182736299-28d19b3b-f19e-4381-ba5b-96f12f49d612.png">
The slider is not visible currently.
#### Expected Behavior
https://user-images.githubusercontent.com/85789734/182736336-9b034b27-0c75-42f8-ac3e-1e75d41d6cfc.mov
Restore the Slider on [layer5.io/community](https://layer5.io/community)
#### Environment:
- Host OS:
- Browser:
---
<img src="https://raw.githubusercontent.com/layer5io/layer5/master/.github/assets/images/layer5/5-light-small.svg" width="16px" align="left" /><h3> Contributor Resources and <a href="https://layer5.io/community/handbook">Handbook</a></h3>
The layer5.io website uses Gatsby, React, and GitHub Pages. Site content is found under the [`master` branch](https://github.com/layer5io/layer5/tree/master).
- 📚 See [contributing instructions](https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md)
- 🎨 Wireframes and designs for Layer5 site in [Figma](https://www.figma.com/file/5ZwEkSJwUPitURD59YHMEN/Layer5-Designs).
- 🙋🏾🙋🏼 Questions: [Discussion Forum](https://discuss.layer5.io) and [Community Slack](http://slack.layer5.io)
|
non_process
|
picture slider not showing on community page description img width alt screenshot at am src the slider is not visible currently expected behavior restore the slider on environment host os browser contributor resources and a href the io website uses gatsby react and github pages site content is found under the 📚 see 🎨 wireframes and designs for site in 🙋🏾🙋🏼 questions and
| 0
|
156,160
| 24,579,974,476
|
IssuesEvent
|
2022-10-13 14:56:40
|
microsoft/pyright
|
https://api.github.com/repos/microsoft/pyright
|
closed
|
Infer type of function that is a `TypeGuard` based on the argument
|
as designed
|
**Is your feature request related to a problem? Please describe.**
We'd like to test if a function annotated as `TypeGuard[some_type]` has a value of `True` when an object of `some_type` is passed.
```python
from __future__ import annotations
from typing_extensions import TypeGuard, reveal_type
def fun(x: int | str) -> TypeGuard[int]:
return isinstance(x, int)
reveal_type(fun(3))
reveal_type(fun("abc"))
```
I would have expected the type of `fun(3)` to be `Literal[True]` and the type of `fun("abc")` to be `Literal[False]`, but pyright says that both of them are `bool`
For testing purposes in the pandas stubs, we would use `assert_type(fun(3), Literal[True])`
**Describe the solution you'd like**
Narrowing of the type of the function to `True` or `False`, when possible
|
1.0
|
Infer type of function that is a `TypeGuard` based on the argument - **Is your feature request related to a problem? Please describe.**
We'd like to test if a function annotated as `TypeGuard[some_type]` has a value of `True` when an object of `some_type` is passed.
```python
from __future__ import annotations
from typing_extensions import TypeGuard, reveal_type
def fun(x: int | str) -> TypeGuard[int]:
return isinstance(x, int)
reveal_type(fun(3))
reveal_type(fun("abc"))
```
I would have expected the type of `fun(3)` to be `Literal[True]` and the type of `fun("abc")` to be `Literal[False]`, but pyright says that both of them are `bool`
For testing purposes in the pandas stubs, we would use `assert_type(fun(3), Literal[True])`
**Describe the solution you'd like**
Narrowing of the type of the function to `True` or `False`, when possible
|
non_process
|
infer type of function that is a typeguard based on the argument is your feature request related to a problem please describe we d like to test if a function annotated as typeguard has a value of true when an object of some type is passed python from future import annotations from typing extensions import typeguard reveal type def fun x int str typeguard return isinstance x int reveal type fun reveal type fun abc i would have expected the type of fun to be literal and the type of fun abc to be literal but pyright says that both of them are bool for testing purposes in the pandas stubs we would use assert type fun literal describe the solution you d like narrowing of the type of the function to true or false when possible
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.