Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
86,557 | 10,510,658,017 | IssuesEvent | 2019-09-27 13:53:39 | lxc/lxd | https://api.github.com/repos/lxc/lxd | closed | Cluster member roles | API Documentation Feature | Right now every cluster member can either be a database node or not with LXD making this decision at the time of join or when promotion is required. To make things more flexible for added functionality moving forward, we should make this a bit more visible by introducing the concept of roles.
Each cluster member would then have the following roles available:
- database
- standby-database (mutually exclusive with `database`)
- event-hub
LXD would still automatically pick those roles on join or when promotion is required, but the API will allow the user to manually add/remove roles so long as it wouldn't break the cluster.
At any time, at least one database node and one event-hub would need to be present.
As right now we don't have `standby-database` nor `event-hub` implemented, the initial pass on this should only consist of:
- Create a `nodes_roles` table
- Record `database` role for all raft nodes
- Expose list of roles as config in `/1.0/cluster/members/NAME`
- Introduce the usual set of `edit/set/unset` commands for `lxc cluster`
- Make it possible to set/unset `role.database` on cluster members | 1.0 | Cluster member roles - Right now every cluster member can either be a database node or not with LXD making this decision at the time of join or when promotion is required. To make things more flexible for added functionality moving forward, we should make this a bit more visible by introducing the concept of roles.
Each cluster member would then have the following roles available:
- database
- standby-database (mutually exclusive with `database`)
- event-hub
LXD would still automatically pick those roles on join or when promotion is required, but the API will allow the user to manually add/remove roles so long as it wouldn't break the cluster.
At any time, at least one database node and one event-hub would need to be present.
As right now we don't have `standby-database` nor `event-hub` implemented, the initial pass on this should only consist of:
- Create a `nodes_roles` table
- Record `database` role for all raft nodes
- Expose list of roles as config in `/1.0/cluster/members/NAME`
- Introduce the usual set of `edit/set/unset` commands for `lxc cluster`
- Make it possible to set/unset `role.database` on cluster members | non_priority | cluster member roles right now every cluster member can either be a database node or not with lxd making this decision at the time of join or when promotion is required to make things more flexible for added functionality moving forward we should make this a bit more visible by introducing the concept of roles each cluster member would then have the following roles available database standby database mutually exclusive with database event hub lxd would still automatically pick those roles on join or when promotion is required but the api will allow the user to manually add remove roles so long as it wouldn t break the cluster at any time at least one database node and one event hub would need to be present as right now we don t have standby database nor event hub implemented the initial pass on this should only consist of create a nodes roles table record database role for all raft nodes expose list of roles as config in cluster members name introduce the usual set of edit set unset commands for lxc cluster make it possible to set unset role database on cluster members | 0 |
747,255 | 26,078,741,588 | IssuesEvent | 2022-12-25 01:02:13 | microsoft/PowerToys | https://api.github.com/repos/microsoft/PowerToys | closed | [Shortcut Guide]: CCA fails on place holder text at 'Exclude apps' edit field. | Issue-Bug Needs-Author-Feedback Product-Settings Priority-1 Area-Accessibility Status-No recent activity Resolution-Can't Repro A11ySev2 | ### Microsoft PowerToys version
0.45.0
### Running as admin
- [ ] Yes
### Area(s) with issue?
Shortcut Guide
FancyZones
Keyboard Manager
### Steps to reproduce
**Steps to reproduce**
Test Environment:
OS: Windows 11 Version Dev
App version: 0.45.0
App name: PowerToys
Screen Reader: Narrator
Tool: Colour Contrast Analyzer
**Repro Steps:**
1. Launch 'Powertoys' application.
2. Navigate to 'Shortcut Guide'.
3. Turn on Colour Contrast Analyzer(CCA).
4. Navigate to 'Exclude Apps' Edit field under 'Exclude apps' section.
5. Click a snip of the 'Exclude apps' editfield when on focus.
5. Observe the Contrast Ratio of the placeholder text when field is empty.
**Note:**
Same issue repro with 'Exclude apps' under FancyZones.
Same issue repro with 'All Apps' under Remap A shortcut.
**User Impact:** ->
Visual user are impacted here as information is conveyed by it with help of Colors.
**Guidelines Reference:-**
https://www.w3.org/TR/WCAG21/#non-text-contrast
**Attachments:**



### ✔️ Expected Behavior
Colour Contrast Ratio should be 4.5:1 on place holder text at 'Exclude apps' edit field.
### ❌ Actual Behavior
Colour Contrast Ratio is 3.3:1 > 4.5>1 on place holder text at 'Exclude apps' edit field.
### Other Software
_No response_ | 1.0 | [Shortcut Guide]: CCA fails on place holder text at 'Exclude apps' edit field. - ### Microsoft PowerToys version
0.45.0
### Running as admin
- [ ] Yes
### Area(s) with issue?
Shortcut Guide
FancyZones
Keyboard Manager
### Steps to reproduce
**Steps to reproduce**
Test Environment:
OS: Windows 11 Version Dev
App version: 0.45.0
App name: PowerToys
Screen Reader: Narrator
Tool: Colour Contrast Analyzer
**Repro Steps:**
1. Launch 'Powertoys' application.
2. Navigate to 'Shortcut Guide'.
3. Turn on Colour Contrast Analyzer(CCA).
4. Navigate to 'Exclude Apps' Edit field under 'Exclude apps' section.
5. Click a snip of the 'Exclude apps' editfield when on focus.
5. Observe the Contrast Ratio of the placeholder text when field is empty.
**Note:**
Same issue repro with 'Exclude apps' under FancyZones.
Same issue repro with 'All Apps' under Remap A shortcut.
**User Impact:** ->
Visual user are impacted here as information is conveyed by it with help of Colors.
**Guidelines Reference:-**
https://www.w3.org/TR/WCAG21/#non-text-contrast
**Attachments:**



### ✔️ Expected Behavior
Colour Contrast Ratio should be 4.5:1 on place holder text at 'Exclude apps' edit field.
### ❌ Actual Behavior
Colour Contrast Ratio is 3.3:1 > 4.5>1 on place holder text at 'Exclude apps' edit field.
### Other Software
_No response_ | priority | cca fails on place holder text at exclude apps edit field microsoft powertoys version running as admin yes area s with issue shortcut guide fancyzones keyboard manager steps to reproduce steps to reproduce test environment os windows version dev app version app name powertoys screen reader narrator tool colour contrast analyzer repro steps launch powertoys application navigate to shortcut guide turn on colour contrast analyzer cca navigate to exclude apps edit field under exclude apps section click a snip of the exclude apps editfield when on focus observe the contrast ratio of the placeholder text when field is empty note same issue repro with exclude apps under fancyzones same issue repro with all apps under remap a shortcut user impact visual user are impacted here as information is conveyed by it with help of colors guidelines reference attachments ✔️ expected behavior colour contrast ratio should be on place holder text at exclude apps edit field ❌ actual behavior colour contrast ratio is on place holder text at exclude apps edit field other software no response | 1 |
813,184 | 30,448,047,032 | IssuesEvent | 2023-07-15 23:45:37 | KingSupernova31/RulesGuru | https://api.github.com/repos/KingSupernova31/RulesGuru | opened | Restructure the editor layout | enhancement medium priority good first issue | It's annoying that you have to constantly scroll up and down on the editor. There's lots of wasted space at the top, so we should collapse the central nevigation column and move the relevant fields upwards. | 1.0 | Restructure the editor layout - It's annoying that you have to constantly scroll up and down on the editor. There's lots of wasted space at the top, so we should collapse the central nevigation column and move the relevant fields upwards. | priority | restructure the editor layout it s annoying that you have to constantly scroll up and down on the editor there s lots of wasted space at the top so we should collapse the central nevigation column and move the relevant fields upwards | 1 |
188,212 | 6,774,122,938 | IssuesEvent | 2017-10-27 09:11:32 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | [Federation] Implement federated pod autoscaler | area/federation priority/important-longterm sig/autoscaling sig/multicluster | This issue is created to track the development of a federated pod autoscaler.
The design document was updated [here ](https://docs.google.com/document/d/1Uacp99LskNHkJ6ZFFOVGorDivYKi4ySkI8E3tTFAZJ8/edit?usp=sharing) some days ago.
cc @kubernetes/sig-cluster-federation
@shashidharatd @kshafiee @deepak-vij | 1.0 | [Federation] Implement federated pod autoscaler - This issue is created to track the development of a federated pod autoscaler.
The design document was updated [here ](https://docs.google.com/document/d/1Uacp99LskNHkJ6ZFFOVGorDivYKi4ySkI8E3tTFAZJ8/edit?usp=sharing) some days ago.
cc @kubernetes/sig-cluster-federation
@shashidharatd @kshafiee @deepak-vij | priority | implement federated pod autoscaler this issue is created to track the development of a federated pod autoscaler the design document was updated some days ago cc kubernetes sig cluster federation shashidharatd kshafiee deepak vij | 1 |
141,277 | 18,957,038,994 | IssuesEvent | 2021-11-18 21:38:26 | CliffCrerar/bootstrap-5-starter | https://api.github.com/repos/CliffCrerar/bootstrap-5-starter | opened | CVE-2018-19838 (Medium) detected in opennmsopennms-source-26.0.0-1 | security vulnerability | ## CVE-2018-19838 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-26.0.0-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bootstrap-5-starter/commit/28feb8d3319a9c4a693d527577bee824eb4e960f">28feb8d3319a9c4a693d527577bee824eb4e960f</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/node-sass/src/libsass/src/ast.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, functions inside ast.cpp for IMPLEMENT_AST_OPERATORS expansion allow attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, as demonstrated by recursive calls involving clone(), cloneChildren(), and copy().
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19838>CVE-2018-19838</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-12-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19838 (Medium) detected in opennmsopennms-source-26.0.0-1 - ## CVE-2018-19838 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-26.0.0-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bootstrap-5-starter/commit/28feb8d3319a9c4a693d527577bee824eb4e960f">28feb8d3319a9c4a693d527577bee824eb4e960f</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/node-sass/src/libsass/src/ast.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, functions inside ast.cpp for IMPLEMENT_AST_OPERATORS expansion allow attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, as demonstrated by recursive calls involving clone(), cloneChildren(), and copy().
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19838>CVE-2018-19838</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-12-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in opennmsopennms source cve medium severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href found in base branch main vulnerable source files node modules node sass src libsass src ast cpp vulnerability details in libsass prior to functions inside ast cpp for implement ast operators expansion allow attackers to cause a denial of service resulting from stack consumption via a crafted sass file as demonstrated by recursive calls involving clone clonechildren and copy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource | 0 |
313,857 | 9,576,578,767 | IssuesEvent | 2019-05-07 09:24:53 | opencaching/opencaching-pl | https://api.github.com/repos/opencaching/opencaching-pl | opened | Reveal the number of users which ignore the cache | Component Cache Priority Medium Type Enhancement | from OCPL forum:
> autor: Wesker » niedziela, 20 stycznia 2019, 21:38
>
> Czy autor swojej skrzynki widzi, że ktoś ignoruje jego skrzynkę? Jeśli nie to czy dałoby się taką opcję dołożyć (także w samej skrzynce obok znalezionych/nieznalezionych)?
> Po co? wydaje mi się, że niektórzy autorzy nie zdają sobie sprawy z tego ile osób ignoruje ich skrzynki. Dla przykładu: skrzynki typu own. Niby są dostępne bo autorzy tak twierdzą, ale nie sposób się z nimi skontaktować i wtedy pozostaje dać do ignorowanych bo to chyba jedyny skuteczny sposób aby dany kesz nie męczył na mapie. Ilość ignorowanych to byłby także sygnał dla innych poszukujących. Skrzynki o których wspominam powyżej były raz przez COGowiczów dodane jako nieaktywne, ale szybko zostały przywrócone jako dostępne przez samych autorów. Zaraz po tym zabiegu próbowałem się z tymi osobami kontaktować ale cisza i 0 odpowiedzi. Tych osób nawet nie widać w zabawie od dłuższego czasu. Celowo nie piszę o jakie skrzynki chodzi, ale są.
My opinion is that the ignoring of the cache is a private decision of every user and should be a secret, but the number of ignoring users can be public. | 1.0 | Reveal the number of users which ignore the cache - from OCPL forum:
> autor: Wesker » niedziela, 20 stycznia 2019, 21:38
>
> Czy autor swojej skrzynki widzi, że ktoś ignoruje jego skrzynkę? Jeśli nie to czy dałoby się taką opcję dołożyć (także w samej skrzynce obok znalezionych/nieznalezionych)?
> Po co? wydaje mi się, że niektórzy autorzy nie zdają sobie sprawy z tego ile osób ignoruje ich skrzynki. Dla przykładu: skrzynki typu own. Niby są dostępne bo autorzy tak twierdzą, ale nie sposób się z nimi skontaktować i wtedy pozostaje dać do ignorowanych bo to chyba jedyny skuteczny sposób aby dany kesz nie męczył na mapie. Ilość ignorowanych to byłby także sygnał dla innych poszukujących. Skrzynki o których wspominam powyżej były raz przez COGowiczów dodane jako nieaktywne, ale szybko zostały przywrócone jako dostępne przez samych autorów. Zaraz po tym zabiegu próbowałem się z tymi osobami kontaktować ale cisza i 0 odpowiedzi. Tych osób nawet nie widać w zabawie od dłuższego czasu. Celowo nie piszę o jakie skrzynki chodzi, ale są.
My opinion is that the ignoring of the cache is a private decision of every user and should be a secret, but the number of ignoring users can be public. | priority | reveal the number of users which ignore the cache from ocpl forum autor wesker » niedziela stycznia czy autor swojej skrzynki widzi że ktoś ignoruje jego skrzynkę jeśli nie to czy dałoby się taką opcję dołożyć także w samej skrzynce obok znalezionych nieznalezionych po co wydaje mi się że niektórzy autorzy nie zdają sobie sprawy z tego ile osób ignoruje ich skrzynki dla przykładu skrzynki typu own niby są dostępne bo autorzy tak twierdzą ale nie sposób się z nimi skontaktować i wtedy pozostaje dać do ignorowanych bo to chyba jedyny skuteczny sposób aby dany kesz nie męczył na mapie ilość ignorowanych to byłby także sygnał dla innych poszukujących skrzynki o których wspominam powyżej były raz przez cogowiczów dodane jako nieaktywne ale szybko zostały przywrócone jako dostępne przez samych autorów zaraz po tym zabiegu próbowałem się z tymi osobami kontaktować ale cisza i odpowiedzi tych osób nawet nie widać w zabawie od dłuższego czasu celowo nie piszę o jakie skrzynki chodzi ale są my opinion is that the ignoring of the cache is a private decision of every user and should be a secret but the number of ignoring users can be public | 1 |
64,837 | 6,924,448,870 | IssuesEvent | 2017-11-30 12:49:18 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | ClientSplitBrainTest.testClientListeners_InSplitBrain | Team: Client Type: Test-Failure | ```
java.lang.AssertionError: id 0 expected:<3> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at com.hazelcast.client.ClientSplitBrainTest$1.run(ClientSplitBrainTest.java:93)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:239)
at com.hazelcast.client.ClientSplitBrainTest.checkEventsEventually(ClientSplitBrainTest.java:90)
at com.hazelcast.client.ClientSplitBrainTest.testClientListeners_InSplitBrain(ClientSplitBrainTest.java:84)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.maintenance-OpenJDK6/com.hazelcast$hazelcast-client/83/testReport/junit/com.hazelcast.client/ClientSplitBrainTest/testClientListeners_InSplitBrain/
| 1.0 | ClientSplitBrainTest.testClientListeners_InSplitBrain - ```
java.lang.AssertionError: id 0 expected:<3> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at com.hazelcast.client.ClientSplitBrainTest$1.run(ClientSplitBrainTest.java:93)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:239)
at com.hazelcast.client.ClientSplitBrainTest.checkEventsEventually(ClientSplitBrainTest.java:90)
at com.hazelcast.client.ClientSplitBrainTest.testClientListeners_InSplitBrain(ClientSplitBrainTest.java:84)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.maintenance-OpenJDK6/com.hazelcast$hazelcast-client/83/testReport/junit/com.hazelcast.client/ClientSplitBrainTest/testClientListeners_InSplitBrain/
| non_priority | clientsplitbraintest testclientlisteners insplitbrain java lang assertionerror id expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at com hazelcast client clientsplitbraintest run clientsplitbraintest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast client clientsplitbraintest checkeventseventually clientsplitbraintest java at com hazelcast client clientsplitbraintest testclientlisteners insplitbrain clientsplitbraintest java | 0 |
509,344 | 14,728,821,745 | IssuesEvent | 2021-01-06 10:29:20 | bounswe/bounswe2020group3 | https://api.github.com/repos/bounswe/bounswe2020group3 | opened | Filter on search results | Frontend Priority: Medium Status: In Progress Type: Enhancement | * **Project: FRONTEND**
* **This is a: FEATURE REQUEST**
* **Description of the issue**
There is no filtering mechanism on the search results. Therefore, the users see also undesired results.
* **For feature requests: Expected functionality of the requested feature**
A menu includes filter related options should be added top of the results on the search page
* **Deadline for resolution:**
11.01.2021
| 1.0 | Filter on search results - * **Project: FRONTEND**
* **This is a: FEATURE REQUEST**
* **Description of the issue**
There is no filtering mechanism on the search results. Therefore, the users see also undesired results.
* **For feature requests: Expected functionality of the requested feature**
A menu includes filter related options should be added top of the results on the search page
* **Deadline for resolution:**
11.01.2021
| priority | filter on search results project frontend this is a feature request description of the issue there is no filtering mechanism on the search results therefore the users see also undesired results for feature requests expected functionality of the requested feature a menu includes filter related options should be added top of the results on the search page deadline for resolution | 1 |
65,251 | 14,708,543,435 | IssuesEvent | 2021-01-04 23:59:18 | mauriciolauffer/NodeGoat | https://api.github.com/repos/mauriciolauffer/NodeGoat | opened | CVE-2018-1107 (High) detected in is-my-json-valid-2.15.0.tgz | security vulnerability | ## CVE-2018-1107 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mauriciolauffer/NodeGoat/commit/7e4baae75a2e9024fd12261d8d99f91cc913a308">7e4baae75a2e9024fd12261d8d99f91cc913a308</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
is-my-json-valid before 1.4.2 and 2.17.2 is vulnerable to ReDoS. This library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1546357">https://bugzilla.redhat.com/show_bug.cgi?id=1546357</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 1.4.2,2.17.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-1107 (High) detected in is-my-json-valid-2.15.0.tgz - ## CVE-2018-1107 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mauriciolauffer/NodeGoat/commit/7e4baae75a2e9024fd12261d8d99f91cc913a308">7e4baae75a2e9024fd12261d8d99f91cc913a308</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
is-my-json-valid before 1.4.2 and 2.17.2 is vulnerable to ReDoS. This library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1546357">https://bugzilla.redhat.com/show_bug.cgi?id=1546357</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 1.4.2,2.17.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in is my json valid tgz cve high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details is my json valid before and is vulnerable to redos this library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
17,550 | 10,728,618,785 | IssuesEvent | 2019-10-28 14:13:53 | cityofaustin/atd-knack-finance-purchasing | https://api.github.com/repos/cityofaustin/atd-knack-finance-purchasing | closed | Log-In Issues: D. Smith | Service: Apps Type: IT Support Workgroup: PE | David Smith
Parking Meters
emailed me
>I was unable to access my account and cannot reset my password for the Purchasing website this morning. Can you please help.
I emailed
>When you try to reset your password what’s it doing? Also, which log in option are you trying to do? If you choose Non-COA Sign-In, I believe you can reset your password. I would try the white “Sign In” button first though. Either describe what’s happening of provide a screen shot of any error messages you’re getting.
He replied
>When I used the white login button it did not recognize my password and would not allow me to rest it.
I tried the Non-COA login, was sent an e-mail with a link that I used to reset my password. When I tried to login in with the new password it showed a username/password error and would not let me in.
Retried the white login in button and it still will not recognize my password.
>Not sure what happened. Use Chrome. Used on Friday with no issues. Normally password and login are saved.
| 1.0 | Log-In Issues: D. Smith - David Smith
Parking Meters
emailed me
>I was unable to access my account and cannot reset my password for the Purchasing website this morning. Can you please help.
I emailed
>When you try to reset your password what’s it doing? Also, which log in option are you trying to do? If you choose Non-COA Sign-In, I believe you can reset your password. I would try the white “Sign In” button first though. Either describe what’s happening of provide a screen shot of any error messages you’re getting.
He replied
>When I used the white login button it did not recognize my password and would not allow me to rest it.
I tried the Non-COA login, was sent an e-mail with a link that I used to reset my password. When I tried to login in with the new password it showed a username/password error and would not let me in.
Retried the white login in button and it still will not recognize my password.
>Not sure what happened. Use Chrome. Used on Friday with no issues. Normally password and login are saved.
| non_priority | log in issues d smith david smith parking meters emailed me i was unable to access my account and cannot reset my password for the purchasing website this morning can you please help i emailed when you try to reset your password what’s it doing also which log in option are you trying to do if you choose non coa sign in i believe you can reset your password i would try the white “sign in” button first though either describe what’s happening of provide a screen shot of any error messages you’re getting he replied when i used the white login button it did not recognize my password and would not allow me to rest it i tried the non coa login was sent an e mail with a link that i used to reset my password when i tried to login in with the new password it showed a username password error and would not let me in retried the white login in button and it still will not recognize my password not sure what happened use chrome used on friday with no issues normally password and login are saved | 0 |
318,517 | 27,309,815,997 | IssuesEvent | 2023-02-24 11:13:49 | dusk-network/rusk | https://api.github.com/repos/dusk-network/rusk | closed | node: Share common structs and functionalities in a distinct crate | mark:testnet area:rusk-node | #### Summary
It is advisable to consider encapsulating common data structures and functionalities in a distinct crate. The new crate should allow both Consensus library and Node library to use any global data structs (like Ledger Block and Ledger Transaction) together with any common traits definitions.
Following items to be moved to this crate:
- Block struct (together with both Header and Certificate structs)
- Ledger Transaction (a wrapper of Phoenix tx)
- Serializable trait
- Message trait
- Message definitions
- Message types (topics)
NB: Wire message-related structs should be addressed together with Kadcast integration.
#### Additional context
> Incorporating the new crate into the consensus and node crates should be done simultaneously with implementing this issue.
| 1.0 | node: Share common structs and functionalities in a distinct crate - #### Summary
It is advisable to consider encapsulating common data structures and functionalities in a distinct crate. The new crate should allow both Consensus library and Node library to use any global data structs (like Ledger Block and Ledger Transaction) together with any common traits definitions.
Following items to be moved to this crate:
- Block struct (together with both Header and Certificate structs)
- Ledger Transaction (a wrapper of Phoenix tx)
- Serializable trait
- Message trait
- Message definitions
- Message types (topics)
NB: Wire message-related structs should be addressed together with Kadcast integration.
#### Additional context
> Incorporating the new crate into the consensus and node crates should be done simultaneously with implementing this issue.
| non_priority | node share common structs and functionalities in a distinct crate summary it is advisable to consider encapsulating common data structures and functionalities in a distinct crate the new crate should allow both consensus library and node library to use any global data structs like ledger block and ledger transaction together with any common traits definitions following items to be moved to this crate block struct together with both header and certificate structs ledger transaction a wrapper of phoenix tx serializable trait message trait message definitions message types topics nb wire message related structs should be addressed together with kadcast integration additional context incorporating the new crate into the consensus and node crates should be done simultaneously with implementing this issue | 0 |
63,801 | 26,518,551,557 | IssuesEvent | 2023-01-18 23:19:31 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | opened | Populate GitHub Repo with Cypress artefacts | ops and shared services Post Upgrade Testing | **Describe the issue**
Add Cypress code to Repo
**What is the Value/Impact?**
To be able to store and use the test scripts to be developed
**What is the plan? How will this get completed?**
We will be using the install procedure as laid out in https://github.com/bcgov/automated-testing
**Identify any dependencies**
N/A
**Definition of done**
Code is added and functional (smoke test can be run)
| 1.0 | Populate GitHub Repo with Cypress artefacts - **Describe the issue**
Add Cypress code to Repo
**What is the Value/Impact?**
To be able to store and use the test scripts to be developed
**What is the plan? How will this get completed?**
We will be using the install procedure as laid out in https://github.com/bcgov/automated-testing
**Identify any dependencies**
N/A
**Definition of done**
Code is added and functional (smoke test can be run)
| non_priority | populate github repo with cypress artefacts describe the issue add cypress code to repo what is the value impact to be able to store and use the test scripts to be developed what is the plan how will this get completed we will be using the install procedure as laid out in identify any dependencies n a definition of done code is added and functional smoke test can be run | 0 |
8,272 | 7,229,362,209 | IssuesEvent | 2018-02-11 19:06:14 | johnnylindbergh/OppBlock | https://api.github.com/repos/johnnylindbergh/OppBlock | opened | Logout | Security Vulnerability | The /logout endpoint calls the req.logout() function but doesn't actually kill the session. | True | Logout - The /logout endpoint calls the req.logout() function but doesn't actually kill the session. | non_priority | logout the logout endpoint calls the req logout function but doesn t actually kill the session | 0 |
661,861 | 22,092,603,581 | IssuesEvent | 2022-06-01 07:27:24 | magento/magento2 | https://api.github.com/repos/magento/magento2 | closed | Controller extending AccountInterface breaks in Production mode | Issue: Confirmed Triage: Dev.Experience Reproduced on 2.4.x Progress: PR in progress Priority: P2 Issue: needs update Area: Framework | <!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.4.0) and any important information on the environment where bug is reproducible.
-->
1. Magento 2.4.2 / 2.4.3-p1
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. Create a new module, route and controller which implements `AccountInterface` , as customer modules do
2. Implement the controller like this:
```
class Someaction implements AccountInterface, HttpGetActionInterface
```
3. Enable production mode
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. User is redirected to the customer login page
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
1. Here is the error when you try to access the frontend route
```
{"0":"Invalid return type","1":"#1 Magento\\Framework\\App\\Bootstrap->run() called at
[crypto\/pub\/index.php:42]\n","url":"\/mycontroller\/someaction\/","script_name":"\/index.php",
"report_id":"64317f2f4266fb92d176ead06a60faf2c27ba54790374689a1d4954d1d7cc5ed"}
```
### Solution
1. You need your controller to extend `DEPRECATED` `\Magento\Framework\App\Action\Action` class
### Important
1. This works in developer mode!
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [x] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
| 1.0 | Controller extending AccountInterface breaks in Production mode - <!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.4.0) and any important information on the environment where bug is reproducible.
-->
1. Magento 2.4.2 / 2.4.3-p1
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. Create a new module, route and controller which implements `AccountInterface` , as customer modules do
2. Implement the controller like this:
```
class Someaction implements AccountInterface, HttpGetActionInterface
```
3. Enable production mode
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. User is redirected to the customer login page
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
1. Here is the error when you try to access the frontend route
```
{"0":"Invalid return type","1":"#1 Magento\\Framework\\App\\Bootstrap->run() called at
[crypto\/pub\/index.php:42]\n","url":"\/mycontroller\/someaction\/","script_name":"\/index.php",
"report_id":"64317f2f4266fb92d176ead06a60faf2c27ba54790374689a1d4954d1d7cc5ed"}
```
### Solution
1. You need your controller to extend `DEPRECATED` `\Magento\Framework\App\Action\Action` class
### Important
1. This works in developer mode!
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [x] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
| priority | controller extending accountinterface breaks in production mode please review our guidelines before adding a new issue fields marked with are required please don t remove the template preconditions provide the exact magento version example and any important information on the environment where bug is reproducible magento steps to reproduce important provide a set of clear steps to reproduce this bug we can not provide support without clear instructions on how to reproduce create a new module route and controller which implements accountinterface as customer modules do implement the controller like this class someaction implements accountinterface httpgetactioninterface enable production mode expected result user is redirected to the customer login page actual result here is the error when you try to access the frontend route invalid return type magento framework app bootstrap run called at n url mycontroller someaction script name index php report id solution you need your controller to extend deprecated magento framework app action action class important this works in developer mode please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability” | 1 |
21,834 | 3,757,733,288 | IssuesEvent | 2016-03-14 03:07:58 | konato-events/web | https://api.github.com/repos/konato-events/web | opened | Fix header issues on mobile screens | design sev:critical type:bug | - menu toggle is glued to the right
- menu buttons are not side-by-side
- header animation is waaaay too slow and laggy (remove it?)
- header shrink stay half hidden | 1.0 | Fix header issues on mobile screens - - menu toggle is glued to the right
- menu buttons are not side-by-side
- header animation is waaaay too slow and laggy (remove it?)
- header shrink stay half hidden | non_priority | fix header issues on mobile screens menu toggle is glued to the right menu buttons are not side by side header animation is waaaay too slow and laggy remove it header shrink stay half hidden | 0 |
68,681 | 7,107,440,212 | IssuesEvent | 2018-01-16 19:59:15 | oakmound/oak | https://api.github.com/repos/oakmound/oak | closed | Benchmarking Pass 1 | tests | For a first benchmarking pass, we should write benchmarks for all of the main loops of the engine:
- Logic Loop
- Event Resolution Loop
- Draw Loop
- Input Loop
... In doing this we should investigate alternative, simpler implementations (that may disable features, and be triggered via config settings). | 1.0 | Benchmarking Pass 1 - For a first benchmarking pass, we should write benchmarks for all of the main loops of the engine:
- Logic Loop
- Event Resolution Loop
- Draw Loop
- Input Loop
... In doing this we should investigate alternative, simpler implementations (that may disable features, and be triggered via config settings). | non_priority | benchmarking pass for a first benchmarking pass we should write benchmarks for all of the main loops of the engine logic loop event resolution loop draw loop input loop in doing this we should investigate alternative simpler implementations that may disable features and be triggered via config settings | 0 |
5,673 | 3,975,542,013 | IssuesEvent | 2016-05-05 06:12:39 | kolliSuman/issues | https://api.github.com/repos/kolliSuman/issues | closed | QA_Grammar_Back to experiments_p2 | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open | Defect Description :
In the "Grammar" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user.
Actual Result :
In the "Grammar" experiment,the back to experiments link is not present in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/virtual-english-iitg/blob/master/test-cases/integration_test-cases/Grammar/Grammar_33_Back%20to%20experiments_p2.org | True | QA_Grammar_Back to experiments_p2 - Defect Description :
In the "Grammar" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user.
Actual Result :
In the "Grammar" experiment,the back to experiments link is not present in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/virtual-english-iitg/blob/master/test-cases/integration_test-cases/Grammar/Grammar_33_Back%20to%20experiments_p2.org | non_priority | qa grammar back to experiments defect description in the grammar experiment the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in order to view the list of experiments by the user actual result in the grammar experiment the back to experiments link is not present in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link | 0 |
409,086 | 11,956,624,064 | IssuesEvent | 2020-04-04 11:23:26 | AY1920S2-CS2103T-W17-3/main | https://api.github.com/repos/AY1920S2-CS2103T-W17-3/main | closed | As an unthrifty traveller, I want to be notified if my spending goes beyond my planned levels | priority.High type.Story | ... so that I can adjust my budget | 1.0 | As an unthrifty traveller, I want to be notified if my spending goes beyond my planned levels - ... so that I can adjust my budget | priority | as an unthrifty traveller i want to be notified if my spending goes beyond my planned levels so that i can adjust my budget | 1 |
738,534 | 25,565,031,568 | IssuesEvent | 2022-11-30 13:43:24 | longhorn/longhorn | https://api.github.com/repos/longhorn/longhorn | closed | [DOC] Add information about encryption algorithm to documentation | kind/doc priority/0 feature/encryption | **Is your feature request related to a problem? Please describe.**
The documentation page for volume encryption contains no information about the encryption algorithms & modes used for encryption in transit or at rest. This information is often required by IT Security managers or compliance authorities.
**Describe the solution you'd like**
Add a summary about what encryption algorithms are used for both encryption in transit & encryption at rest.
Eg, "Encryption at rest uses AES256 bit in XYZ mode and encryption in transit uses AES256 bit in ABC mode."
**Describe alternatives you've considered**
I also looked in the longhorn/longhorn repo for this information but can't find what I'm looking for.
**Additional context**
The documentation page is https://longhorn.io/docs/1.2.2/advanced-resources/volume-encryption/.
Also, it's amazing that encryption was added, very much appreciated.
| 1.0 | [DOC] Add information about encryption algorithm to documentation - **Is your feature request related to a problem? Please describe.**
The documentation page for volume encryption contains no information about the encryption algorithms & modes used for encryption in transit or at rest. This information is often required by IT Security managers or compliance authorities.
**Describe the solution you'd like**
Add a summary about what encryption algorithms are used for both encryption in transit & encryption at rest.
Eg, "Encryption at rest uses AES256 bit in XYZ mode and encryption in transit uses AES256 bit in ABC mode."
**Describe alternatives you've considered**
I also looked in the longhorn/longhorn repo for this information but can't find what I'm looking for.
**Additional context**
The documentation page is https://longhorn.io/docs/1.2.2/advanced-resources/volume-encryption/.
Also, it's amazing that encryption was added, very much appreciated.
| priority | add information about encryption algorithm to documentation is your feature request related to a problem please describe the documentation page for volume encryption contains no information about the encryption algorithms modes used for encryption in transit or at rest this information is often required by it security managers or compliance authorities describe the solution you d like add a summary about what encryption algorithms are used for both encryption in transit encryption at rest eg encryption at rest uses bit in xyz mode and encryption in transit uses bit in abc mode describe alternatives you ve considered i also looked in the longhorn longhorn repo for this information but can t find what i m looking for additional context the documentation page is also it s amazing that encryption was added very much appreciated | 1 |
152,746 | 24,011,517,494 | IssuesEvent | 2022-09-14 19:16:41 | GCTC-NTGC/gc-digital-talent | https://api.github.com/repos/GCTC-NTGC/gc-digital-talent | opened | Admin - Tables missing visual null state | design | ## Description
When there are no results for in a table, there is nothing indicating that there are 0 results (null state).
## Screenshot
 | 1.0 | Admin - Tables missing visual null state - ## Description
When there are no results for in a table, there is nothing indicating that there are 0 results (null state).
## Screenshot
 | non_priority | admin tables missing visual null state description when there are no results for in a table there is nothing indicating that there are results null state screenshot | 0 |
680,516 | 23,274,297,472 | IssuesEvent | 2022-08-05 04:57:11 | prgrms-web-devcourse/Team-Books-CheckMoi-FE | https://api.github.com/repos/prgrms-web-devcourse/Team-Books-CheckMoi-FE | opened | Feat/userProfile/[id] | 프론트 Type: 기능추가 Priority: 중간 | ## 기능 요청
### 📌 설명
userProfile 페이지에 userid를 전달받아 특정 유저의 프로필을 띄울 수 있도록 한다.
### 🎨 구현할 내용
userid 받아오고 유저정보 보여주는 작업
### 예상 구현 시간
1일
### 시급한 정도
🏃🏻 보통 | 1.0 | Feat/userProfile/[id] - ## 기능 요청
### 📌 설명
userProfile 페이지에 userid를 전달받아 특정 유저의 프로필을 띄울 수 있도록 한다.
### 🎨 구현할 내용
userid 받아오고 유저정보 보여주는 작업
### 예상 구현 시간
1일
### 시급한 정도
🏃🏻 보통 | priority | feat userprofile 기능 요청 📌 설명 userprofile 페이지에 userid를 전달받아 특정 유저의 프로필을 띄울 수 있도록 한다 🎨 구현할 내용 userid 받아오고 유저정보 보여주는 작업 예상 구현 시간 시급한 정도 🏃🏻 보통 | 1 |
249,194 | 21,111,220,987 | IssuesEvent | 2022-04-05 01:58:28 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | lvm2 v2.03.09-2 (r4.1) | r4.1-dom0-cur-test r4.1-dom0-sec-test | Update of lvm2 to v2.03.09-2 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-lvm2/commit/9716b12b2905eb28aa2ae0bfe92e82926ba4bb88
[Changes since previous version](https://github.com/QubesOS/qubes-lvm2/compare/72c190c108153ae8503111e196a34fe2580b8c53...v2.03.09-2):
QubesOS/qubes-lvm2@9716b12 version 2.03.09-2
QubesOS/qubes-lvm2@2a17e61 Exclude most devices from scanning
QubesOS/qubes-lvm2@c5a7b9f Fix handling built-in defaults for device filters
QubesOS/qubes-lvm2@2cd7e02 ci: add config
QubesOS/qubes-lvm2@d79c246 Set python3 explicitly
QubesOS/qubes-lvm2@fbf400e Add qubes-builder integration
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 current repo` (available 7 days from now)
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| 2.0 | lvm2 v2.03.09-2 (r4.1) - Update of lvm2 to v2.03.09-2 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-lvm2/commit/9716b12b2905eb28aa2ae0bfe92e82926ba4bb88
[Changes since previous version](https://github.com/QubesOS/qubes-lvm2/compare/72c190c108153ae8503111e196a34fe2580b8c53...v2.03.09-2):
QubesOS/qubes-lvm2@9716b12 version 2.03.09-2
QubesOS/qubes-lvm2@2a17e61 Exclude most devices from scanning
QubesOS/qubes-lvm2@c5a7b9f Fix handling built-in defaults for device filters
QubesOS/qubes-lvm2@2cd7e02 ci: add config
QubesOS/qubes-lvm2@d79c246 Set python3 explicitly
QubesOS/qubes-lvm2@fbf400e Add qubes-builder integration
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 current repo` (available 7 days from now)
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload lvm2 9716b12b2905eb28aa2ae0bfe92e82926ba4bb88 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| non_priority | update of to for qubes see comments below for details built from qubesos qubes version qubesos qubes exclude most devices from scanning qubesos qubes fix handling built in defaults for device filters qubesos qubes ci add config qubesos qubes set explicitly qubesos qubes add qubes builder integration referenced issues if you re release manager you can issue gpg inline signed command upload current repo available days from now upload current dists repo you can choose subset of distributions like vm vm available days from now upload security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it | 0 |
170,465 | 20,883,725,895 | IssuesEvent | 2022-03-23 01:06:00 | snowdensb/dependabot-core | https://api.github.com/repos/snowdensb/dependabot-core | reopened | CVE-2021-32804 (High) detected in multiple libraries | security vulnerability | ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b>, <b>tar-4.4.8.tgz</b>, <b>tar-2.2.1.tgz</b>, <b>tar-6.1.0.tgz</b>, <b>tar-4.4.6.tgz</b></p></summary>
<p>
<details><summary><b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/helpers/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/helpers/node_modules/npm/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- npm-6.14.13.tgz (Root Library)
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/tar</p>
<p>
Dependency Hierarchy:
- npm-5.8.0.tgz (Root Library)
- :x: **tar-4.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.8.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/node_modules/pacote/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/lerna/node_modules/pacote/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.6.0.tgz (Root Library)
- libnpm-2.0.1.tgz
- pacote-9.2.3.tgz
- :x: **tar-4.4.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/tar,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/tar,/npm_and_yarn/spec/fixtures/projects/npm7/lerna/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/tar</p>
<p>
Dependency Hierarchy:
- jest-cli-22.0.4.tgz (Root Library)
- jest-haste-map-22.0.3.tgz
- sane-2.2.0.tgz
- fsevents-1.1.3.tgz
- node-pre-gyp-0.6.39.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-6.1.0.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/helpers/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/helpers/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- arborist-2.7.1.tgz (Root Library)
- :x: **tar-6.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.6.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.6.tgz">https://registry.npmjs.org/tar/-/tar-4.4.6.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/node_modules/tar</p>
<p>
Dependency Hierarchy:
- bull-arena-2.4.5.tgz (Root Library)
- nodemon-1.18.4.tgz
- chokidar-2.0.4.tgz
- fsevents-1.2.4.tgz
- node-pre-gyp-0.10.3.tgz
- :x: **tar-4.4.6.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (npm): 6.14.14</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (npm): 5.9.0-next.0</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (lerna): 3.7.0</p><p>Fix Resolution (tar): 3.2.2</p>
<p>Direct dependency fix Resolution (jest-cli): 22.0.5</p><p>Fix Resolution (tar): 6.1.1</p>
<p>Direct dependency fix Resolution (@npmcli/arborist): 2.8.0</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (bull-arena): 2.5.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"6.14.13","packageFilePaths":["/npm_and_yarn/helpers/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:6.14.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.14.14","isBinary":false},{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"5.8.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:5.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.9.0-next.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lerna","packageVersion":"3.6.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/package.json"],"isTransitiveDependency":false,"dependencyTree":"lerna:3.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.7.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"jest-cli","packageVersion":"22.0.4","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/package.json"],"isTransitiveDependency":false,"dependencyTree":"jest-cli:22.0.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"22.0.5","isBinary":false},{"packageType":"javascript/Node.js","packageName":"@npmcli/arborist","packageVersion":"2.7.1","packageFilePaths":["/npm_and_yarn/helpers/package.json"],"isTransitiveDependency":false,"dependencyTree":"@npmcli/arborist:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"bull-arena","packageVersion":"2.4.5","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/package.json"],"isTransitiveDependency":false,"dependencyTree":"bull-arena:2.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.5.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-32804 (High) detected in multiple libraries - ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b>, <b>tar-4.4.8.tgz</b>, <b>tar-2.2.1.tgz</b>, <b>tar-6.1.0.tgz</b>, <b>tar-4.4.6.tgz</b></p></summary>
<p>
<details><summary><b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/helpers/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/helpers/node_modules/npm/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- npm-6.14.13.tgz (Root Library)
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/tar</p>
<p>
Dependency Hierarchy:
- npm-5.8.0.tgz (Root Library)
- :x: **tar-4.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.8.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/node_modules/pacote/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/lerna/node_modules/pacote/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.6.0.tgz (Root Library)
- libnpm-2.0.1.tgz
- pacote-9.2.3.tgz
- :x: **tar-4.4.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/tar,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/tar,/npm_and_yarn/spec/fixtures/projects/npm7/lerna/node_modules/tar/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/tar</p>
<p>
Dependency Hierarchy:
- jest-cli-22.0.4.tgz (Root Library)
- jest-haste-map-22.0.3.tgz
- sane-2.2.0.tgz
- fsevents-1.1.3.tgz
- node-pre-gyp-0.6.39.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-6.1.0.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/helpers/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/helpers/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- arborist-2.7.1.tgz (Root Library)
- :x: **tar-6.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.6.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.6.tgz">https://registry.npmjs.org/tar/-/tar-4.4.6.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/node_modules/tar</p>
<p>
Dependency Hierarchy:
- bull-arena-2.4.5.tgz (Root Library)
- nodemon-1.18.4.tgz
- chokidar-2.0.4.tgz
- fsevents-1.2.4.tgz
- node-pre-gyp-0.10.3.tgz
- :x: **tar-4.4.6.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (npm): 6.14.14</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (npm): 5.9.0-next.0</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (lerna): 3.7.0</p><p>Fix Resolution (tar): 3.2.2</p>
<p>Direct dependency fix Resolution (jest-cli): 22.0.5</p><p>Fix Resolution (tar): 6.1.1</p>
<p>Direct dependency fix Resolution (@npmcli/arborist): 2.8.0</p><p>Fix Resolution (tar): 4.4.14</p>
<p>Direct dependency fix Resolution (bull-arena): 2.5.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"6.14.13","packageFilePaths":["/npm_and_yarn/helpers/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:6.14.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.14.14","isBinary":false},{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"5.8.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:5.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.9.0-next.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lerna","packageVersion":"3.6.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/lerna/package.json"],"isTransitiveDependency":false,"dependencyTree":"lerna:3.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.7.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"jest-cli","packageVersion":"22.0.4","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/package.json"],"isTransitiveDependency":false,"dependencyTree":"jest-cli:22.0.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"22.0.5","isBinary":false},{"packageType":"javascript/Node.js","packageName":"@npmcli/arborist","packageVersion":"2.7.1","packageFilePaths":["/npm_and_yarn/helpers/package.json"],"isTransitiveDependency":false,"dependencyTree":"@npmcli/arborist:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"bull-arena","packageVersion":"2.4.5","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/github_dependency_slash/package.json"],"isTransitiveDependency":false,"dependencyTree":"bull-arena:2.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.5.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar tgz tar tgz tar tgz tar tgz tar tgz tar for node library home page a href path to dependency file npm and yarn helpers package json path to vulnerable library npm and yarn helpers node modules npm node modules tar package json dependency hierarchy npm tgz root library x tar tgz vulnerable library tar tgz tar for node library home page a href path to dependency file npm and yarn spec fixtures projects yarn dist tag package json path to vulnerable library npm and yarn spec fixtures projects yarn dist tag node modules tar dependency hierarchy npm tgz root library x tar tgz vulnerable library tar tgz tar for node library home page a href path to dependency file npm and yarn spec fixtures projects and yarn lerna package json path to vulnerable library npm and yarn spec fixtures projects and yarn lerna node modules pacote node modules tar package json npm and yarn spec fixtures projects lerna node modules pacote node modules tar package json dependency hierarchy lerna tgz root library libnpm tgz pacote tgz x tar tgz vulnerable library tar tgz tar for node library home page a href path to dependency file npm and yarn spec fixtures projects yarn no lockfile change package json path to vulnerable library npm and yarn spec fixtures projects yarn no lockfile change node modules tar npm and yarn spec fixtures projects and yarn lerna node modules tar package json npm and yarn spec fixtures projects yarn lockfile only change node modules tar npm and yarn spec fixtures projects lerna node modules tar package json npm and yarn spec fixtures projects yarn dist tag node modules tar dependency hierarchy jest cli tgz root library jest haste map tgz sane tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library tar tgz tar for node library home page a href path to dependency file npm and yarn helpers package json path to vulnerable library npm and yarn helpers node modules tar package json dependency hierarchy arborist tgz root library x tar tgz vulnerable library tar tgz tar for node library home page a href path to dependency file npm and yarn spec fixtures projects yarn github dependency slash package json path to vulnerable library npm and yarn spec fixtures projects yarn github dependency slash node modules tar dependency hierarchy bull arena tgz root library nodemon tgz chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch main vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution npm fix resolution tar direct dependency fix resolution npm next fix resolution tar direct dependency fix resolution lerna fix resolution tar direct dependency fix resolution jest cli fix resolution tar direct dependency fix resolution npmcli arborist fix resolution tar direct dependency fix resolution bull arena rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree npm isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename npm packageversion packagefilepaths istransitivedependency false dependencytree npm isminimumfixversionavailable true minimumfixversion next isbinary false packagetype javascript node js packagename lerna packageversion packagefilepaths istransitivedependency false dependencytree lerna isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename jest cli packageversion packagefilepaths istransitivedependency false dependencytree jest cli isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename npmcli arborist packageversion packagefilepaths istransitivedependency false dependencytree npmcli arborist isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename bull arena packageversion packagefilepaths istransitivedependency false dependencytree bull arena isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar vulnerabilityurl | 0 |
121,510 | 4,817,430,448 | IssuesEvent | 2016-11-04 13:41:19 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | closed | Ignore first rows while filtering and sorting | Plugin: column sorting Priority: normal Type: Feature | I have a custom dropdown for filtering, also need the ability to sort the data. Currently I am not able to do so since filtering and sorting applies to all the data including for headers. How can I implement a similar behavior while Ignoring the first row after header?
In the following example the first row has values of "Do not sort" which would be great to ignore for sorting and filtering.
http://jsfiddle.net/9rofomd5/
| 1.0 | Ignore first rows while filtering and sorting - I have a custom dropdown for filtering, also need the ability to sort the data. Currently I am not able to do so since filtering and sorting applies to all the data including for headers. How can I implement a similar behavior while Ignoring the first row after header?
In the following example the first row has values of "Do not sort" which would be great to ignore for sorting and filtering.
http://jsfiddle.net/9rofomd5/
| priority | ignore first rows while filtering and sorting i have a custom dropdown for filtering also need the ability to sort the data currently i am not able to do so since filtering and sorting applies to all the data including for headers how can i implement a similar behavior while ignoring the first row after header in the following example the first row has values of do not sort which would be great to ignore for sorting and filtering | 1 |
199,531 | 15,046,798,686 | IssuesEvent | 2021-02-03 07:59:46 | genenetwork/gn-deploy-servers | https://api.github.com/repos/genenetwork/gn-deploy-servers | closed | Move GN1 from EC2 to Lily (phase 1) | in progress testing | - [x] Make sure Lily is up-to-date (Pj)
- [x] Make an incremental offsite backup on S3 (Pj)
- [x] Point DNS http://gn1.genenework.org/ to Lily (Pj)
- [ ] Test GN1 and OPAR on Lily (Rob, Arthur, Zach)
when you all sign off this week, over the weekend:
- [ ] Disable syncronization script EC2 -> Lily so Lily becomes master (Pj)
- [ ] Point DNS http://www.genenetwork.org/ and OPAR to Lily (Pj)
(So Arthur, keep updating EC2 these days until I tell you to switch to
Lily).
- [ ] Halt EC2 image (but keep it for a possible restart) (Pj)
- [ ] Arthur can run update script for LODs (Arthur)
EC2 will remain for a few weeks, so we can easily switch back by updating the database and repointing DNS. But I think we'll be fine.
Note that we may get bitten by router configs in Memphis. Keep an eye on that and call Daniel if anything is weird.
| 1.0 | Move GN1 from EC2 to Lily (phase 1) - - [x] Make sure Lily is up-to-date (Pj)
- [x] Make an incremental offsite backup on S3 (Pj)
- [x] Point DNS http://gn1.genenework.org/ to Lily (Pj)
- [ ] Test GN1 and OPAR on Lily (Rob, Arthur, Zach)
when you all sign off this week, over the weekend:
- [ ] Disable syncronization script EC2 -> Lily so Lily becomes master (Pj)
- [ ] Point DNS http://www.genenetwork.org/ and OPAR to Lily (Pj)
(So Arthur, keep updating EC2 these days until I tell you to switch to
Lily).
- [ ] Halt EC2 image (but keep it for a possible restart) (Pj)
- [ ] Arthur can run update script for LODs (Arthur)
EC2 will remain for a few weeks, so we can easily switch back by updating the database and repointing DNS. But I think we'll be fine.
Note that we may get bitten by router configs in Memphis. Keep an eye on that and call Daniel if anything is weird.
| non_priority | move from to lily phase make sure lily is up to date pj make an incremental offsite backup on pj point dns to lily pj test and opar on lily rob arthur zach when you all sign off this week over the weekend disable syncronization script lily so lily becomes master pj point dns and opar to lily pj so arthur keep updating these days until i tell you to switch to lily halt image but keep it for a possible restart pj arthur can run update script for lods arthur will remain for a few weeks so we can easily switch back by updating the database and repointing dns but i think we ll be fine note that we may get bitten by router configs in memphis keep an eye on that and call daniel if anything is weird | 0 |
376,748 | 11,150,890,284 | IssuesEvent | 2019-12-24 01:15:42 | containous/traefik | https://api.github.com/repos/containous/traefik | closed | Memory usage keeps growing and continuously slow growing | area/provider/k8s area/provider/k8s/crd kind/bug/possible priority/P2 | ### Do you want to request a *feature* or report a *bug*?
Only my question
When I used traefik, I found the memory usage keeps growing, continuously slow growing

. So I use Jmeter for pressure test(traefik memeory is limited 2G),the result is:

And then I have stoped it, the memory also keep at 2G, it looks like strange.
### Output of `traefik version`: (*What version of Traefik are you using?*)
image: traefik:v2.1.0-rc2
### What did you do?
My deployment:
```yml
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
labels:
app: traefik-canary
name: traefik-canary
namespace: microsvc
spec:
progressDeadlineSeconds: 600
replicas: 1
revisionHistoryLimit: 10
selector:
matchLabels:
app: traefik-canary
strategy:
rollingUpdate:
maxSurge: 25%
maxUnavailable: 25%
type: RollingUpdate
template:
metadata:
creationTimestamp: null
labels:
app: traefik-canary
date: "1573641385"
name: traefik-canary
spec:
serviceAccountName: traefik-ingress-controller
containers:
- args:
- --configFile=/config/traefik-static.yaml
image: traefik:v2.1.0-rc2
imagePullPolicy: IfNotPresent
name: traefik
ports:
- containerPort: 80
name: api
protocol: TCP
- containerPort: 8080
name: admin
protocol: TCP
- containerPort: 8090
name: open
protocol: TCP
resources: {}
terminationMessagePath: /dev/termination-log
terminationMessagePolicy: File
resources:
limits:
cpu: "2"
memory: 2Gi
requests:
cpu: 100m
memory: 1Gi
volumeMounts:
- mountPath: /config
name: config-volume
dnsPolicy: ClusterFirst
restartPolicy: Always
schedulerName: default-scheduler
securityContext: {}
terminationGracePeriodSeconds: 30
volumes:
- configMap:
defaultMode: 420
name: traefik
name: config-volume
```
And my configMap:
```yml
apiVersion: v1
data:
traefik-static.yaml: |
entryPoints:
api:
address: ":80"
transport:
respondingTimeouts:
readTimeout: 15
writeTimeout: 15
open:
address: ":8090"
forwardedHeaders:
insecure: true
transport:
respondingTimeouts:
readTimeout: 15
writeTimeout: 15
api:
dashboard: true
insecure: true
providers:
file:
directory: /config
watch: true
kubernetesCRD: {}
tracing:
jaeger:
collector:
endpoint: http://jaeger-collector.arch.svc.cluster.local:14268/api/traces
samplingType: probabilistic
samplingParam: 0.3
accessLog:
filePath: "/log/traefik.log"
format: json
bufferingSize: 200
fields:
defaultMode: keep
headers:
defaultMode: keep
metrics:
prometheus: {}
ping:
entryPoint: "traefik"
health-check.yaml: |
http:
routers:
health-check:
rule: "Path(`/ping`)"
entryPoints:
- api
- open
service: health-check
services:
health-check:
loadBalancer:
servers:
- url: http://localhost:8080/ping
kind: ConfigMap
metadata:
name: traefik
namespace: microsvc
``` | 1.0 | Memory usage keeps growing and continuously slow growing - ### Do you want to request a *feature* or report a *bug*?
Only my question
When I used traefik, I found the memory usage keeps growing, continuously slow growing

. So I use Jmeter for pressure test(traefik memeory is limited 2G),the result is:

And then I have stoped it, the memory also keep at 2G, it looks like strange.
### Output of `traefik version`: (*What version of Traefik are you using?*)
image: traefik:v2.1.0-rc2
### What did you do?
My deployment:
```yml
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
labels:
app: traefik-canary
name: traefik-canary
namespace: microsvc
spec:
progressDeadlineSeconds: 600
replicas: 1
revisionHistoryLimit: 10
selector:
matchLabels:
app: traefik-canary
strategy:
rollingUpdate:
maxSurge: 25%
maxUnavailable: 25%
type: RollingUpdate
template:
metadata:
creationTimestamp: null
labels:
app: traefik-canary
date: "1573641385"
name: traefik-canary
spec:
serviceAccountName: traefik-ingress-controller
containers:
- args:
- --configFile=/config/traefik-static.yaml
image: traefik:v2.1.0-rc2
imagePullPolicy: IfNotPresent
name: traefik
ports:
- containerPort: 80
name: api
protocol: TCP
- containerPort: 8080
name: admin
protocol: TCP
- containerPort: 8090
name: open
protocol: TCP
resources: {}
terminationMessagePath: /dev/termination-log
terminationMessagePolicy: File
resources:
limits:
cpu: "2"
memory: 2Gi
requests:
cpu: 100m
memory: 1Gi
volumeMounts:
- mountPath: /config
name: config-volume
dnsPolicy: ClusterFirst
restartPolicy: Always
schedulerName: default-scheduler
securityContext: {}
terminationGracePeriodSeconds: 30
volumes:
- configMap:
defaultMode: 420
name: traefik
name: config-volume
```
And my configMap:
```yml
apiVersion: v1
data:
traefik-static.yaml: |
entryPoints:
api:
address: ":80"
transport:
respondingTimeouts:
readTimeout: 15
writeTimeout: 15
open:
address: ":8090"
forwardedHeaders:
insecure: true
transport:
respondingTimeouts:
readTimeout: 15
writeTimeout: 15
api:
dashboard: true
insecure: true
providers:
file:
directory: /config
watch: true
kubernetesCRD: {}
tracing:
jaeger:
collector:
endpoint: http://jaeger-collector.arch.svc.cluster.local:14268/api/traces
samplingType: probabilistic
samplingParam: 0.3
accessLog:
filePath: "/log/traefik.log"
format: json
bufferingSize: 200
fields:
defaultMode: keep
headers:
defaultMode: keep
metrics:
prometheus: {}
ping:
entryPoint: "traefik"
health-check.yaml: |
http:
routers:
health-check:
rule: "Path(`/ping`)"
entryPoints:
- api
- open
service: health-check
services:
health-check:
loadBalancer:
servers:
- url: http://localhost:8080/ping
kind: ConfigMap
metadata:
name: traefik
namespace: microsvc
``` | priority | memory usage keeps growing and continuously slow growing do you want to request a feature or report a bug only my question when i used traefik i found the memory usage keeps growing continuously slow growing so i use jmeter for pressure test(traefik memeory is limited ,the result is and then i have stoped it the memory also keep at it looks like strange output of traefik version what version of traefik are you using image traefik what did you do my deployment yml apiversion extensions kind deployment metadata labels app traefik canary name traefik canary namespace microsvc spec progressdeadlineseconds replicas revisionhistorylimit selector matchlabels app traefik canary strategy rollingupdate maxsurge maxunavailable type rollingupdate template metadata creationtimestamp null labels app traefik canary date name traefik canary spec serviceaccountname traefik ingress controller containers args configfile config traefik static yaml image traefik imagepullpolicy ifnotpresent name traefik ports containerport name api protocol tcp containerport name admin protocol tcp containerport name open protocol tcp resources terminationmessagepath dev termination log terminationmessagepolicy file resources limits cpu memory requests cpu memory volumemounts mountpath config name config volume dnspolicy clusterfirst restartpolicy always schedulername default scheduler securitycontext terminationgraceperiodseconds volumes configmap defaultmode name traefik name config volume and my configmap yml apiversion data traefik static yaml entrypoints api address transport respondingtimeouts readtimeout writetimeout open address forwardedheaders insecure true transport respondingtimeouts readtimeout writetimeout api dashboard true insecure true providers file directory config watch true kubernetescrd tracing jaeger collector endpoint samplingtype probabilistic samplingparam accesslog filepath log traefik log format json bufferingsize fields defaultmode keep headers defaultmode keep metrics prometheus ping entrypoint traefik health check yaml http routers health check rule path ping entrypoints api open service health check services health check loadbalancer servers url kind configmap metadata name traefik namespace microsvc | 1 |
78,612 | 7,655,847,587 | IssuesEvent | 2018-05-10 14:31:47 | PlasmaPy/PlasmaPy | https://api.github.com/repos/PlasmaPy/PlasmaPy | closed | Expand coverage for `PlasmaBlob` tests | Effort: trivial Good first contribution Priority: medium Testing | Most of the work here is just adding multiple plasma configurations to hit the various possible regimes for coupling and degeneracy. | 1.0 | Expand coverage for `PlasmaBlob` tests - Most of the work here is just adding multiple plasma configurations to hit the various possible regimes for coupling and degeneracy. | non_priority | expand coverage for plasmablob tests most of the work here is just adding multiple plasma configurations to hit the various possible regimes for coupling and degeneracy | 0 |
451,963 | 32,046,448,489 | IssuesEvent | 2023-09-23 03:37:39 | ISIS3510-202320-Team21/Wiki | https://api.github.com/repos/ISIS3510-202320-Team21/Wiki | closed | End 4 views and connect Figma diagram. | documentation | Generate five distinct views and establish connections between them and others within the Figma diagram. | 1.0 | End 4 views and connect Figma diagram. - Generate five distinct views and establish connections between them and others within the Figma diagram. | non_priority | end views and connect figma diagram generate five distinct views and establish connections between them and others within the figma diagram | 0 |
43,039 | 12,965,169,076 | IssuesEvent | 2020-07-20 21:49:18 | jtimberlake/griffin | https://api.github.com/repos/jtimberlake/griffin | opened | CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz, lodash-4.17.4.tgz | security vulnerability | ## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-4.17.4.tgz</b></p></summary>
<p>
<details><summary><b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/griffin/ui/angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/griffin/ui/angular/node_modules/karma/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-4.17.4.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/griffin/ui/angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/griffin/ui/angular/node_modules/angular-tree-component/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- angular-tree-component-4.1.0.tgz (Root Library)
- :x: **lodash-4.17.4.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/griffin/commit/7b8d4cb53c4eab239eecb18da5b2a6048b2fce60">7b8d4cb53c4eab239eecb18da5b2a6048b2fce60</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"karma:1.7.1;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.5"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.4","isTransitiveDependency":true,"dependencyTree":"angular-tree-component:4.1.0;lodash:4.17.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.5"}],"vulnerabilityIdentifier":"CVE-2018-3721","vulnerabilityDetails":"lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz, lodash-4.17.4.tgz - ## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-4.17.4.tgz</b></p></summary>
<p>
<details><summary><b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/griffin/ui/angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/griffin/ui/angular/node_modules/karma/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-4.17.4.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/griffin/ui/angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/griffin/ui/angular/node_modules/angular-tree-component/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- angular-tree-component-4.1.0.tgz (Root Library)
- :x: **lodash-4.17.4.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/griffin/commit/7b8d4cb53c4eab239eecb18da5b2a6048b2fce60">7b8d4cb53c4eab239eecb18da5b2a6048b2fce60</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"karma:1.7.1;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.5"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.4","isTransitiveDependency":true,"dependencyTree":"angular-tree-component:4.1.0;lodash:4.17.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.5"}],"vulnerabilityIdentifier":"CVE-2018-3721","vulnerabilityDetails":"lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in lodash tgz lodash tgz cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file tmp ws scm griffin ui angular package json path to vulnerable library tmp ws scm griffin ui angular node modules karma node modules lodash package json dependency hierarchy karma tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm griffin ui angular package json path to vulnerable library tmp ws scm griffin ui angular node modules angular tree component node modules lodash package json dependency hierarchy angular tree component tgz root library x lodash tgz vulnerable library found in head commit a href vulnerability details lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects vulnerabilityurl | 0 |
12,923 | 8,039,009,414 | IssuesEvent | 2018-07-30 17:00:33 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Drastic network io performance degradation in Docker | area-System.Net needs more info tenet-performance | Hi.
We've dockerized one of our apps recently and everything seemed to be fine except we just couldn't make it pass load test. Symptoms seem pretty strange to me, later about them. However after 'full desperado mode' took over we tried to run it on host machine with just dotnet binaries (no dotnet installed) and test passed with ease.
The app itself is simple fully async .NetCore 2.1.0 CRUD Web Api with network related work mostly - it communicates with mongodb, rabbitmq and a couple of WCF services and that's it.
So symptoms that are 100% reproducible under load:
any request takes enormous amount of time to complete, as simple as this controller is:
```
[HttpGet]
[Route("test2")]
public async Task<string> Test2()
{
return "Ok";
}
```
after hitting some load rate(~50% ) the time for completion grows linearly and shortly after takes >1m so 504 from Nginx. However our logging shows that processing of the whole request pipeline takes ~3ms, so we inspected what actually happens with tcpdump:
0.00 - packet with that Get received
~2.00 - processing started (according to app logs)
~2.00 - processing finished (according to app logs)
~4.00 - packet with 'Ok' sent.
(these timings are just after completion time starts to grow up)
So we thought this is Kestrel related and these 2s + 2s are wasted somewhere in there, but second test contoller came in:
```
private readonly HttpClient _client = new HttpClient();
[HttpGet]
public async Task<string> Test()
{
_logger.LogInformation("starting test async io");
var response = await _client.GetAsync("http://<some intranet url that just returns static ~40 bytes string in no time>");
var result = await response.Content.ReadAsStringAsync();
_logger.LogInformation("finished test async io");
return result;
}
```
and there's ~2s between first and second log entries(and it behaves same as in test 1 - delay before and after processing). Tcpdump shows that response from url comes almost immediately after request packet has been sent, as expected.
There's absolutely no ThreadPool starvation, as number of threads is ~50 and minThreads is set to (1024, 1024) anyways.
I believe there's no network issues neither with host or container itself.
What we tried:
1. We tried .netCore 2.0.0, .netCore 2.1.0, .netCore 2.1.0 with UseLibuv
2. Alpine, Debian and Ubuntu base images
3. --net=host
4. perfcollect, but couldnt figure out the bottleneck
Host os information:
```
cat /etc/lsb-release
DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=16.04
DISTRIB_CODENAME=xenial
DISTRIB_DESCRIPTION="Ubuntu 16.04.2 LTS"
```
```
lscpu
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 8
On-line CPU(s) list: 0-7
Thread(s) per core: 1
Core(s) per socket: 1
Socket(s): 8
NUMA node(s): 1
Vendor ID: GenuineIntel
CPU family: 6
Model: 79
Model name: Intel(R) Xeon(R) CPU E5-2697A v4 @ 2.60GHz
Stepping: 1
CPU MHz: 2593.993
BogoMIPS: 5187.98
Hypervisor vendor: VMware
Virtualization type: full
L1d cache: 32K
L1i cache: 32K
L2 cache: 256K
L3 cache: 40960K
NUMA node0 CPU(s): 0-7
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts nopl xtopology tsc_reliable nonstop_tsc eagerfpu pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 invpcid rtm rdseed adx smap xsaveopt arat
```
Docker info:
```
sudo docker info
Containers: xxx
Running: xxx
Paused: xxx
Stopped: xxx
Images: xxx
Server Version: 17.12.0-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 89623f28b87a6004d4b785663257362d1658a729
runc version: b2567b37d7b75eb4cf325b77297b140ea686ce8f
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.4.0-62-generic
Operating System: Ubuntu 16.04.2 LTS
OSType: linux
Architecture: x86_64
CPUs: 8
Total Memory: 7.796GiB
Name: xxxxxxxx
ID: 7IDX:SAWS:SUKX:MJW5:RNDJ:TVCR:PZ2S:V2GB:KWPF:ZEGU:MLPV:KU6C
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
HTTPS Proxy: xxxxxxxx
Registry: https://index.docker.io/v1/
Labels:
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
WARNING: No swap limit support
```
Dockerfile:
```
FROM microsoft/dotnet:2.1-sdk AS build-env
WORKDIR src
ENV TZ=Europe/Moscow
RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
COPY src ./
RUN dotnet restore xxx.sln -s https://api.nuget.org/v3/index.json --packages /nuget/packages
RUN dotnet build xxx.sln -c Release --no-restore
RUN dotnet publish xxx.Web -c Release -o ../published/Web --no-restore
FROM microsoft/dotnet:2.1-aspnetcore-runtime AS base
WORKDIR /app
ENV TZ=Europe/Moscow
RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
FROM base AS production
WORKDIR /app
COPY --from=build-env ./src/published/Web ./
RUN apt-get update -y && \
apt-get upgrade -y && \
apt-get dist-upgrade -y && \
apt-get install -y net-tools && \
apt-get install -y procps && \
apt-get install -y tcpdump
ENTRYPOINT ["dotnet", "xxx.Web.dll"]
```
I can provide any necessary debug information/traces/dumps
P.S. Again i'd like to point out that it's not just some 'user code' failing, everything is ok on the host machine, but **not** in docker. | True | Drastic network io performance degradation in Docker - Hi.
We've dockerized one of our apps recently and everything seemed to be fine except we just couldn't make it pass load test. Symptoms seem pretty strange to me, later about them. However after 'full desperado mode' took over we tried to run it on host machine with just dotnet binaries (no dotnet installed) and test passed with ease.
The app itself is simple fully async .NetCore 2.1.0 CRUD Web Api with network related work mostly - it communicates with mongodb, rabbitmq and a couple of WCF services and that's it.
So symptoms that are 100% reproducible under load:
any request takes enormous amount of time to complete, as simple as this controller is:
```
[HttpGet]
[Route("test2")]
public async Task<string> Test2()
{
return "Ok";
}
```
after hitting some load rate(~50% ) the time for completion grows linearly and shortly after takes >1m so 504 from Nginx. However our logging shows that processing of the whole request pipeline takes ~3ms, so we inspected what actually happens with tcpdump:
0.00 - packet with that Get received
~2.00 - processing started (according to app logs)
~2.00 - processing finished (according to app logs)
~4.00 - packet with 'Ok' sent.
(these timings are just after completion time starts to grow up)
So we thought this is Kestrel related and these 2s + 2s are wasted somewhere in there, but second test contoller came in:
```
private readonly HttpClient _client = new HttpClient();
[HttpGet]
public async Task<string> Test()
{
_logger.LogInformation("starting test async io");
var response = await _client.GetAsync("http://<some intranet url that just returns static ~40 bytes string in no time>");
var result = await response.Content.ReadAsStringAsync();
_logger.LogInformation("finished test async io");
return result;
}
```
and there's ~2s between first and second log entries(and it behaves same as in test 1 - delay before and after processing). Tcpdump shows that response from url comes almost immediately after request packet has been sent, as expected.
There's absolutely no ThreadPool starvation, as number of threads is ~50 and minThreads is set to (1024, 1024) anyways.
I believe there's no network issues neither with host or container itself.
What we tried:
1. We tried .netCore 2.0.0, .netCore 2.1.0, .netCore 2.1.0 with UseLibuv
2. Alpine, Debian and Ubuntu base images
3. --net=host
4. perfcollect, but couldnt figure out the bottleneck
Host os information:
```
cat /etc/lsb-release
DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=16.04
DISTRIB_CODENAME=xenial
DISTRIB_DESCRIPTION="Ubuntu 16.04.2 LTS"
```
```
lscpu
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 8
On-line CPU(s) list: 0-7
Thread(s) per core: 1
Core(s) per socket: 1
Socket(s): 8
NUMA node(s): 1
Vendor ID: GenuineIntel
CPU family: 6
Model: 79
Model name: Intel(R) Xeon(R) CPU E5-2697A v4 @ 2.60GHz
Stepping: 1
CPU MHz: 2593.993
BogoMIPS: 5187.98
Hypervisor vendor: VMware
Virtualization type: full
L1d cache: 32K
L1i cache: 32K
L2 cache: 256K
L3 cache: 40960K
NUMA node0 CPU(s): 0-7
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts nopl xtopology tsc_reliable nonstop_tsc eagerfpu pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 invpcid rtm rdseed adx smap xsaveopt arat
```
Docker info:
```
sudo docker info
Containers: xxx
Running: xxx
Paused: xxx
Stopped: xxx
Images: xxx
Server Version: 17.12.0-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 89623f28b87a6004d4b785663257362d1658a729
runc version: b2567b37d7b75eb4cf325b77297b140ea686ce8f
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.4.0-62-generic
Operating System: Ubuntu 16.04.2 LTS
OSType: linux
Architecture: x86_64
CPUs: 8
Total Memory: 7.796GiB
Name: xxxxxxxx
ID: 7IDX:SAWS:SUKX:MJW5:RNDJ:TVCR:PZ2S:V2GB:KWPF:ZEGU:MLPV:KU6C
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
HTTPS Proxy: xxxxxxxx
Registry: https://index.docker.io/v1/
Labels:
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
WARNING: No swap limit support
```
Dockerfile:
```
FROM microsoft/dotnet:2.1-sdk AS build-env
WORKDIR src
ENV TZ=Europe/Moscow
RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
COPY src ./
RUN dotnet restore xxx.sln -s https://api.nuget.org/v3/index.json --packages /nuget/packages
RUN dotnet build xxx.sln -c Release --no-restore
RUN dotnet publish xxx.Web -c Release -o ../published/Web --no-restore
FROM microsoft/dotnet:2.1-aspnetcore-runtime AS base
WORKDIR /app
ENV TZ=Europe/Moscow
RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
FROM base AS production
WORKDIR /app
COPY --from=build-env ./src/published/Web ./
RUN apt-get update -y && \
apt-get upgrade -y && \
apt-get dist-upgrade -y && \
apt-get install -y net-tools && \
apt-get install -y procps && \
apt-get install -y tcpdump
ENTRYPOINT ["dotnet", "xxx.Web.dll"]
```
I can provide any necessary debug information/traces/dumps
P.S. Again i'd like to point out that it's not just some 'user code' failing, everything is ok on the host machine, but **not** in docker. | non_priority | drastic network io performance degradation in docker hi we ve dockerized one of our apps recently and everything seemed to be fine except we just couldn t make it pass load test symptoms seem pretty strange to me later about them however after full desperado mode took over we tried to run it on host machine with just dotnet binaries no dotnet installed and test passed with ease the app itself is simple fully async netcore crud web api with network related work mostly it communicates with mongodb rabbitmq and a couple of wcf services and that s it so symptoms that are reproducible under load any request takes enormous amount of time to complete as simple as this controller is public async task return ok after hitting some load rate the time for completion grows linearly and shortly after takes so from nginx however our logging shows that processing of the whole request pipeline takes so we inspected what actually happens with tcpdump packet with that get received processing started according to app logs processing finished according to app logs packet with ok sent these timings are just after completion time starts to grow up so we thought this is kestrel related and these are wasted somewhere in there but second test contoller came in private readonly httpclient client new httpclient public async task test logger loginformation starting test async io var response await client getasync intranet url that just returns static bytes string in no time var result await response content readasstringasync logger loginformation finished test async io return result and there s between first and second log entries and it behaves same as in test delay before and after processing tcpdump shows that response from url comes almost immediately after request packet has been sent as expected there s absolutely no threadpool starvation as number of threads is and minthreads is set to anyways i believe there s no network issues neither with host or container itself what we tried we tried netcore netcore netcore with uselibuv alpine debian and ubuntu base images net host perfcollect but couldnt figure out the bottleneck host os information cat etc lsb release distrib id ubuntu distrib release distrib codename xenial distrib description ubuntu lts lscpu architecture cpu op mode s bit bit byte order little endian cpu s on line cpu s list thread s per core core s per socket socket s numa node s vendor id genuineintel cpu family model model name intel r xeon r cpu stepping cpu mhz bogomips hypervisor vendor vmware virtualization type full cache cache cache cache numa cpu s flags fpu vme de pse tsc msr pae mce apic sep mtrr pge mca cmov pat clflush dts mmx fxsr sse ss syscall nx rdtscp lm constant tsc arch perfmon pebs bts nopl xtopology tsc reliable nonstop tsc eagerfpu pni pclmulqdq fma pcid movbe popcnt tsc deadline timer aes xsave avx rdrand hypervisor lahf lm abm fsgsbase tsc adjust hle smep invpcid rtm rdseed adx smap xsaveopt arat docker info sudo docker info containers xxx running xxx paused xxx stopped xxx images xxx server version ce storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host macvlan null overlay log awslogs fluentd gcplogs gelf journald json file logentries splunk syslog swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options apparmor seccomp profile default kernel version generic operating system ubuntu lts ostype linux architecture cpus total memory name xxxxxxxx id saws sukx rndj tvcr kwpf zegu mlpv docker root dir var lib docker debug mode client false debug mode server false https proxy xxxxxxxx registry labels experimental false insecure registries live restore enabled false warning no swap limit support dockerfile from microsoft dotnet sdk as build env workdir src env tz europe moscow run ln snf usr share zoneinfo tz etc localtime echo tz etc timezone copy src run dotnet restore xxx sln s packages nuget packages run dotnet build xxx sln c release no restore run dotnet publish xxx web c release o published web no restore from microsoft dotnet aspnetcore runtime as base workdir app env tz europe moscow run ln snf usr share zoneinfo tz etc localtime echo tz etc timezone from base as production workdir app copy from build env src published web run apt get update y apt get upgrade y apt get dist upgrade y apt get install y net tools apt get install y procps apt get install y tcpdump entrypoint i can provide any necessary debug information traces dumps p s again i d like to point out that it s not just some user code failing everything is ok on the host machine but not in docker | 0 |
412,084 | 12,035,073,053 | IssuesEvent | 2020-04-13 17:13:18 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | closed | CATS: Properties not working as expected. | Priority: Medium Type: Bug | **Describe the bug**
Setting following properties in `values.yaml` is expected run only `internetless` CATs suite, however in practice I have noticed it running `apps` suite in addition to `internetless`. This issue might not be specific to running any particular suite.
```
properties:
acceptance-tests:
acceptance-tests:
acceptance_tests:
include: '=internetless'
```
**To Reproduce**
Set above properties in your `values.yaml` and run CATs: `bazel run //testing/acceptance_tests`
**Expected behavior**
A clear and concise description of what you expected to happen.
**Environment**
- Component/Plugin Name: CATs
| 1.0 | CATS: Properties not working as expected. - **Describe the bug**
Setting following properties in `values.yaml` is expected run only `internetless` CATs suite, however in practice I have noticed it running `apps` suite in addition to `internetless`. This issue might not be specific to running any particular suite.
```
properties:
acceptance-tests:
acceptance-tests:
acceptance_tests:
include: '=internetless'
```
**To Reproduce**
Set above properties in your `values.yaml` and run CATs: `bazel run //testing/acceptance_tests`
**Expected behavior**
A clear and concise description of what you expected to happen.
**Environment**
- Component/Plugin Name: CATs
| priority | cats properties not working as expected describe the bug setting following properties in values yaml is expected run only internetless cats suite however in practice i have noticed it running apps suite in addition to internetless this issue might not be specific to running any particular suite properties acceptance tests acceptance tests acceptance tests include internetless to reproduce set above properties in your values yaml and run cats bazel run testing acceptance tests expected behavior a clear and concise description of what you expected to happen environment component plugin name cats | 1 |
84,001 | 3,645,746,186 | IssuesEvent | 2016-02-15 15:53:00 | gdg-x/frisbee | https://api.github.com/repos/gdg-x/frisbee | closed | Remove build timestamp from BuildConfig | enhancement good first issue in progress priority: high | Use paperwork instead of BuildConfig for gitSha and timestamp
Paperwork is available at
https://github.com/zsoltk/paperwork | 1.0 | Remove build timestamp from BuildConfig - Use paperwork instead of BuildConfig for gitSha and timestamp
Paperwork is available at
https://github.com/zsoltk/paperwork | priority | remove build timestamp from buildconfig use paperwork instead of buildconfig for gitsha and timestamp paperwork is available at | 1 |
320,072 | 27,419,986,027 | IssuesEvent | 2023-03-01 16:05:59 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix meta.test_reptile_step | Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499789861" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4305405432/jobs/7507808950" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4161702642/jobs/7199989710" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4161702642/jobs/7199989710" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>Not found</summary>
Not found
</details>
| 1.0 | Fix meta.test_reptile_step - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499789861" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4305405432/jobs/7507808950" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4161702642/jobs/7199989710" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4161702642/jobs/7199989710" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>Not found</summary>
Not found
</details>
| non_priority | fix meta test reptile step tensorflow img src torch img src numpy img src jax img src not found not found | 0 |
299,340 | 9,205,377,419 | IssuesEvent | 2019-03-08 10:24:06 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | Labels are not printed on Windows XP | Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **spclark -** (spclark -)
Original Redmine Issue: 737, https://issues.qgis.org/issues/737
Original Assignee: nobody -
---
When a sample shape file is symbolized with labels, the labels are not printed in print composer not when the map is exported out to png.
| 1.0 | Labels are not printed on Windows XP - ---
Author Name: **spclark -** (spclark -)
Original Redmine Issue: 737, https://issues.qgis.org/issues/737
Original Assignee: nobody -
---
When a sample shape file is symbolized with labels, the labels are not printed in print composer not when the map is exported out to png.
| priority | labels are not printed on windows xp author name spclark spclark original redmine issue original assignee nobody when a sample shape file is symbolized with labels the labels are not printed in print composer not when the map is exported out to png | 1 |
239,298 | 7,788,491,271 | IssuesEvent | 2018-06-07 05:03:30 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | No completions and suggestions inside the public and private blocks of object Definition | Priority/High Severity/Major Type/Bug bug component/LanguageServer | **Description:**
```
type TestObject object {
public {
string testString;
int hello;
i<cursor>
}
private {
int a = 12;
}
new (testString) {}
function testFunction (string input) returns string;
};
```
In the above code snippet, completions are not given at the cursor position
**Affected Versions:**
v0.973.0 | 1.0 | No completions and suggestions inside the public and private blocks of object Definition - **Description:**
```
type TestObject object {
public {
string testString;
int hello;
i<cursor>
}
private {
int a = 12;
}
new (testString) {}
function testFunction (string input) returns string;
};
```
In the above code snippet, completions are not given at the cursor position
**Affected Versions:**
v0.973.0 | priority | no completions and suggestions inside the public and private blocks of object definition description type testobject object public string teststring int hello i private int a new teststring function testfunction string input returns string in the above code snippet completions are not given at the cursor position affected versions | 1 |
686,158 | 23,479,017,086 | IssuesEvent | 2022-08-17 08:55:50 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.instagram.com - site is not usable | priority-critical browser-focus-geckoview engine-gecko | <!-- @browser: Firefox Mobile 103.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:103.0) Gecko/103.0 Firefox/103.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/109217 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.instagram.com/reel/CfVIk46jNhW/?igshid=MDJmNzVkMjY%3D
**Browser / Version**: Firefox Mobile 103.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Webpage loaded once initially after accepting required cookies only. When I refresh the page it doesn't load again. If i close and reopen Firefox Focus, it never loads the webpage. Only once.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/a885c849-d6e2-47d1-8cf4-295119ac8438.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220802163236</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/8/a6825a81-1587-4471-a8df-d029168ac640)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.instagram.com - site is not usable - <!-- @browser: Firefox Mobile 103.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:103.0) Gecko/103.0 Firefox/103.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/109217 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.instagram.com/reel/CfVIk46jNhW/?igshid=MDJmNzVkMjY%3D
**Browser / Version**: Firefox Mobile 103.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Webpage loaded once initially after accepting required cookies only. When I refresh the page it doesn't load again. If i close and reopen Firefox Focus, it never loads the webpage. Only once.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/a885c849-d6e2-47d1-8cf4-295119ac8438.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220802163236</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/8/a6825a81-1587-4471-a8df-d029168ac640)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description page not loading correctly steps to reproduce webpage loaded once initially after accepting required cookies only when i refresh the page it doesn t load again if i close and reopen firefox focus it never loads the webpage only once view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
3,571 | 9,807,650,322 | IssuesEvent | 2019-06-12 14:07:54 | trezor/trezor-firmware | https://api.github.com/repos/trezor/trezor-firmware | opened | touch-event-driven tests for TT | P4 Low S4 Low architecture core emulator feature firmware tests | Currently the "user input" for automated tests is driven by "confirm" signals from `DebugLinkDecision`. These need to be specially handled in TT application code, and this handling is inconsistent (e.g. it's impossible to select host/device in passphrase prompt).
Instead of (and in addition to) this mechanism, it should be possible to send touch events over the debug link. The special handling would only happen in the core loop, and we could simulate all kinds of touch events more accurately.
Note that this would not be usable for PIN tests, and cumbersome for entering mnemonics or passphrases.
This mechanism should also provide feedback, perhaps a reply-message over debuglink, to indicate that handling is done and screen is redrawn, so that we don't have to hard-code delays in tests. | 1.0 | touch-event-driven tests for TT - Currently the "user input" for automated tests is driven by "confirm" signals from `DebugLinkDecision`. These need to be specially handled in TT application code, and this handling is inconsistent (e.g. it's impossible to select host/device in passphrase prompt).
Instead of (and in addition to) this mechanism, it should be possible to send touch events over the debug link. The special handling would only happen in the core loop, and we could simulate all kinds of touch events more accurately.
Note that this would not be usable for PIN tests, and cumbersome for entering mnemonics or passphrases.
This mechanism should also provide feedback, perhaps a reply-message over debuglink, to indicate that handling is done and screen is redrawn, so that we don't have to hard-code delays in tests. | non_priority | touch event driven tests for tt currently the user input for automated tests is driven by confirm signals from debuglinkdecision these need to be specially handled in tt application code and this handling is inconsistent e g it s impossible to select host device in passphrase prompt instead of and in addition to this mechanism it should be possible to send touch events over the debug link the special handling would only happen in the core loop and we could simulate all kinds of touch events more accurately note that this would not be usable for pin tests and cumbersome for entering mnemonics or passphrases this mechanism should also provide feedback perhaps a reply message over debuglink to indicate that handling is done and screen is redrawn so that we don t have to hard code delays in tests | 0 |
39,616 | 10,367,524,340 | IssuesEvent | 2019-09-07 09:01:54 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | Ballerina build not working relative paths | Area/BuildTools Priority/High Type/Bug | **Description:**
When I give full path it is working.
ex.
```
test$ ballerina run /home/waruna/Downloads/main.bal
34
test$ ballerina run ../main.bal
error: no ballerina source files found in module 'main.bal'
test$ md5sum ../main.bal
5ee5b0538f2d3fb2c246f45a97029d6d ../main.bal
``` | 1.0 | Ballerina build not working relative paths - **Description:**
When I give full path it is working.
ex.
```
test$ ballerina run /home/waruna/Downloads/main.bal
34
test$ ballerina run ../main.bal
error: no ballerina source files found in module 'main.bal'
test$ md5sum ../main.bal
5ee5b0538f2d3fb2c246f45a97029d6d ../main.bal
``` | non_priority | ballerina build not working relative paths description when i give full path it is working ex test ballerina run home waruna downloads main bal test ballerina run main bal error no ballerina source files found in module main bal test main bal main bal | 0 |
635,407 | 20,387,022,895 | IssuesEvent | 2022-02-22 08:12:26 | enso-org/enso | https://api.github.com/repos/enso-org/enso | closed | up_to doesnt work correctly when range start is 0 | Type: Bug Status: Help Wanted Difficulty: Beginner Change: Non-Breaking Priority: High Size: Small Status: Good First Issue Category: Stdlib | <!--
Please ensure that you are running the latest version of Enso before reporting
the bug! It may have been fixed since.
-->
### General Summary
the up_to operator doesnt work correctly when range starts at 0.
### Steps to Reproduce
1. open the enso IDE
2. enter the graph below (this is the error case - range 0..6 returns a list [0..4] with a length of 5)

3. change the range start to 1 -( correct use - range 1..6 returns a list [1..5] with a length of 5)

| 1.0 | up_to doesnt work correctly when range start is 0 - <!--
Please ensure that you are running the latest version of Enso before reporting
the bug! It may have been fixed since.
-->
### General Summary
the up_to operator doesnt work correctly when range starts at 0.
### Steps to Reproduce
1. open the enso IDE
2. enter the graph below (this is the error case - range 0..6 returns a list [0..4] with a length of 5)

3. change the range start to 1 -( correct use - range 1..6 returns a list [1..5] with a length of 5)

| priority | up to doesnt work correctly when range start is please ensure that you are running the latest version of enso before reporting the bug it may have been fixed since general summary the up to operator doesnt work correctly when range starts at steps to reproduce open the enso ide enter the graph below this is the error case range returns a list with a length of change the range start to correct use range returns a list with a length of | 1 |
103,224 | 16,602,029,692 | IssuesEvent | 2021-06-01 20:56:41 | samq-ghdemo/SEARCH-NCJIS-nibrs | https://api.github.com/repos/samq-ghdemo/SEARCH-NCJIS-nibrs | opened | CVE-2018-19362 (High) detected in multiple libraries | security vulnerability | ## CVE-2018-19362 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19362","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-19362 (High) detected in multiple libraries - ## CVE-2018-19362 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19362","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs fbi service pom xml path to vulnerable library search ncjis nibrs tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs validation pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar search ncjis nibrs web nibrs web target nibrs web web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs flatfile pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the jboss common core class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the jboss common core class from polymorphic deserialization vulnerabilityurl | 0 |
65,655 | 6,971,574,107 | IssuesEvent | 2017-12-11 14:28:30 | Kademi/kademi-dev | https://api.github.com/repos/Kademi/kademi-dev | closed | Points allocation enhancements | enhancement High priority Ready to Test QA | Here's an example: http://viri.admin.kademi.com.au/sales/sales-by-dealer/sources/source-dealer-rewards/#general-tab

Please make the following enhancements
- Add an option to allow either allocate individual points records for each sale (as it does now) or to roll up sales into a single value and award a single points record for that. This option is for situations where individual sales are very small and the points records are rounded to zero, by adding them together the rounding will be avoided. This process should mark all included records as allocated.
- Add an optional expression field to "ignore" records. If this expression evaluates to true the record will not have any points allocated, and will not be marked as allocated, so it could be included in the next processing (if the expression output changes). Make this a textarea field to allow for more complex rules
- Change the existing expression field to a textarea to allow for more complex rules | 1.0 | Points allocation enhancements - Here's an example: http://viri.admin.kademi.com.au/sales/sales-by-dealer/sources/source-dealer-rewards/#general-tab

Please make the following enhancements
- Add an option to allow either allocate individual points records for each sale (as it does now) or to roll up sales into a single value and award a single points record for that. This option is for situations where individual sales are very small and the points records are rounded to zero, by adding them together the rounding will be avoided. This process should mark all included records as allocated.
- Add an optional expression field to "ignore" records. If this expression evaluates to true the record will not have any points allocated, and will not be marked as allocated, so it could be included in the next processing (if the expression output changes). Make this a textarea field to allow for more complex rules
- Change the existing expression field to a textarea to allow for more complex rules | non_priority | points allocation enhancements here s an example please make the following enhancements add an option to allow either allocate individual points records for each sale as it does now or to roll up sales into a single value and award a single points record for that this option is for situations where individual sales are very small and the points records are rounded to zero by adding them together the rounding will be avoided this process should mark all included records as allocated add an optional expression field to ignore records if this expression evaluates to true the record will not have any points allocated and will not be marked as allocated so it could be included in the next processing if the expression output changes make this a textarea field to allow for more complex rules change the existing expression field to a textarea to allow for more complex rules | 0 |
767,724 | 26,937,978,045 | IssuesEvent | 2023-02-07 22:30:31 | ZTL-ARTCC/Current_Website | https://api.github.com/repos/ZTL-ARTCC/Current_Website | closed | Update tower training flow | high priority | **Description**
Updates the flow of tower training sessions IAW training order v3.
**Acceptance Criteria**
Checkbox list of the acceptance criteria for this story
_Accepted By:_
**Dev Notes**
**Working Branch**
**PR**
**QA Notes**
**Deployment Notes**
| 1.0 | Update tower training flow - **Description**
Updates the flow of tower training sessions IAW training order v3.
**Acceptance Criteria**
Checkbox list of the acceptance criteria for this story
_Accepted By:_
**Dev Notes**
**Working Branch**
**PR**
**QA Notes**
**Deployment Notes**
| priority | update tower training flow description updates the flow of tower training sessions iaw training order acceptance criteria checkbox list of the acceptance criteria for this story accepted by dev notes working branch pr qa notes deployment notes | 1 |
510,710 | 14,814,842,965 | IssuesEvent | 2021-01-14 05:57:49 | threefoldtech/js-sdk | https://api.github.com/repos/threefoldtech/js-sdk | opened | 3Bot name entry chat screen: add description of name can only be lowercase letter. | priority_major type_bug | You only find out that you cannot use number, special characters and spaces after you have put on of them in the name and press ```NEXT```. Add in the description a guideline of what is accepted
"A 3bot name can only consist of lowercase letter, no special character, numbers or spaces allowed" | 1.0 | 3Bot name entry chat screen: add description of name can only be lowercase letter. - You only find out that you cannot use number, special characters and spaces after you have put on of them in the name and press ```NEXT```. Add in the description a guideline of what is accepted
"A 3bot name can only consist of lowercase letter, no special character, numbers or spaces allowed" | priority | name entry chat screen add description of name can only be lowercase letter you only find out that you cannot use number special characters and spaces after you have put on of them in the name and press next add in the description a guideline of what is accepted a name can only consist of lowercase letter no special character numbers or spaces allowed | 1 |
83,414 | 10,349,565,722 | IssuesEvent | 2019-09-04 23:00:22 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Brave feedback component | QA/Yes design/needs-mock-up priority/P4 | Brave needs a UI component to be able to solicit feedback on our various features (in particular if it is in beta) at the moment the user is using the feature. Since we are a privacy protecting browser and don't collect user data, we need to give the average user a way to express feedback so we can be more diligent about making well-informed updates. | 1.0 | Brave feedback component - Brave needs a UI component to be able to solicit feedback on our various features (in particular if it is in beta) at the moment the user is using the feature. Since we are a privacy protecting browser and don't collect user data, we need to give the average user a way to express feedback so we can be more diligent about making well-informed updates. | non_priority | brave feedback component brave needs a ui component to be able to solicit feedback on our various features in particular if it is in beta at the moment the user is using the feature since we are a privacy protecting browser and don t collect user data we need to give the average user a way to express feedback so we can be more diligent about making well informed updates | 0 |
726,573 | 25,003,863,987 | IssuesEvent | 2022-11-03 10:14:06 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | [bug] Conan v2 think that macOS SDK is an iOS SDK | stage: in-progress priority: medium complex: low bug | <!--
Please don't forget to update the issue title.
Include all applicable information to help us reproduce your problem.
To help us debug your issue please explain:
-->
### Environment Details (include every applicable attribute)
* Operating System+version: macos latest
* Compiler+version: clang 14
* Conan version: 1.53
* Python version: 3.9
### Steps to reproduce (Include if Applicable)
```
git clone https://github.com/nativium/nativium.git -b fix-conan-1-53
cd nativium
python3 nativium.py conan setup
python3 nativium.py target ios setup --arch=arm64 --build=debug --group=ios
python3 nativium.py target ios build --arch=arm64 --build=debug --group=ios
```
### Logs (Executed commands with output) (Include/Attach if Applicable)
```
CMake Error at /opt/homebrew/Cellar/cmake/3.24.2/share/cmake/Modules/Platform/iOS-Initialize.cmake:4 (message):
/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX12.3.sdk
is not an iOS SDK
Call Stack (most recent call first):
/opt/homebrew/Cellar/cmake/3.24.2/share/cmake/Modules/CMakeSystemSpecificInitialize.cmake:21 (include)
CMakeLists.txt:8 (project)
```
The problem is here:
https://github.com/nativium/nativium/blob/fix-conan-1-53/conan/darwin-toolchain/conanfile.py#L102
The `XCRun` object from conan don't check correctly the settings. It get the settings from conanfile, but need be something like i do, that is try to get `settings_target`. Or make XCRun receive `settings` and not `conanfile`:
```
settings_target = None
if hasattr(self, "settings_target"):
settings_target = self.settings_target
self.output.info("Settings: Target")
else:
settings_target = self.settings
self.output.info("Settings: Default")
```
The conan XCRun lines is here:
https://github.com/conan-io/conan/blob/develop/conan/tools/apple/apple.py#L82-L87 | 1.0 | [bug] Conan v2 think that macOS SDK is an iOS SDK - <!--
Please don't forget to update the issue title.
Include all applicable information to help us reproduce your problem.
To help us debug your issue please explain:
-->
### Environment Details (include every applicable attribute)
* Operating System+version: macos latest
* Compiler+version: clang 14
* Conan version: 1.53
* Python version: 3.9
### Steps to reproduce (Include if Applicable)
```
git clone https://github.com/nativium/nativium.git -b fix-conan-1-53
cd nativium
python3 nativium.py conan setup
python3 nativium.py target ios setup --arch=arm64 --build=debug --group=ios
python3 nativium.py target ios build --arch=arm64 --build=debug --group=ios
```
### Logs (Executed commands with output) (Include/Attach if Applicable)
```
CMake Error at /opt/homebrew/Cellar/cmake/3.24.2/share/cmake/Modules/Platform/iOS-Initialize.cmake:4 (message):
/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX12.3.sdk
is not an iOS SDK
Call Stack (most recent call first):
/opt/homebrew/Cellar/cmake/3.24.2/share/cmake/Modules/CMakeSystemSpecificInitialize.cmake:21 (include)
CMakeLists.txt:8 (project)
```
The problem is here:
https://github.com/nativium/nativium/blob/fix-conan-1-53/conan/darwin-toolchain/conanfile.py#L102
The `XCRun` object from conan don't check correctly the settings. It get the settings from conanfile, but need be something like i do, that is try to get `settings_target`. Or make XCRun receive `settings` and not `conanfile`:
```
settings_target = None
if hasattr(self, "settings_target"):
settings_target = self.settings_target
self.output.info("Settings: Target")
else:
settings_target = self.settings
self.output.info("Settings: Default")
```
The conan XCRun lines is here:
https://github.com/conan-io/conan/blob/develop/conan/tools/apple/apple.py#L82-L87 | priority | conan think that macos sdk is an ios sdk please don t forget to update the issue title include all applicable information to help us reproduce your problem to help us debug your issue please explain environment details include every applicable attribute operating system version macos latest compiler version clang conan version python version steps to reproduce include if applicable git clone b fix conan cd nativium nativium py conan setup nativium py target ios setup arch build debug group ios nativium py target ios build arch build debug group ios logs executed commands with output include attach if applicable cmake error at opt homebrew cellar cmake share cmake modules platform ios initialize cmake message applications xcode app contents developer platforms macosx platform developer sdks sdk is not an ios sdk call stack most recent call first opt homebrew cellar cmake share cmake modules cmakesystemspecificinitialize cmake include cmakelists txt project the problem is here the xcrun object from conan don t check correctly the settings it get the settings from conanfile but need be something like i do that is try to get settings target or make xcrun receive settings and not conanfile settings target none if hasattr self settings target settings target self settings target self output info settings target else settings target self settings self output info settings default the conan xcrun lines is here | 1 |
502,153 | 14,541,307,728 | IssuesEvent | 2020-12-15 14:25:15 | visual-framework/vf-core | https://api.github.com/repos/visual-framework/vf-core | closed | PROPOSAL - Move document widths to design tokens | Priority: Low Type: Enhancement Type: Normalisation | This could be a `vf-document-widths` or `vf-page-grid` or such.
The use case is two-fold:
1. Moving some of the document widths to design tokens would save a lot of repetition (you can find 288px, and 76.5 em many times in the code; see screenshot)
2. It would also allow for more flexibility for non-core-EMBL projects to have a different width

Also possible inclusions for the tokens is `vf-sass-config/variables/vf-global-custom-properties.scss`
```css
// Native CSS properties, currently limited to document spacing.
:root {
--page-grid-gap: 16px;
--embl-grid-module--prime: 200px;
--embl-grid-spacing-normaliser: 6px;
@media (min-width: 1200px) {
--page-grid-gap: 30px;
}
}
```
And i think these might not be used anywhere `vf-sass-config/variables/vf-global-variables.scss`:
```css
// grid variables
$global-grid-column-gap: 1em !default;
$global-page-max-width: 76.5em !default;
``` | 1.0 | PROPOSAL - Move document widths to design tokens - This could be a `vf-document-widths` or `vf-page-grid` or such.
The use case is two-fold:
1. Moving some of the document widths to design tokens would save a lot of repetition (you can find 288px, and 76.5 em many times in the code; see screenshot)
2. It would also allow for more flexibility for non-core-EMBL projects to have a different width

Also possible inclusions for the tokens is `vf-sass-config/variables/vf-global-custom-properties.scss`
```css
// Native CSS properties, currently limited to document spacing.
:root {
--page-grid-gap: 16px;
--embl-grid-module--prime: 200px;
--embl-grid-spacing-normaliser: 6px;
@media (min-width: 1200px) {
--page-grid-gap: 30px;
}
}
```
And i think these might not be used anywhere `vf-sass-config/variables/vf-global-variables.scss`:
```css
// grid variables
$global-grid-column-gap: 1em !default;
$global-page-max-width: 76.5em !default;
``` | priority | proposal move document widths to design tokens this could be a vf document widths or vf page grid or such the use case is two fold moving some of the document widths to design tokens would save a lot of repetition you can find and em many times in the code see screenshot it would also allow for more flexibility for non core embl projects to have a different width also possible inclusions for the tokens is vf sass config variables vf global custom properties scss css native css properties currently limited to document spacing root page grid gap embl grid module prime embl grid spacing normaliser media min width page grid gap and i think these might not be used anywhere vf sass config variables vf global variables scss css grid variables global grid column gap default global page max width default | 1 |
81,633 | 3,592,967,926 | IssuesEvent | 2016-02-01 17:55:05 | leeensminger/DelDOT-NPDES-Field-Tool | https://api.github.com/repos/leeensminger/DelDOT-NPDES-Field-Tool | opened | Investigation Reason of Targeted | bug - high priority | If "Targeted" is selected for Investigation Reason, then Reported By and External Report Type should grey out.

| 1.0 | Investigation Reason of Targeted - If "Targeted" is selected for Investigation Reason, then Reported By and External Report Type should grey out.

| priority | investigation reason of targeted if targeted is selected for investigation reason then reported by and external report type should grey out | 1 |
525,764 | 15,264,466,903 | IssuesEvent | 2021-02-22 05:27:39 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | XML template does not work properly with query expressions | CompilerSLDump Priority/Blocker SwanLakeDump Team/CompilerFE Type/SpecDeviation | Consider the following example
```ballerina
import ballerina/io;
type Person record {|
string name;
string country;
|};
function personsToXml(Person[] persons) returns xml {
return xml`<data>${
from var {name, country} in Persons
select xml`<person country="${country}">${name}</person>`
}</data>`;
}
public function main() {
Person p = {name: "FDF", country: "FFFF"};
Person e = {name:"EEEE", country: "EWFRWRFW"};
Person r = {name: "FFDF", country: "FDFD"};
Person[] arr = [];
arr.push(p);
arr.push(e);
arr.push(r);
xml ee = personsToXml(arr);
io:println(ee);
}
```
When compiling the above code, the following errors are printed. But the program should compile propery and the xml should be printed
```
ERROR [test.bal:(11:7,12:64)] incompatible types: expected '(int|float|decimal|string|boolean|xml)', found 'xml:Element[]'
ERROR [test.bal:(11:16,11:31)] invalid record binding pattern with type 'other'
ERROR [test.bal:(11:35,11:42)] undefined symbol 'Persons'
ERROR [test.bal:(12:37,12:44)] undefined symbol 'country'
ERROR [test.bal:(12:49,12:53)] undefined symbol 'name'
error: compilation contains errors
``` | 1.0 | XML template does not work properly with query expressions - Consider the following example
```ballerina
import ballerina/io;
type Person record {|
string name;
string country;
|};
function personsToXml(Person[] persons) returns xml {
return xml`<data>${
from var {name, country} in Persons
select xml`<person country="${country}">${name}</person>`
}</data>`;
}
public function main() {
Person p = {name: "FDF", country: "FFFF"};
Person e = {name:"EEEE", country: "EWFRWRFW"};
Person r = {name: "FFDF", country: "FDFD"};
Person[] arr = [];
arr.push(p);
arr.push(e);
arr.push(r);
xml ee = personsToXml(arr);
io:println(ee);
}
```
When compiling the above code, the following errors are printed. But the program should compile propery and the xml should be printed
```
ERROR [test.bal:(11:7,12:64)] incompatible types: expected '(int|float|decimal|string|boolean|xml)', found 'xml:Element[]'
ERROR [test.bal:(11:16,11:31)] invalid record binding pattern with type 'other'
ERROR [test.bal:(11:35,11:42)] undefined symbol 'Persons'
ERROR [test.bal:(12:37,12:44)] undefined symbol 'country'
ERROR [test.bal:(12:49,12:53)] undefined symbol 'name'
error: compilation contains errors
``` | priority | xml template does not work properly with query expressions consider the following example ballerina import ballerina io type person record string name string country function personstoxml person persons returns xml return xml from var name country in persons select xml name public function main person p name fdf country ffff person e name eeee country ewfrwrfw person r name ffdf country fdfd person arr arr push p arr push e arr push r xml ee personstoxml arr io println ee when compiling the above code the following errors are printed but the program should compile propery and the xml should be printed error incompatible types expected int float decimal string boolean xml found xml element error invalid record binding pattern with type other error undefined symbol persons error undefined symbol country error undefined symbol name error compilation contains errors | 1 |
401,997 | 27,346,134,708 | IssuesEvent | 2023-02-27 05:13:40 | Eu4ng/TIL | https://api.github.com/repos/Eu4ng/TIL | opened | [UE5/GAS] 입력 액션 바인딩 | documentation | ```cpp
void AGASPlayerCharacter::SetupPlayerInputComponent(class UInputComponent* PlayerInputComponent)
{
UGASEnhancedInputComponent* GASEnhancedInputComponent = CastChecked<UGASEnhancedInputComponent>(PlayerInputComponent);
//Make sure to set your input component class in the InputSettings->DefaultClasses
check(GASEnhancedInputComponent);
const FGASGameplayTags& GameplayTags = FGASGameplayTags::Get();
//Bind Input actions by tag
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Move, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Move);
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Look, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Look);
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Jump, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Jump);
BindAscInput();
}
```
```cpp
void AGASPlayerCharacter::BindAscInput()
{
if(!ASCInputBound && AbilitySystemComponent.IsValid() && IsValid(InputComponent))
{
AbilitySystemComponent->BindAbilityActivationToInputComponent(
InputComponent,
FGameplayAbilityInputBinds(
FString("ConfirmTarget"),
FString("CancelTarget"),
FString("GASAbilityID"),
static_cast<int32>(EGASAbilityID::Confirm),
static_cast<int32>(EGASAbilityID::Cancel)));
ASCInputBound = true;
}
}
``` | 1.0 | [UE5/GAS] 입력 액션 바인딩 - ```cpp
void AGASPlayerCharacter::SetupPlayerInputComponent(class UInputComponent* PlayerInputComponent)
{
UGASEnhancedInputComponent* GASEnhancedInputComponent = CastChecked<UGASEnhancedInputComponent>(PlayerInputComponent);
//Make sure to set your input component class in the InputSettings->DefaultClasses
check(GASEnhancedInputComponent);
const FGASGameplayTags& GameplayTags = FGASGameplayTags::Get();
//Bind Input actions by tag
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Move, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Move);
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Look, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Look);
GASEnhancedInputComponent->BindActionByTag(InputConfig, GameplayTags.InputTag_Jump, ETriggerEvent::Triggered, this, &AGASPlayerCharacter::Input_Jump);
BindAscInput();
}
```
```cpp
void AGASPlayerCharacter::BindAscInput()
{
if(!ASCInputBound && AbilitySystemComponent.IsValid() && IsValid(InputComponent))
{
AbilitySystemComponent->BindAbilityActivationToInputComponent(
InputComponent,
FGameplayAbilityInputBinds(
FString("ConfirmTarget"),
FString("CancelTarget"),
FString("GASAbilityID"),
static_cast<int32>(EGASAbilityID::Confirm),
static_cast<int32>(EGASAbilityID::Cancel)));
ASCInputBound = true;
}
}
``` | non_priority | 입력 액션 바인딩 cpp void agasplayercharacter setupplayerinputcomponent class uinputcomponent playerinputcomponent ugasenhancedinputcomponent gasenhancedinputcomponent castchecked playerinputcomponent make sure to set your input component class in the inputsettings defaultclasses check gasenhancedinputcomponent const fgasgameplaytags gameplaytags fgasgameplaytags get bind input actions by tag gasenhancedinputcomponent bindactionbytag inputconfig gameplaytags inputtag move etriggerevent triggered this agasplayercharacter input move gasenhancedinputcomponent bindactionbytag inputconfig gameplaytags inputtag look etriggerevent triggered this agasplayercharacter input look gasenhancedinputcomponent bindactionbytag inputconfig gameplaytags inputtag jump etriggerevent triggered this agasplayercharacter input jump bindascinput cpp void agasplayercharacter bindascinput if ascinputbound abilitysystemcomponent isvalid isvalid inputcomponent abilitysystemcomponent bindabilityactivationtoinputcomponent inputcomponent fgameplayabilityinputbinds fstring confirmtarget fstring canceltarget fstring gasabilityid static cast egasabilityid confirm static cast egasabilityid cancel ascinputbound true | 0 |
21,892 | 14,919,091,430 | IssuesEvent | 2021-01-22 23:08:15 | aguirre-lab/ml4c3 | https://api.github.com/repos/aguirre-lab/ml4c3 | closed | New ml4c3 recipe: clustering | infrastructure 🚇 | <!--- Provide a general summary of the issue in the Title above -->
## Why it Matters
<!--- Describe why this new feature or enhancement should be implemented, etc -->
## Possible Solution
<!-- High-level overview of how you propose to address -->
## Acceptance Criteria
<!--- Unambiguous milestones; if any are incomplete, the PR cannot be merged -->
## Blocked by or Pending
<!--- Links to other issues or PRs that block this issue and should be addressed or merged first. Also, label this issue with "`blocked 🚫`" -->
| 1.0 | New ml4c3 recipe: clustering - <!--- Provide a general summary of the issue in the Title above -->
## Why it Matters
<!--- Describe why this new feature or enhancement should be implemented, etc -->
## Possible Solution
<!-- High-level overview of how you propose to address -->
## Acceptance Criteria
<!--- Unambiguous milestones; if any are incomplete, the PR cannot be merged -->
## Blocked by or Pending
<!--- Links to other issues or PRs that block this issue and should be addressed or merged first. Also, label this issue with "`blocked 🚫`" -->
| non_priority | new recipe clustering why it matters possible solution acceptance criteria blocked by or pending | 0 |
293,509 | 8,996,805,600 | IssuesEvent | 2019-02-02 04:55:41 | minio/minio | https://api.github.com/repos/minio/minio | closed | Docker Swarm and minio distributed - secrets issue | community priority: medium triage |
Following this doc https://docs.minio.io/docs/deploy-minio-on-docker-swarm - results in the error below.
Maybe this is docker secrets issue - can I set up the stack on swarm via a different method?
**"starting container failed: RemoveSecretsPath failed: remove /var/lib/docker/containers/c6c4623343663d14cf3d551ae9067d32b2df7239e4f6d32fb36f7cd8cc13704b/secrets/access_key: read-only file system"**
## Expected Behavior
Swarm services should start
## Current Behavior
"starting container failed: RemoveSecretsPath failed: remove /var/lib/docker/containers/c6c4623343663d14cf3d551ae9067d32b2df7239e4f6d32fb36f7cd8cc13704b/secrets/access_key: read-only file system"
## Steps to Reproduce (for bugs)
Follow this doc for Centos 7 https://docs.minio.io/docs/deploy-minio-on-docker-swarm
| 1.0 | Docker Swarm and minio distributed - secrets issue -
Following this doc https://docs.minio.io/docs/deploy-minio-on-docker-swarm - results in the error below.
Maybe this is docker secrets issue - can I set up the stack on swarm via a different method?
**"starting container failed: RemoveSecretsPath failed: remove /var/lib/docker/containers/c6c4623343663d14cf3d551ae9067d32b2df7239e4f6d32fb36f7cd8cc13704b/secrets/access_key: read-only file system"**
## Expected Behavior
Swarm services should start
## Current Behavior
"starting container failed: RemoveSecretsPath failed: remove /var/lib/docker/containers/c6c4623343663d14cf3d551ae9067d32b2df7239e4f6d32fb36f7cd8cc13704b/secrets/access_key: read-only file system"
## Steps to Reproduce (for bugs)
Follow this doc for Centos 7 https://docs.minio.io/docs/deploy-minio-on-docker-swarm
| priority | docker swarm and minio distributed secrets issue following this doc results in the error below maybe this is docker secrets issue can i set up the stack on swarm via a different method starting container failed removesecretspath failed remove var lib docker containers secrets access key read only file system expected behavior swarm services should start current behavior starting container failed removesecretspath failed remove var lib docker containers secrets access key read only file system steps to reproduce for bugs follow this doc for centos | 1 |
524,818 | 15,223,837,541 | IssuesEvent | 2021-02-18 03:40:35 | ppy/osu | https://api.github.com/repos/ppy/osu | closed | Icon in volume mixer is too big | framework-fix-required low-priority platform:windows | It's a small bug, but it's one

Icon resolution is 64px, should be 32px | 1.0 | Icon in volume mixer is too big - It's a small bug, but it's one

Icon resolution is 64px, should be 32px | priority | icon in volume mixer is too big it s a small bug but it s one icon resolution is should be | 1 |
81,652 | 15,785,004,802 | IssuesEvent | 2021-04-01 15:48:02 | YSMull/blog | https://api.github.com/repos/YSMull/blog | opened | Minimum Window Substring | /leetcode/76/ leetcode | <div>原文链接: <a href="https://ysmull.cn/leetcode/76/">https://ysmull.cn/leetcode/76/</a></div><br><p><a href="https://leetcode.com/problems/minimum-window-substring/#/description">原题链接</a></p>
<h1 id="id-题目">题目</h1>
<blockquote>
<p>Given a string S and a string T, find the minimum window in S which will contain all the characters in T in complexity O(n).<br />
For example,<br />
S = “ADOBECODEBANC”<br />
T = “ABC”</p>
</blockquote>
<h1 id="id-解法">解法</h1>
<p>这道题目,我一开始用动态规划做,感觉有点难哦,只能看答案。<br />
讨论区有个人给了一个解决一类字符串问题的<a href="https://discuss.leetcode.com/topic/30941/here-is-a-10-line-template-that-can-solve-most-substring-problems">模板</a>:</p>
<blockquote>
<p>For most substring problem, we are given a string and need to find a substring of it which satisfy some restrictions. A general way is to use a hashmap assisted with two pointers. The template is given below.</p>
</blockquote>
<pre><code class="language-c">int findSubstring(string s) {
vector<int> map(128,0);
int counter; // check whether the substring is valid
int begin=0, end=0; //two pointers, one point to tail and one head
int d; //the length of substring
for() { /* initialize the hash map here */ }
while(end<s.size()){
if(map[s[end++]]-- ?){ /* modify counter here */ }
while(/* counter condition */){
/* update d here if finding minimum*/
//increase begin to make it invalid/valid again
if(map[s[begin++]]++ ?){ /*modify counter here*/ }
}
/* update d here if finding maximum*/
}
return d;
}
</code></pre>
<p>这个模板暗示的算法是,维护两个指针start和end分别指向子串的起始位置和结束位置,end向后遍历,当满足子串的性质之后,<br />
向后移动start破坏该性质并寻找下一个满足性质的位置。</p>
<p>用这个思想可以解决<a href="/blog/leetcode-3">Longest Substring Without Repeating Characters</a></p>
<h1 id="id-代码">代码</h1>
<pre><code class="language-c">string minWindow(string s, string t) {
int ascii[256];
memset(ascii, 0, 256 * sizeof(int));
int count = 0;
for (int i = 0; i < t.length(); i++) {
ascii[t[i]]++;
count++;
}
int start = 0, end = 0;
int minLen = 999999, minStart = 0;
while (end < s.length()) {
if (ascii[s[end]]-- > 0) {
count--;
}
while (count == 0) {
if (end - start + 1 < minLen) {
minStart = start;
minLen = end - start + 1;
}
if (++ascii[s[start++]] > 0) {
count++;
}
// if (start > end) break;
}
end++;
}
if (minLen == 999999 ) return "";
return s.substr(minStart, minLen);
}
</code></pre>
| 2.0 | Minimum Window Substring - <div>原文链接: <a href="https://ysmull.cn/leetcode/76/">https://ysmull.cn/leetcode/76/</a></div><br><p><a href="https://leetcode.com/problems/minimum-window-substring/#/description">原题链接</a></p>
<h1 id="id-题目">题目</h1>
<blockquote>
<p>Given a string S and a string T, find the minimum window in S which will contain all the characters in T in complexity O(n).<br />
For example,<br />
S = “ADOBECODEBANC”<br />
T = “ABC”</p>
</blockquote>
<h1 id="id-解法">解法</h1>
<p>这道题目,我一开始用动态规划做,感觉有点难哦,只能看答案。<br />
讨论区有个人给了一个解决一类字符串问题的<a href="https://discuss.leetcode.com/topic/30941/here-is-a-10-line-template-that-can-solve-most-substring-problems">模板</a>:</p>
<blockquote>
<p>For most substring problem, we are given a string and need to find a substring of it which satisfy some restrictions. A general way is to use a hashmap assisted with two pointers. The template is given below.</p>
</blockquote>
<pre><code class="language-c">int findSubstring(string s) {
vector<int> map(128,0);
int counter; // check whether the substring is valid
int begin=0, end=0; //two pointers, one point to tail and one head
int d; //the length of substring
for() { /* initialize the hash map here */ }
while(end<s.size()){
if(map[s[end++]]-- ?){ /* modify counter here */ }
while(/* counter condition */){
/* update d here if finding minimum*/
//increase begin to make it invalid/valid again
if(map[s[begin++]]++ ?){ /*modify counter here*/ }
}
/* update d here if finding maximum*/
}
return d;
}
</code></pre>
<p>这个模板暗示的算法是,维护两个指针start和end分别指向子串的起始位置和结束位置,end向后遍历,当满足子串的性质之后,<br />
向后移动start破坏该性质并寻找下一个满足性质的位置。</p>
<p>用这个思想可以解决<a href="/blog/leetcode-3">Longest Substring Without Repeating Characters</a></p>
<h1 id="id-代码">代码</h1>
<pre><code class="language-c">string minWindow(string s, string t) {
int ascii[256];
memset(ascii, 0, 256 * sizeof(int));
int count = 0;
for (int i = 0; i < t.length(); i++) {
ascii[t[i]]++;
count++;
}
int start = 0, end = 0;
int minLen = 999999, minStart = 0;
while (end < s.length()) {
if (ascii[s[end]]-- > 0) {
count--;
}
while (count == 0) {
if (end - start + 1 < minLen) {
minStart = start;
minLen = end - start + 1;
}
if (++ascii[s[start++]] > 0) {
count++;
}
// if (start > end) break;
}
end++;
}
if (minLen == 999999 ) return "";
return s.substr(minStart, minLen);
}
</code></pre>
| non_priority | minimum window substring 原文链接 a href href 题目 given a string s and a string t find the minimum window in s which will contain all the characters in t in complexity o n for example s “adobecodebanc” t “abc” 解法 这道题目,我一开始用动态规划做,感觉有点难哦,只能看答案。 讨论区有个人给了一个解决一类字符串问题的 a href for most substring problem we are given a string and need to find a substring of it which satisfy some restrictions a general way is to use a hashmap assisted with two pointers the template is given below int findsubstring string s vector lt int gt map int counter check whether the substring is valid int begin end two pointers one point to tail and one head int d the length of substring for initialize the hash map here while end lt s size if map modify counter here while counter condition update d here if finding minimum increase begin to make it invalid valid again if map modify counter here update d here if finding maximum return d 这个模板暗示的算法是,维护两个指针start和end分别指向子串的起始位置和结束位置,end向后遍历,当满足子串的性质之后, 向后移动start破坏该性质并寻找下一个满足性质的位置。 用这个思想可以解决 longest substring without repeating characters 代码 string minwindow string s string t int ascii memset ascii sizeof int int count for int i i lt t length i ascii count int start end int minlen minstart while end lt s length if ascii gt count while count if end start lt minlen minstart start minlen end start if ascii gt count if start gt end break end if minlen return return s substr minstart minlen | 0 |
312,873 | 23,446,535,515 | IssuesEvent | 2022-08-15 20:16:10 | Tombodil/Trope22 | https://api.github.com/repos/Tombodil/Trope22 | closed | Trope Website Content | documentation | Hey Tom
I have created a Google Doc with the content for the home page elements as well as changes to the blog.
https://docs.google.com/document/d/1HMMKFD-qi341cRV7aP9DyUMzfuCx2fwbU5nmvUSoYQg/edit | 1.0 | Trope Website Content - Hey Tom
I have created a Google Doc with the content for the home page elements as well as changes to the blog.
https://docs.google.com/document/d/1HMMKFD-qi341cRV7aP9DyUMzfuCx2fwbU5nmvUSoYQg/edit | non_priority | trope website content hey tom i have created a google doc with the content for the home page elements as well as changes to the blog | 0 |
642,467 | 20,888,499,402 | IssuesEvent | 2022-03-23 08:35:25 | returntocorp/semgrep | https://api.github.com/repos/returntocorp/semgrep | closed | FP: string constant inference not correct for non-literal-require | bug priority:medium lang:typescript feature:const-propagation | **Describe the bug**
https://semgrep.dev/s/0y95 should not trigger for the third case. Full string inference is hard but this is pretty basic and common and one line. The ? is to decide whether you are in prod/dev mode, e.g..
| 1.0 | FP: string constant inference not correct for non-literal-require - **Describe the bug**
https://semgrep.dev/s/0y95 should not trigger for the third case. Full string inference is hard but this is pretty basic and common and one line. The ? is to decide whether you are in prod/dev mode, e.g..
| priority | fp string constant inference not correct for non literal require describe the bug should not trigger for the third case full string inference is hard but this is pretty basic and common and one line the is to decide whether you are in prod dev mode e g | 1 |
324,380 | 9,888,744,837 | IssuesEvent | 2019-06-25 12:19:48 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | learningsynths.ableton.com - site is not usable | browser-fenix engine-gecko priority-normal | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://learningsynths.ableton.com/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Loading spinner forever
**Steps to Reproduce**:
Loaded the website, it just shows an (unnecessary) loading spinner that never goes away
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | learningsynths.ableton.com - site is not usable - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://learningsynths.ableton.com/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Loading spinner forever
**Steps to Reproduce**:
Loaded the website, it just shows an (unnecessary) loading spinner that never goes away
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | learningsynths ableton com site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description loading spinner forever steps to reproduce loaded the website it just shows an unnecessary loading spinner that never goes away browser configuration none from with ❤️ | 1 |
237,762 | 18,168,792,628 | IssuesEvent | 2021-09-27 17:24:04 | adafruit/Adafruit_CircuitPython_SD | https://api.github.com/repos/adafruit/Adafruit_CircuitPython_SD | opened | Missing Type Annotations | documentation good first issue | There are missing type annotations for some functions in this library.
The `typing` module does not exist on CircuitPython devices so the import needs to be wrapped in try/except to catch the error for missing import. There is an example of how that is done here:
```python
try:
from typing import List, Tuple
except ImportError:
pass
```
Once imported the typing annotations for the argument type(s), and return type(s) can be added to the function signature. Here is an example of a function that has had this done already:
```python
def wrap_text_to_pixels(
string: str, max_width: int, font=None, indent0: str = "", indent1: str = ""
) -> List[str]:
```
If you are new to Git or Github we have a guide about contributing to our projects here: https://learn.adafruit.com/contribute-to-circuitpython-with-git-and-github
There is also a guide that covers our CI utilities and how to run them locally to ensure they will pass in Github Actions here: https://learn.adafruit.com/creating-and-sharing-a-circuitpython-library/check-your-code In particular the pages: `Sharing docs on ReadTheDocs` and `Check your code with pre-commit` contain the tools to install and commands to run locally to run the checks.
If you are attempting to resolve this issue and need help, you can post a comment on this issue and tag both @foamyguy and @kattni or reach out to us on Discord: https://adafru.it/discord in the `#circuitpython-dev` channel.
The following locations are reported by mypy to be missing type annotations:
- [ ] adafruit_sdcard.py:89
- [ ] adafruit_sdcard.py:105
- [ ] adafruit_sdcard.py:155
- [ ] adafruit_sdcard.py:164
- [ ] adafruit_sdcard.py:183
- [ ] adafruit_sdcard.py:198
- [ ] adafruit_sdcard.py:255
- [ ] adafruit_sdcard.py:304
- [ ] adafruit_sdcard.py:322
- [ ] adafruit_sdcard.py:345
- [ ] adafruit_sdcard.py:398
- [ ] adafruit_sdcard.py:432
- [ ] adafruit_sdcard.py:490 | 1.0 | Missing Type Annotations - There are missing type annotations for some functions in this library.
The `typing` module does not exist on CircuitPython devices so the import needs to be wrapped in try/except to catch the error for missing import. There is an example of how that is done here:
```python
try:
from typing import List, Tuple
except ImportError:
pass
```
Once imported the typing annotations for the argument type(s), and return type(s) can be added to the function signature. Here is an example of a function that has had this done already:
```python
def wrap_text_to_pixels(
string: str, max_width: int, font=None, indent0: str = "", indent1: str = ""
) -> List[str]:
```
If you are new to Git or Github we have a guide about contributing to our projects here: https://learn.adafruit.com/contribute-to-circuitpython-with-git-and-github
There is also a guide that covers our CI utilities and how to run them locally to ensure they will pass in Github Actions here: https://learn.adafruit.com/creating-and-sharing-a-circuitpython-library/check-your-code In particular the pages: `Sharing docs on ReadTheDocs` and `Check your code with pre-commit` contain the tools to install and commands to run locally to run the checks.
If you are attempting to resolve this issue and need help, you can post a comment on this issue and tag both @foamyguy and @kattni or reach out to us on Discord: https://adafru.it/discord in the `#circuitpython-dev` channel.
The following locations are reported by mypy to be missing type annotations:
- [ ] adafruit_sdcard.py:89
- [ ] adafruit_sdcard.py:105
- [ ] adafruit_sdcard.py:155
- [ ] adafruit_sdcard.py:164
- [ ] adafruit_sdcard.py:183
- [ ] adafruit_sdcard.py:198
- [ ] adafruit_sdcard.py:255
- [ ] adafruit_sdcard.py:304
- [ ] adafruit_sdcard.py:322
- [ ] adafruit_sdcard.py:345
- [ ] adafruit_sdcard.py:398
- [ ] adafruit_sdcard.py:432
- [ ] adafruit_sdcard.py:490 | non_priority | missing type annotations there are missing type annotations for some functions in this library the typing module does not exist on circuitpython devices so the import needs to be wrapped in try except to catch the error for missing import there is an example of how that is done here python try from typing import list tuple except importerror pass once imported the typing annotations for the argument type s and return type s can be added to the function signature here is an example of a function that has had this done already python def wrap text to pixels string str max width int font none str str list if you are new to git or github we have a guide about contributing to our projects here there is also a guide that covers our ci utilities and how to run them locally to ensure they will pass in github actions here in particular the pages sharing docs on readthedocs and check your code with pre commit contain the tools to install and commands to run locally to run the checks if you are attempting to resolve this issue and need help you can post a comment on this issue and tag both foamyguy and kattni or reach out to us on discord in the circuitpython dev channel the following locations are reported by mypy to be missing type annotations adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py adafruit sdcard py | 0 |
479,675 | 13,804,485,839 | IssuesEvent | 2020-10-11 09:12:41 | OlofSjogren/GoAyo | https://api.github.com/repos/OlofSjogren/GoAyo | opened | Refactor PickUserFragment to allow it take in a destination | enhancement high priority | ## Problem
Currently there is a problem with _PickUserFragment_ requiring to be initialized **after** a fragment so that it may pop backstack to the previous fragment. This works but it gets pretty complicated if you want to use _PickUserFragment_ **first**.
See _CreateGroupFragment_ for instance, when creating this fragment it checks for a condition to determine whether or not it want to navigate to _PickUserFragment_ and as a result you cannot pop backstack back to _CreateGroupFragment_ unless you have selected something.
## Solution
I suggest to refactor _PickUserViewModel_ and _PickUserFragment_ so that you can set a continue destination. You can still pop backstack with back button, but you should navigate to the given destination after having pressed the continue button **or** pop backstack if the destination is not set.
### Note
By doing so you wouldn't have to change any existing usages of _PickUserFragment_. You only need to specify a destination when you want to use _PickUserFragment_ first. | 1.0 | Refactor PickUserFragment to allow it take in a destination - ## Problem
Currently there is a problem with _PickUserFragment_ requiring to be initialized **after** a fragment so that it may pop backstack to the previous fragment. This works but it gets pretty complicated if you want to use _PickUserFragment_ **first**.
See _CreateGroupFragment_ for instance, when creating this fragment it checks for a condition to determine whether or not it want to navigate to _PickUserFragment_ and as a result you cannot pop backstack back to _CreateGroupFragment_ unless you have selected something.
## Solution
I suggest to refactor _PickUserViewModel_ and _PickUserFragment_ so that you can set a continue destination. You can still pop backstack with back button, but you should navigate to the given destination after having pressed the continue button **or** pop backstack if the destination is not set.
### Note
By doing so you wouldn't have to change any existing usages of _PickUserFragment_. You only need to specify a destination when you want to use _PickUserFragment_ first. | priority | refactor pickuserfragment to allow it take in a destination problem currently there is a problem with pickuserfragment requiring to be initialized after a fragment so that it may pop backstack to the previous fragment this works but it gets pretty complicated if you want to use pickuserfragment first see creategroupfragment for instance when creating this fragment it checks for a condition to determine whether or not it want to navigate to pickuserfragment and as a result you cannot pop backstack back to creategroupfragment unless you have selected something solution i suggest to refactor pickuserviewmodel and pickuserfragment so that you can set a continue destination you can still pop backstack with back button but you should navigate to the given destination after having pressed the continue button or pop backstack if the destination is not set note by doing so you wouldn t have to change any existing usages of pickuserfragment you only need to specify a destination when you want to use pickuserfragment first | 1 |
428,917 | 12,418,467,261 | IssuesEvent | 2020-05-23 00:26:59 | eclipse-ee4j/glassfish | https://api.github.com/repos/eclipse-ee4j/glassfish | closed | IE9 and Google Chrome only: Export a LB config xml not working properly | Component: admin_gui ERR: Assignee Priority: Major Stale Type: Bug | This issue is browser specific(Only with google chrome browser)
Steps
1) Login Admin Console and Open a LB.
2) Go to Export Tab and click Export Now.
We can see the loadbalancer xml file is downloaded. But after that "A long-running process has been detected. Please wait..." dialog comes up and the admin console page hangs.
The dialog doesn't fade away at all. If we close it manually and try to access Export Tab in a LB , the dialog comes up again and hangs
#### Environment
OEL, Google Chrome browser 16.0
#### Affected Versions
[3.1.2_dev] | 1.0 | IE9 and Google Chrome only: Export a LB config xml not working properly - This issue is browser specific(Only with google chrome browser)
Steps
1) Login Admin Console and Open a LB.
2) Go to Export Tab and click Export Now.
We can see the loadbalancer xml file is downloaded. But after that "A long-running process has been detected. Please wait..." dialog comes up and the admin console page hangs.
The dialog doesn't fade away at all. If we close it manually and try to access Export Tab in a LB , the dialog comes up again and hangs
#### Environment
OEL, Google Chrome browser 16.0
#### Affected Versions
[3.1.2_dev] | priority | and google chrome only export a lb config xml not working properly this issue is browser specific only with google chrome browser steps login admin console and open a lb go to export tab and click export now we can see the loadbalancer xml file is downloaded but after that a long running process has been detected please wait dialog comes up and the admin console page hangs the dialog doesn t fade away at all if we close it manually and try to access export tab in a lb the dialog comes up again and hangs environment oel google chrome browser affected versions | 1 |
43,341 | 23,201,801,791 | IssuesEvent | 2022-08-01 22:28:01 | keras-team/keras | https://api.github.com/repos/keras-team/keras | closed | Sub-classed constraint doesn't appear to be called when added via add_weight | type:bug/performance stat:awaiting response from contributor stalled | **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): A proprietary variant of RedHat, outside of my control
- TensorFlow installed from (source or binary): Binary
- TensorFlow version (use command below): 2.5.0
- Python version: 3.7.2
- CUDA/cuDNN version: 11.1.0/8.1.1
- GPU model and memory: Nvidia K40 12GB
**Describe the current behavior**
I have created a sub-classed layer that performs a weighted sum of the outputs of two sub_models.
In this layer, I create weights with the `add_weight` function which includes the use of a constraint.
The intention of the constraint is to normalize the weights of each input dimension/feature (i.e. sum to one).
I have created a custom weight normalization constraint which is sub-classed from `krs.constraints.Constraint`. The output weights from the constraint function for model consisting of 2 sub-models with each producing a 200 dimension/feature output. Thus, the input to the constraint would be 2 x 200 weight tensor and the columnwise sums should all be 1.0.
When providing an instance of my constraint class to the `add_weight` function, it appears that there it is never called. It is initialized though. When running `VectorWeightedSum_1` below the `tf.print` statement in the call function shows that the weights are never normalized. Also, the `tf.print` statement in `NormalizeSumWeights.call()` doesn't ever produce an output. (It also doesn't exit the program if I put a `sys.exit()` call inside the contraint's `call`)
If I instead use `VectorWeightedSum_2`, the `tf.print` statement demonstrates that weights are indeed normalized properly. This was done by creating a new function inside the `VectorWeightedSum_2` class as opposed to a constraint class. The line for this code is commented out in the code below.
**Describe the expected behavior**
The sub-classed constraint should have it's `call` function called.
The weights produced by the sub-classed constraint should sum to 1.0 along each column.
In the example code, the `tf.print` statements should be printed for both `_constraint` and the use of the `NormalizeSumWeights` class.
**[Contributing](https://www.tensorflow.org/community/contribute)**
- Do you want to contribute a PR? (yes/no): No
**Standalone code to reproduce the issue**
```
import numpy as np
import tensorflow as tf
import tensorflow.keras as krs
class NormalizeSumWeights(krs.constraints.Constraint):
def __init__(self,**kwargs):
super(NormalizeSumWeights,self).__init__(**kwargs)
def call(self,w):
tf.print("NormalizeSumWeights_________________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
class WeightedSum(krs.layers.Layer):
def __init__( self, n_models = 2, **kwargs):
super( WeightedSum, self ).__init__( **kwargs)
self.n_models = n_models
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,1),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = NormalizeSumWeights(),
trainable = True)
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,0],inputs[0].dtype))
return output
def get_config(self):
data = { "n_models": self.n_models}
return data
class VectorWeightedSum_1(krs.layers.Layer):
def __init__( self, n_models = 2,n_dimensions=200, **kwargs):
super( VectorWeightedSum_1, self ).__init__( **kwargs)
self.n_models = n_models
self.n_dimensions = n_dimensions
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,self.n_dimensions),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = NormalizeSumWeights(),
trainable = True)
def _constraint(self,w):
tf.print("Weights Normalization__________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,:],inputs[0].dtype))
tf.print("vectorWeightedSum___weights________________")
tf.print(self.sum_weights)
return output
class VectorWeightedSum_2(krs.layers.Layer):
def __init__( self, n_models = 2,n_dimensions=200, **kwargs):
super( VectorWeightedSum_2, self ).__init__( **kwargs)
self.n_models = n_models
self.n_dimensions = n_dimensions
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,self.n_dimensions),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = self._constraint,
trainable = True)
def _constraint(self,w):
tf.print("Weights Normalization__________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,:],inputs[0].dtype))
tf.print("vectorWeightedSum___weights________________")
tf.print(self.sum_weights)
return output
input_lf = krs.Input((4,))
x = input_lf
x = krs.layers.Dense(10,activation = 'relu')(x)
x = krs.layers.Dense(10,activation = 'relu')(x)
lf_out = krs.layers.Dense(200,activation = 'relu')(x)
lf_mod = krs.Model(input_lf,lf_out,name='lf')
input_hf_lin = krs.Input((204,))
x = input_hf_lin
x = krs.layers.Dense(10)(x)
x = krs.layers.Dense(10)(x)
hf_lin_out = krs.layers.Dense(200,activation = 'relu')(x)
hf_lin_mod = krs.Model(input_hf_lin,hf_lin_out,name='hf_linear')
input_hf_nonlin = krs.Input((14,))
x = input_hf_nonlin
x = krs.layers.Dense(10,activation = 'relu')(x)
x = krs.layers.Dense(10,activation = 'relu')(x)
hf_nonlin_out = krs.layers.Dense(200,activation = 'relu')(x)
hf_nonlin_mod = krs.Model(input_hf_lin,hf_lin_out,name='hf_nonlinear')
input_hf = krs.Input((204,))
x = input_hf
lin = hf_lin_mod(x)
nonlin = hf_nonlin_mod(x)
summed_out = VectorWeightedSum_1(n_models=2)([lin,nonlin])
#summed_out = VectorWeightedSum_2(n_models=2)([lin,nonlin])
hf_mod = krs.Model(input_hf,summed_out,name='hf')
input_full_mod = krs.Input((4,))
x = input_full_mod
low = lf_mod(x)
x = krs.layers.Concatenate()([low,x])
full_out = hf_mod(x)
full_mod = krs.Model(input_full_mod,outputs = {'low_fidelity':low,'high_fidelity':full_out},name='full_model')
opt = krs.optimizers.Adam()
loss = krs.losses.MSE
full_mod.compile(optimizer = opt,loss = loss)
x_train = np.random.uniform(0,10,(20,4))
y_train_low = np.random.uniform(0,10,(20,200))
y_train_high = np.random.uniform(0,10,(20,200))
y = {"low_fidelity": y_train_low,
"high_fidelity": y_train_high}
full_mod.fit(x_train,y,epochs=5,batch_size=1)
```
| True | Sub-classed constraint doesn't appear to be called when added via add_weight - **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): A proprietary variant of RedHat, outside of my control
- TensorFlow installed from (source or binary): Binary
- TensorFlow version (use command below): 2.5.0
- Python version: 3.7.2
- CUDA/cuDNN version: 11.1.0/8.1.1
- GPU model and memory: Nvidia K40 12GB
**Describe the current behavior**
I have created a sub-classed layer that performs a weighted sum of the outputs of two sub_models.
In this layer, I create weights with the `add_weight` function which includes the use of a constraint.
The intention of the constraint is to normalize the weights of each input dimension/feature (i.e. sum to one).
I have created a custom weight normalization constraint which is sub-classed from `krs.constraints.Constraint`. The output weights from the constraint function for model consisting of 2 sub-models with each producing a 200 dimension/feature output. Thus, the input to the constraint would be 2 x 200 weight tensor and the columnwise sums should all be 1.0.
When providing an instance of my constraint class to the `add_weight` function, it appears that there it is never called. It is initialized though. When running `VectorWeightedSum_1` below the `tf.print` statement in the call function shows that the weights are never normalized. Also, the `tf.print` statement in `NormalizeSumWeights.call()` doesn't ever produce an output. (It also doesn't exit the program if I put a `sys.exit()` call inside the contraint's `call`)
If I instead use `VectorWeightedSum_2`, the `tf.print` statement demonstrates that weights are indeed normalized properly. This was done by creating a new function inside the `VectorWeightedSum_2` class as opposed to a constraint class. The line for this code is commented out in the code below.
**Describe the expected behavior**
The sub-classed constraint should have it's `call` function called.
The weights produced by the sub-classed constraint should sum to 1.0 along each column.
In the example code, the `tf.print` statements should be printed for both `_constraint` and the use of the `NormalizeSumWeights` class.
**[Contributing](https://www.tensorflow.org/community/contribute)**
- Do you want to contribute a PR? (yes/no): No
**Standalone code to reproduce the issue**
```
import numpy as np
import tensorflow as tf
import tensorflow.keras as krs
class NormalizeSumWeights(krs.constraints.Constraint):
def __init__(self,**kwargs):
super(NormalizeSumWeights,self).__init__(**kwargs)
def call(self,w):
tf.print("NormalizeSumWeights_________________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
class WeightedSum(krs.layers.Layer):
def __init__( self, n_models = 2, **kwargs):
super( WeightedSum, self ).__init__( **kwargs)
self.n_models = n_models
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,1),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = NormalizeSumWeights(),
trainable = True)
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,0],inputs[0].dtype))
return output
def get_config(self):
data = { "n_models": self.n_models}
return data
class VectorWeightedSum_1(krs.layers.Layer):
def __init__( self, n_models = 2,n_dimensions=200, **kwargs):
super( VectorWeightedSum_1, self ).__init__( **kwargs)
self.n_models = n_models
self.n_dimensions = n_dimensions
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,self.n_dimensions),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = NormalizeSumWeights(),
trainable = True)
def _constraint(self,w):
tf.print("Weights Normalization__________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,:],inputs[0].dtype))
tf.print("vectorWeightedSum___weights________________")
tf.print(self.sum_weights)
return output
class VectorWeightedSum_2(krs.layers.Layer):
def __init__( self, n_models = 2,n_dimensions=200, **kwargs):
super( VectorWeightedSum_2, self ).__init__( **kwargs)
self.n_models = n_models
self.n_dimensions = n_dimensions
def build(self,input_shape):
self.sum_weights = self.add_weight(name="sum_weights",shape=(self.n_models,self.n_dimensions),
initializer = krs.initializers.RandomUniform(0,1.0),
constraint = self._constraint,
trainable = True)
def _constraint(self,w):
tf.print("Weights Normalization__________________________")
return tf.math.divide(w,tf.math.reduce_sum(w,axis=0))
def call(self,inputs):
output = tf.multiply(tf.cast(self.sum_weights[0,:],inputs[0].dtype),inputs[0])
for i in range(1,len(inputs)):
output += tf.multiply(inputs[i],tf.cast(self.sum_weights[i,:],inputs[0].dtype))
tf.print("vectorWeightedSum___weights________________")
tf.print(self.sum_weights)
return output
input_lf = krs.Input((4,))
x = input_lf
x = krs.layers.Dense(10,activation = 'relu')(x)
x = krs.layers.Dense(10,activation = 'relu')(x)
lf_out = krs.layers.Dense(200,activation = 'relu')(x)
lf_mod = krs.Model(input_lf,lf_out,name='lf')
input_hf_lin = krs.Input((204,))
x = input_hf_lin
x = krs.layers.Dense(10)(x)
x = krs.layers.Dense(10)(x)
hf_lin_out = krs.layers.Dense(200,activation = 'relu')(x)
hf_lin_mod = krs.Model(input_hf_lin,hf_lin_out,name='hf_linear')
input_hf_nonlin = krs.Input((14,))
x = input_hf_nonlin
x = krs.layers.Dense(10,activation = 'relu')(x)
x = krs.layers.Dense(10,activation = 'relu')(x)
hf_nonlin_out = krs.layers.Dense(200,activation = 'relu')(x)
hf_nonlin_mod = krs.Model(input_hf_lin,hf_lin_out,name='hf_nonlinear')
input_hf = krs.Input((204,))
x = input_hf
lin = hf_lin_mod(x)
nonlin = hf_nonlin_mod(x)
summed_out = VectorWeightedSum_1(n_models=2)([lin,nonlin])
#summed_out = VectorWeightedSum_2(n_models=2)([lin,nonlin])
hf_mod = krs.Model(input_hf,summed_out,name='hf')
input_full_mod = krs.Input((4,))
x = input_full_mod
low = lf_mod(x)
x = krs.layers.Concatenate()([low,x])
full_out = hf_mod(x)
full_mod = krs.Model(input_full_mod,outputs = {'low_fidelity':low,'high_fidelity':full_out},name='full_model')
opt = krs.optimizers.Adam()
loss = krs.losses.MSE
full_mod.compile(optimizer = opt,loss = loss)
x_train = np.random.uniform(0,10,(20,4))
y_train_low = np.random.uniform(0,10,(20,200))
y_train_high = np.random.uniform(0,10,(20,200))
y = {"low_fidelity": y_train_low,
"high_fidelity": y_train_high}
full_mod.fit(x_train,y,epochs=5,batch_size=1)
```
| non_priority | sub classed constraint doesn t appear to be called when added via add weight system information have i written custom code as opposed to using a stock example script provided in tensorflow yes os platform and distribution e g linux ubuntu a proprietary variant of redhat outside of my control tensorflow installed from source or binary binary tensorflow version use command below python version cuda cudnn version gpu model and memory nvidia describe the current behavior i have created a sub classed layer that performs a weighted sum of the outputs of two sub models in this layer i create weights with the add weight function which includes the use of a constraint the intention of the constraint is to normalize the weights of each input dimension feature i e sum to one i have created a custom weight normalization constraint which is sub classed from krs constraints constraint the output weights from the constraint function for model consisting of sub models with each producing a dimension feature output thus the input to the constraint would be x weight tensor and the columnwise sums should all be when providing an instance of my constraint class to the add weight function it appears that there it is never called it is initialized though when running vectorweightedsum below the tf print statement in the call function shows that the weights are never normalized also the tf print statement in normalizesumweights call doesn t ever produce an output it also doesn t exit the program if i put a sys exit call inside the contraint s call if i instead use vectorweightedsum the tf print statement demonstrates that weights are indeed normalized properly this was done by creating a new function inside the vectorweightedsum class as opposed to a constraint class the line for this code is commented out in the code below describe the expected behavior the sub classed constraint should have it s call function called the weights produced by the sub classed constraint should sum to along each column in the example code the tf print statements should be printed for both constraint and the use of the normalizesumweights class do you want to contribute a pr yes no no standalone code to reproduce the issue import numpy as np import tensorflow as tf import tensorflow keras as krs class normalizesumweights krs constraints constraint def init self kwargs super normalizesumweights self init kwargs def call self w tf print normalizesumweights return tf math divide w tf math reduce sum w axis class weightedsum krs layers layer def init self n models kwargs super weightedsum self init kwargs self n models n models def build self input shape self sum weights self add weight name sum weights shape self n models initializer krs initializers randomuniform constraint normalizesumweights trainable true def call self inputs output tf multiply tf cast self sum weights inputs dtype inputs for i in range len inputs output tf multiply inputs tf cast self sum weights inputs dtype return output def get config self data n models self n models return data class vectorweightedsum krs layers layer def init self n models n dimensions kwargs super vectorweightedsum self init kwargs self n models n models self n dimensions n dimensions def build self input shape self sum weights self add weight name sum weights shape self n models self n dimensions initializer krs initializers randomuniform constraint normalizesumweights trainable true def constraint self w tf print weights normalization return tf math divide w tf math reduce sum w axis def call self inputs output tf multiply tf cast self sum weights inputs dtype inputs for i in range len inputs output tf multiply inputs tf cast self sum weights inputs dtype tf print vectorweightedsum weights tf print self sum weights return output class vectorweightedsum krs layers layer def init self n models n dimensions kwargs super vectorweightedsum self init kwargs self n models n models self n dimensions n dimensions def build self input shape self sum weights self add weight name sum weights shape self n models self n dimensions initializer krs initializers randomuniform constraint self constraint trainable true def constraint self w tf print weights normalization return tf math divide w tf math reduce sum w axis def call self inputs output tf multiply tf cast self sum weights inputs dtype inputs for i in range len inputs output tf multiply inputs tf cast self sum weights inputs dtype tf print vectorweightedsum weights tf print self sum weights return output input lf krs input x input lf x krs layers dense activation relu x x krs layers dense activation relu x lf out krs layers dense activation relu x lf mod krs model input lf lf out name lf input hf lin krs input x input hf lin x krs layers dense x x krs layers dense x hf lin out krs layers dense activation relu x hf lin mod krs model input hf lin hf lin out name hf linear input hf nonlin krs input x input hf nonlin x krs layers dense activation relu x x krs layers dense activation relu x hf nonlin out krs layers dense activation relu x hf nonlin mod krs model input hf lin hf lin out name hf nonlinear input hf krs input x input hf lin hf lin mod x nonlin hf nonlin mod x summed out vectorweightedsum n models summed out vectorweightedsum n models hf mod krs model input hf summed out name hf input full mod krs input x input full mod low lf mod x x krs layers concatenate full out hf mod x full mod krs model input full mod outputs low fidelity low high fidelity full out name full model opt krs optimizers adam loss krs losses mse full mod compile optimizer opt loss loss x train np random uniform y train low np random uniform y train high np random uniform y low fidelity y train low high fidelity y train high full mod fit x train y epochs batch size | 0 |
308,693 | 9,442,092,763 | IssuesEvent | 2019-04-15 05:14:35 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | opened | www.google.com - site is not usable | browser-firefox-mobile priority-critical | <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=http%3A%2F%2Fvcurb.us%2F4is3%3Fc%3Dz7mu9%26pn%3D2ul7h6j&ie=utf-8&oe=utf-8&client=firefox-b-1-m
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: can't connect
**Steps to Reproduce**:
Message and no connection symptom and Google side kick couldn't either
[](https://webcompat.com/uploads/2019/4/636fa37e-6e33-4a30-b894-c1f630757a86.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190331141835</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.google.com - site is not usable - <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=http%3A%2F%2Fvcurb.us%2F4is3%3Fc%3Dz7mu9%26pn%3D2ul7h6j&ie=utf-8&oe=utf-8&client=firefox-b-1-m
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: can't connect
**Steps to Reproduce**:
Message and no connection symptom and Google side kick couldn't either
[](https://webcompat.com/uploads/2019/4/636fa37e-6e33-4a30-b894-c1f630757a86.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190331141835</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description can t connect steps to reproduce message and no connection symptom and google side kick couldn t either browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️ | 1 |
114,852 | 17,266,863,268 | IssuesEvent | 2021-07-22 14:43:57 | turkdevops/php-src | https://api.github.com/repos/turkdevops/php-src | closed | CVE-2019-11036 (High) detected in php-srcphp-7.1.0RC3 - autoclosed | security vulnerability | ## CVE-2019-11036 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary>
<p>
<p>The PHP Interpreter</p>
<p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p>
<p>Found in base branch: <b>microseconds</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When processing certain files, PHP EXIF extension in versions 7.1.x below 7.1.29, 7.2.x below 7.2.18 and 7.3.x below 7.3.5 can be caused to read past allocated buffer in exif_process_IFD_TAG function. This may lead to information disclosure or crash.
<p>Publish Date: 2019-05-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11036>CVE-2019-11036</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11036">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11036</a></p>
<p>Release Date: 2019-05-03</p>
<p>Fix Resolution: No_FIX_Version</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-11036 (High) detected in php-srcphp-7.1.0RC3 - autoclosed - ## CVE-2019-11036 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary>
<p>
<p>The PHP Interpreter</p>
<p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p>
<p>Found in base branch: <b>microseconds</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When processing certain files, PHP EXIF extension in versions 7.1.x below 7.1.29, 7.2.x below 7.2.18 and 7.3.x below 7.3.5 can be caused to read past allocated buffer in exif_process_IFD_TAG function. This may lead to information disclosure or crash.
<p>Publish Date: 2019-05-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11036>CVE-2019-11036</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11036">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11036</a></p>
<p>Release Date: 2019-05-03</p>
<p>Fix Resolution: No_FIX_Version</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in php srcphp autoclosed cve high severity vulnerability vulnerable library php srcphp the php interpreter library home page a href found in head commit a href found in base branch microseconds vulnerable source files php src ext exif exif c vulnerability details when processing certain files php exif extension in versions x below x below and x below can be caused to read past allocated buffer in exif process ifd tag function this may lead to information disclosure or crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution no fix version step up your open source security game with whitesource | 0 |
189,019 | 6,793,170,805 | IssuesEvent | 2017-11-01 05:46:40 | wso2/product-ei | https://api.github.com/repos/wso2/product-ei | closed | NoSuchMethodError thrown when running sample "Introduction to XQuery Mediator" | Component/Integrator Priority/High Severity/Major Type/Bug | **Description:**
When sample 390 is run java.lang.NoSuchMethodError is thrown the server failed to respond.
**Affected Product Version:**
Observed the issue in synapse version (2.1.7-wso2v23-SNAPSHOT) that is in development.
**Steps to reproduce:**
1. Create a proxy with the following configuration
```xml
<definitions xmlns="http://ws.apache.org/ns/synapse">
<!-- the SimpleURLRegistry allows access to a URL based registry (e.g. file:/// or http://) -->
<registry provider="org.apache.synapse.registry.url.SimpleURLRegistry">
<!-- the root property of the simple URL registry helps resolve a resource URL as root + key -->
<parameter name="root">file:repository/conf/sample/resources/</parameter>
<!-- all resources loaded from the URL registry would be cached for this number of milli seconds -->
<parameter name="cachableDuration">15000</parameter>
</registry>
<localEntry key="xquery-key-req"
src="file:repository/conf/sample/resources/xquery/xquery_req.xq"/>
<proxy name="StockQuoteProxy">
<target>
<inSequence>
<property name="body" expression="$body/child::*[position()=1]"/>
<xquery key="xquery-key-req">
<variable name="payload" type="ELEMENT"/>
</xquery>
<send>
<endpoint>
<address uri="http://localhost:9000/services/SimpleStockQuoteService"/>
</endpoint>
</send>
</inSequence>
<outSequence>
<out>
<xquery key="xquery/xquery_res.xq">
<variable name="payload" type="ELEMENT"/>
<variable xmlns:m0="http://services.samples"
xmlns:ax21="http://services.samples/xsd" name="code" type="STRING"
expression="self::node()//m0:return/ax21:symbol/child::text()"/>
<variable xmlns:m0="http://services.samples"
xmlns:ax21="http://services.samples/xsd" name="price"
type="DOUBLE"
expression="self::node()//m0:return/ax21:last/child::text()"/>
</xquery>
<send/>
</out>
</outSequence>
</target>
<publishWSDL uri="file:repository/conf/sample/resources/proxy/sample_proxy_1.wsdl"/>
</proxy>
</definitions>
```
2. Send a request using stockquote client
```
ant stockquote -Daddurl=http://localhost:8280/services/StockQuoteProxy -Dmode=customquote
``` | 1.0 | NoSuchMethodError thrown when running sample "Introduction to XQuery Mediator" - **Description:**
When sample 390 is run java.lang.NoSuchMethodError is thrown the server failed to respond.
**Affected Product Version:**
Observed the issue in synapse version (2.1.7-wso2v23-SNAPSHOT) that is in development.
**Steps to reproduce:**
1. Create a proxy with the following configuration
```xml
<definitions xmlns="http://ws.apache.org/ns/synapse">
<!-- the SimpleURLRegistry allows access to a URL based registry (e.g. file:/// or http://) -->
<registry provider="org.apache.synapse.registry.url.SimpleURLRegistry">
<!-- the root property of the simple URL registry helps resolve a resource URL as root + key -->
<parameter name="root">file:repository/conf/sample/resources/</parameter>
<!-- all resources loaded from the URL registry would be cached for this number of milli seconds -->
<parameter name="cachableDuration">15000</parameter>
</registry>
<localEntry key="xquery-key-req"
src="file:repository/conf/sample/resources/xquery/xquery_req.xq"/>
<proxy name="StockQuoteProxy">
<target>
<inSequence>
<property name="body" expression="$body/child::*[position()=1]"/>
<xquery key="xquery-key-req">
<variable name="payload" type="ELEMENT"/>
</xquery>
<send>
<endpoint>
<address uri="http://localhost:9000/services/SimpleStockQuoteService"/>
</endpoint>
</send>
</inSequence>
<outSequence>
<out>
<xquery key="xquery/xquery_res.xq">
<variable name="payload" type="ELEMENT"/>
<variable xmlns:m0="http://services.samples"
xmlns:ax21="http://services.samples/xsd" name="code" type="STRING"
expression="self::node()//m0:return/ax21:symbol/child::text()"/>
<variable xmlns:m0="http://services.samples"
xmlns:ax21="http://services.samples/xsd" name="price"
type="DOUBLE"
expression="self::node()//m0:return/ax21:last/child::text()"/>
</xquery>
<send/>
</out>
</outSequence>
</target>
<publishWSDL uri="file:repository/conf/sample/resources/proxy/sample_proxy_1.wsdl"/>
</proxy>
</definitions>
```
2. Send a request using stockquote client
```
ant stockquote -Daddurl=http://localhost:8280/services/StockQuoteProxy -Dmode=customquote
``` | priority | nosuchmethoderror thrown when running sample introduction to xquery mediator description when sample is run java lang nosuchmethoderror is thrown the server failed to respond affected product version observed the issue in synapse version snapshot that is in development steps to reproduce create a proxy with the following configuration xml definitions xmlns file repository conf sample resources localentry key xquery key req src file repository conf sample resources xquery xquery req xq address uri variable xmlns xmlns name code type string expression self node return symbol child text variable xmlns xmlns name price type double expression self node return last child text send a request using stockquote client ant stockquote daddurl dmode customquote | 1 |
410,393 | 27,784,027,636 | IssuesEvent | 2023-03-17 00:31:55 | apache/daffodil-vscode | https://api.github.com/repos/apache/daffodil-vscode | closed | Document required glibc for Linux | documentation dependencies | The Ωedit Linux build is being done on Ubuntu 20.04 LTS. Linux OSes using different a version of glibc than then system we built it on will not be compatible. | 1.0 | Document required glibc for Linux - The Ωedit Linux build is being done on Ubuntu 20.04 LTS. Linux OSes using different a version of glibc than then system we built it on will not be compatible. | non_priority | document required glibc for linux the ωedit linux build is being done on ubuntu lts linux oses using different a version of glibc than then system we built it on will not be compatible | 0 |
93,389 | 3,899,529,354 | IssuesEvent | 2016-04-17 20:00:26 | TerraTerma/Core | https://api.github.com/repos/TerraTerma/Core | opened | [ToDo] Lobby commands | Easy Low priority Phase 1 | we need a way to set the map of the lobby by default. I would recommend copying the current skywars commands and changing them it suit the Lobby | 1.0 | [ToDo] Lobby commands - we need a way to set the map of the lobby by default. I would recommend copying the current skywars commands and changing them it suit the Lobby | priority | lobby commands we need a way to set the map of the lobby by default i would recommend copying the current skywars commands and changing them it suit the lobby | 1 |
10,176 | 4,716,311,958 | IssuesEvent | 2016-10-16 00:13:25 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Consider building the standard library with Cargo | A-build | This is distinct from #16091. Once we've hit `1.0.0`, there's no better way to dogfood that stability than have the standard library a normal part of the Cargo ecosystem. The main benefit is that cross-compiling suddenly becomes a lot easier, especially for bare-metal targets: just use the normal `--target` with `cargo`. It also means that the answer to "do I need to add a line to my `Cargo.toml`" becomes a lot easier.
Cargo would inject references to `std` etc automatically, with an opt-out flag at the package level for things such as `zinc` or operating system kernels.
Building `rustc` would become slightly more interesting, but would probably just involve juggling with the `PATH` so that Cargo is pointed at the right place. | 1.0 | Consider building the standard library with Cargo - This is distinct from #16091. Once we've hit `1.0.0`, there's no better way to dogfood that stability than have the standard library a normal part of the Cargo ecosystem. The main benefit is that cross-compiling suddenly becomes a lot easier, especially for bare-metal targets: just use the normal `--target` with `cargo`. It also means that the answer to "do I need to add a line to my `Cargo.toml`" becomes a lot easier.
Cargo would inject references to `std` etc automatically, with an opt-out flag at the package level for things such as `zinc` or operating system kernels.
Building `rustc` would become slightly more interesting, but would probably just involve juggling with the `PATH` so that Cargo is pointed at the right place. | non_priority | consider building the standard library with cargo this is distinct from once we ve hit there s no better way to dogfood that stability than have the standard library a normal part of the cargo ecosystem the main benefit is that cross compiling suddenly becomes a lot easier especially for bare metal targets just use the normal target with cargo it also means that the answer to do i need to add a line to my cargo toml becomes a lot easier cargo would inject references to std etc automatically with an opt out flag at the package level for things such as zinc or operating system kernels building rustc would become slightly more interesting but would probably just involve juggling with the path so that cargo is pointed at the right place | 0 |
311,022 | 9,527,147,687 | IssuesEvent | 2019-04-29 02:01:25 | Sonarr/Sonarr | https://api.github.com/repos/Sonarr/Sonarr | closed | Add TeamCity build for mono compile | priority:high suboptimal task | We need a separate TC build queue to compile and test using the mono compiler (but not as deliverable artifact). So we can verify the build keeps working.
Checklist
- [x] Create Build.sh
- [ ] Add Build queue
- [ ] Add VM to run build and tests (archlinux, freebsd, or something other than debian)
| 1.0 | Add TeamCity build for mono compile - We need a separate TC build queue to compile and test using the mono compiler (but not as deliverable artifact). So we can verify the build keeps working.
Checklist
- [x] Create Build.sh
- [ ] Add Build queue
- [ ] Add VM to run build and tests (archlinux, freebsd, or something other than debian)
| priority | add teamcity build for mono compile we need a separate tc build queue to compile and test using the mono compiler but not as deliverable artifact so we can verify the build keeps working checklist create build sh add build queue add vm to run build and tests archlinux freebsd or something other than debian | 1 |
785,530 | 27,616,992,960 | IssuesEvent | 2023-03-09 20:12:52 | Ore-Design/Ore-3D-Reports-Changelog | https://api.github.com/repos/Ore-Design/Ore-3D-Reports-Changelog | closed | Feature: Multi Select Add Label | enhancement in progress low priority | Can we gain the ability to multi select to add labels on fuse? | 1.0 | Feature: Multi Select Add Label - Can we gain the ability to multi select to add labels on fuse? | priority | feature multi select add label can we gain the ability to multi select to add labels on fuse | 1 |
629,577 | 20,036,420,762 | IssuesEvent | 2022-02-02 12:24:23 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | Event Status Banner Not Showing in Default Twenty Twenty-Two Theme | category:frontend-ui-&-ux type:theme/plugin-conflict 💥 priority:normal status:has-fix | The event Status banner does show in the default Twenty Twenty-One theme, but not the Twenty Twenty-Two theme:
Twenty Twenty-One:

Twenty Twenty-Two:

| 1.0 | Event Status Banner Not Showing in Default Twenty Twenty-Two Theme - The event Status banner does show in the default Twenty Twenty-One theme, but not the Twenty Twenty-Two theme:
Twenty Twenty-One:

Twenty Twenty-Two:

| priority | event status banner not showing in default twenty twenty two theme the event status banner does show in the default twenty twenty one theme but not the twenty twenty two theme twenty twenty one twenty twenty two | 1 |
72,370 | 13,822,794,071 | IssuesEvent | 2020-10-13 05:54:14 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | opened | CSharpUpgradeProjectCodeFixProvider sometimes reports the wrong version | Area-IDE Bug IDE-CodeStyle Integration-Test help wanted | **Version Used**: Periodically appearing in integration tests, e.g. #48372
**Steps to Reproduce**:
Run the `LegacyProject_AllConfigurationsUpdated` test.
**Expected Behavior**:
> Upgrade this project to C# language version '7.3'
**Actual Behavior**:
> Upgrade this project to C# language version 'default'

| 1.0 | CSharpUpgradeProjectCodeFixProvider sometimes reports the wrong version - **Version Used**: Periodically appearing in integration tests, e.g. #48372
**Steps to Reproduce**:
Run the `LegacyProject_AllConfigurationsUpdated` test.
**Expected Behavior**:
> Upgrade this project to C# language version '7.3'
**Actual Behavior**:
> Upgrade this project to C# language version 'default'

| non_priority | csharpupgradeprojectcodefixprovider sometimes reports the wrong version version used periodically appearing in integration tests e g steps to reproduce run the legacyproject allconfigurationsupdated test expected behavior upgrade this project to c language version actual behavior upgrade this project to c language version default | 0 |
101,866 | 16,529,728,071 | IssuesEvent | 2021-05-27 03:11:18 | alpersonalwebsite/postME | https://api.github.com/repos/alpersonalwebsite/postME | opened | WS-2019-0427 (Medium) detected in elliptic-6.4.1.tgz | security vulnerability | ## WS-2019-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.1.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/elliptic</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-4.44.2.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.0.4.tgz
- :x: **elliptic-6.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2
<p>Publish Date: 2019-11-22
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p>
<p>Release Date: 2020-05-24</p>
<p>Fix Resolution: v6.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0427 (Medium) detected in elliptic-6.4.1.tgz - ## WS-2019-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.1.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/elliptic</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-4.44.2.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.0.4.tgz
- :x: **elliptic-6.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2
<p>Publish Date: 2019-11-22
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p>
<p>Release Date: 2020-05-24</p>
<p>Fix Resolution: v6.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file postme frontend package json path to vulnerable library postme frontend node modules elliptic dependency hierarchy react scripts tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library found in head commit a href vulnerability details the function getnaf in elliptic library has information leakage this issue is mitigated in version publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
288,665 | 31,863,722,452 | IssuesEvent | 2023-09-15 12:48:34 | SecOpsNews/news | https://api.github.com/repos/SecOpsNews/news | opened | [SecurityWeek] Deduce Raises $9 Million to Fight AI-Generated Identity Fraud | SecurityWeek |
Deduce has raised $9 million in a new funding round led by Freestyle Capital, to launch its AI-generated identity fraud prevention platform.
The post [Deduce Raises $9 Million to Fight AI-Generated Identity Fraud](https://www.securityweek.com/deduce-raises-9-million-to-fight-ai-generated-identity-fraud/) appeared first on [SecurityWeek](https://www.securityweek.com).
<https://www.securityweek.com/deduce-raises-9-million-to-fight-ai-generated-identity-fraud/>
| True | [SecurityWeek] Deduce Raises $9 Million to Fight AI-Generated Identity Fraud -
Deduce has raised $9 million in a new funding round led by Freestyle Capital, to launch its AI-generated identity fraud prevention platform.
The post [Deduce Raises $9 Million to Fight AI-Generated Identity Fraud](https://www.securityweek.com/deduce-raises-9-million-to-fight-ai-generated-identity-fraud/) appeared first on [SecurityWeek](https://www.securityweek.com).
<https://www.securityweek.com/deduce-raises-9-million-to-fight-ai-generated-identity-fraud/>
| non_priority | deduce raises million to fight ai generated identity fraud deduce has raised million in a new funding round led by freestyle capital to launch its ai generated identity fraud prevention platform the post appeared first on | 0 |
193,337 | 22,216,134,738 | IssuesEvent | 2022-06-08 01:59:32 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | reopened | CVE-2017-18216 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2017-18216 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ocfs2/cluster/nodemanager.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ocfs2/cluster/nodemanager.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In fs/ocfs2/cluster/nodemanager.c in the Linux kernel before 4.15, local users can cause a denial of service (NULL pointer dereference and BUG) because a required mutex is not used.
<p>Publish Date: 2018-03-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18216>CVE-2017-18216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-18216">https://nvd.nist.gov/vuln/detail/CVE-2017-18216</a></p>
<p>Release Date: 2018-03-05</p>
<p>Fix Resolution: 4.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-18216 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2017-18216 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ocfs2/cluster/nodemanager.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ocfs2/cluster/nodemanager.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In fs/ocfs2/cluster/nodemanager.c in the Linux kernel before 4.15, local users can cause a denial of service (NULL pointer dereference and BUG) because a required mutex is not used.
<p>Publish Date: 2018-03-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18216>CVE-2017-18216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-18216">https://nvd.nist.gov/vuln/detail/CVE-2017-18216</a></p>
<p>Release Date: 2018-03-05</p>
<p>Fix Resolution: 4.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files fs cluster nodemanager c fs cluster nodemanager c vulnerability details in fs cluster nodemanager c in the linux kernel before local users can cause a denial of service null pointer dereference and bug because a required mutex is not used publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
451,030 | 32,004,618,606 | IssuesEvent | 2023-09-21 14:11:09 | bluecadet/launchpad | https://api.github.com/repos/bluecadet/launchpad | reopened | Move detailed docs to vitepress | documentation | Discussed in #116
- Move detailed docs to vitepress, build to github pages
- re-implement autogenerated docs from JSDoc comments (broken by #118) | 1.0 | Move detailed docs to vitepress - Discussed in #116
- Move detailed docs to vitepress, build to github pages
- re-implement autogenerated docs from JSDoc comments (broken by #118) | non_priority | move detailed docs to vitepress discussed in move detailed docs to vitepress build to github pages re implement autogenerated docs from jsdoc comments broken by | 0 |
36,050 | 2,795,130,321 | IssuesEvent | 2015-05-11 20:22:19 | cswarth/hiv-sim | https://api.github.com/repos/cswarth/hiv-sim | closed | Add consensus sequence as a method | priority | This will allow us to compare to the McCloskey results (at least more than we are currently able to do). | 1.0 | Add consensus sequence as a method - This will allow us to compare to the McCloskey results (at least more than we are currently able to do). | priority | add consensus sequence as a method this will allow us to compare to the mccloskey results at least more than we are currently able to do | 1 |
724,599 | 24,935,988,439 | IssuesEvent | 2022-10-31 15:10:42 | Poobslag/turbofat | https://api.github.com/repos/Poobslag/turbofat | opened | Losing "lives" should persist until the end of the career day | priority-2 | Currently, you can reach the end of the career day even if you top out a lot. We should try making "lives" persistent, so that if you top out 3 times across different levels your career day ends. | 1.0 | Losing "lives" should persist until the end of the career day - Currently, you can reach the end of the career day even if you top out a lot. We should try making "lives" persistent, so that if you top out 3 times across different levels your career day ends. | priority | losing lives should persist until the end of the career day currently you can reach the end of the career day even if you top out a lot we should try making lives persistent so that if you top out times across different levels your career day ends | 1 |
267,738 | 8,392,236,986 | IssuesEvent | 2018-10-09 17:02:48 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | oc adm diagnostics: unable to parse requirement | component/diagnostics kind/bug lifecycle/rotten priority/P2 | Running `ocm adm diagnostics` produces an error about the diagnostic pod.
##### Version
openshift v3.7.0+7ed6862
kubernetes v1.7.6+a08f5eeb62
etcd 3.2.8
##### Steps To Reproduce
1. run `oc adm diagnostics`
##### Current Result
```
ERROR: [DCli2001 from diagnostic DiagnosticPod@openshift/origin/pkg/diagnostics/client/run_diagnostics_pod.go:81]
Creating diagnostic pod with image openshift/origin-deployer:v3.7.0 failed. Error: (*errors.StatusError) unable to parse requirement: found '', expected: '='
```
##### Expected Result
No error.
##### Additional Information
The router deployment config also seems to exhibit this error:
```
[root@atomicmaster ~]# oc describe dc/router
Name: router
Namespace: default
Created: 5 weeks ago
Labels: router=router
Annotations: <none>
Latest Version: 9
Selector: router=router
Replicas: 1
Triggers: Config
Strategy: Rolling
Template:
Pod Template:
Labels: router=router
Service Account: router
Containers:
router:
Image: openshift/origin-haproxy-router:v3.7.0
Ports: 80/TCP, 443/TCP, 1936/TCP
Requests:
cpu: 100m
memory: 256Mi
Liveness: http-get http://localhost:1936/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Readiness: http-get http://localhost:1936/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Environment:
DEFAULT_CERTIFICATE_DIR: /etc/pki/tls/private
DEFAULT_CERTIFICATE_PATH: /etc/pki/tls/private/tls.crt
ROUTER_CIPHERS:
ROUTER_EXTERNAL_HOST_HOSTNAME:
ROUTER_EXTERNAL_HOST_HTTPS_VSERVER:
ROUTER_EXTERNAL_HOST_HTTP_VSERVER:
ROUTER_EXTERNAL_HOST_INSECURE: false
ROUTER_EXTERNAL_HOST_INTERNAL_ADDRESS:
ROUTER_EXTERNAL_HOST_PARTITION_PATH:
ROUTER_EXTERNAL_HOST_PASSWORD:
ROUTER_EXTERNAL_HOST_PRIVKEY: /etc/secret-volume/router.pem
ROUTER_EXTERNAL_HOST_USERNAME:
ROUTER_EXTERNAL_HOST_VXLAN_GW_CIDR:
ROUTER_LISTEN_ADDR: 0.0.0.0:1936
ROUTER_METRICS_TYPE: haproxy
ROUTER_SERVICE_HTTPS_PORT: 443
ROUTER_SERVICE_HTTP_PORT: 80
ROUTER_SERVICE_NAME: router
ROUTER_SERVICE_NAMESPACE: default
ROUTER_SUBDOMAIN:
STATS_PASSWORD: xxx
STATS_PORT: 1936
STATS_USERNAME: admin
Mounts:
/etc/pki/tls/private from server-certificate (ro)
Volumes:
server-certificate:
Type: Secret (a volume populated by a Secret)
SecretName: router-certs
Optional: false
Deployment #9 (latest):
Created: 27 minutes ago
Status: New
Replicas: 0 current / 0 desired
Deployment #8:
Created: 43 minutes ago
Status: Failed
Replicas: 0 current / 0 desired
Deployment #7:
Created: 2 weeks ago
Status: Failed
Replicas: 0 current / 0 desired
Events:
FirstSeen LastSeen Count From SubObjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
43m 43m 1 deploymentconfig-controller Normal DeploymentCreated Created new replication controller "router-8" for version 8
41m 41m 1 deployer-controller Warning FailedRetry Stop retrying: couldn't create deployer pod for "default/router-8": unable to parse requirement: found '', expected: '='
27m 27m 1 deployer-controller Normal RolloutCancelled Rollout for "default/router-8" cancelled
27m 27m 5 deploymentconfig-controller Normal DeploymentAwaitingCancellation Deployment of version 9 awaiting cancellation of older running deployments
27m 27m 1 deploymentconfig-controller Normal DeploymentCancelled Cancelled deployment "router-8" superceded by version 9
27m 27m 1 deploymentconfig-controller Normal DeploymentCreated Created new replication controller "router-9" for version 9
43m 27m 26 deployer-controller Warning FailedCreate Error creating deployer pod: unable to parse requirement: found '', expected: '='
```
| 1.0 | oc adm diagnostics: unable to parse requirement - Running `ocm adm diagnostics` produces an error about the diagnostic pod.
##### Version
openshift v3.7.0+7ed6862
kubernetes v1.7.6+a08f5eeb62
etcd 3.2.8
##### Steps To Reproduce
1. run `oc adm diagnostics`
##### Current Result
```
ERROR: [DCli2001 from diagnostic DiagnosticPod@openshift/origin/pkg/diagnostics/client/run_diagnostics_pod.go:81]
Creating diagnostic pod with image openshift/origin-deployer:v3.7.0 failed. Error: (*errors.StatusError) unable to parse requirement: found '', expected: '='
```
##### Expected Result
No error.
##### Additional Information
The router deployment config also seems to exhibit this error:
```
[root@atomicmaster ~]# oc describe dc/router
Name: router
Namespace: default
Created: 5 weeks ago
Labels: router=router
Annotations: <none>
Latest Version: 9
Selector: router=router
Replicas: 1
Triggers: Config
Strategy: Rolling
Template:
Pod Template:
Labels: router=router
Service Account: router
Containers:
router:
Image: openshift/origin-haproxy-router:v3.7.0
Ports: 80/TCP, 443/TCP, 1936/TCP
Requests:
cpu: 100m
memory: 256Mi
Liveness: http-get http://localhost:1936/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Readiness: http-get http://localhost:1936/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Environment:
DEFAULT_CERTIFICATE_DIR: /etc/pki/tls/private
DEFAULT_CERTIFICATE_PATH: /etc/pki/tls/private/tls.crt
ROUTER_CIPHERS:
ROUTER_EXTERNAL_HOST_HOSTNAME:
ROUTER_EXTERNAL_HOST_HTTPS_VSERVER:
ROUTER_EXTERNAL_HOST_HTTP_VSERVER:
ROUTER_EXTERNAL_HOST_INSECURE: false
ROUTER_EXTERNAL_HOST_INTERNAL_ADDRESS:
ROUTER_EXTERNAL_HOST_PARTITION_PATH:
ROUTER_EXTERNAL_HOST_PASSWORD:
ROUTER_EXTERNAL_HOST_PRIVKEY: /etc/secret-volume/router.pem
ROUTER_EXTERNAL_HOST_USERNAME:
ROUTER_EXTERNAL_HOST_VXLAN_GW_CIDR:
ROUTER_LISTEN_ADDR: 0.0.0.0:1936
ROUTER_METRICS_TYPE: haproxy
ROUTER_SERVICE_HTTPS_PORT: 443
ROUTER_SERVICE_HTTP_PORT: 80
ROUTER_SERVICE_NAME: router
ROUTER_SERVICE_NAMESPACE: default
ROUTER_SUBDOMAIN:
STATS_PASSWORD: xxx
STATS_PORT: 1936
STATS_USERNAME: admin
Mounts:
/etc/pki/tls/private from server-certificate (ro)
Volumes:
server-certificate:
Type: Secret (a volume populated by a Secret)
SecretName: router-certs
Optional: false
Deployment #9 (latest):
Created: 27 minutes ago
Status: New
Replicas: 0 current / 0 desired
Deployment #8:
Created: 43 minutes ago
Status: Failed
Replicas: 0 current / 0 desired
Deployment #7:
Created: 2 weeks ago
Status: Failed
Replicas: 0 current / 0 desired
Events:
FirstSeen LastSeen Count From SubObjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
43m 43m 1 deploymentconfig-controller Normal DeploymentCreated Created new replication controller "router-8" for version 8
41m 41m 1 deployer-controller Warning FailedRetry Stop retrying: couldn't create deployer pod for "default/router-8": unable to parse requirement: found '', expected: '='
27m 27m 1 deployer-controller Normal RolloutCancelled Rollout for "default/router-8" cancelled
27m 27m 5 deploymentconfig-controller Normal DeploymentAwaitingCancellation Deployment of version 9 awaiting cancellation of older running deployments
27m 27m 1 deploymentconfig-controller Normal DeploymentCancelled Cancelled deployment "router-8" superceded by version 9
27m 27m 1 deploymentconfig-controller Normal DeploymentCreated Created new replication controller "router-9" for version 9
43m 27m 26 deployer-controller Warning FailedCreate Error creating deployer pod: unable to parse requirement: found '', expected: '='
```
| priority | oc adm diagnostics unable to parse requirement running ocm adm diagnostics produces an error about the diagnostic pod version openshift kubernetes etcd steps to reproduce run oc adm diagnostics current result error creating diagnostic pod with image openshift origin deployer failed error errors statuserror unable to parse requirement found expected expected result no error additional information the router deployment config also seems to exhibit this error oc describe dc router name router namespace default created weeks ago labels router router annotations latest version selector router router replicas triggers config strategy rolling template pod template labels router router service account router containers router image openshift origin haproxy router ports tcp tcp tcp requests cpu memory liveness http get delay timeout period success failure readiness http get delay timeout period success failure environment default certificate dir etc pki tls private default certificate path etc pki tls private tls crt router ciphers router external host hostname router external host https vserver router external host http vserver router external host insecure false router external host internal address router external host partition path router external host password router external host privkey etc secret volume router pem router external host username router external host vxlan gw cidr router listen addr router metrics type haproxy router service https port router service http port router service name router router service namespace default router subdomain stats password xxx stats port stats username admin mounts etc pki tls private from server certificate ro volumes server certificate type secret a volume populated by a secret secretname router certs optional false deployment latest created minutes ago status new replicas current desired deployment created minutes ago status failed replicas current desired deployment created weeks ago status failed replicas current desired events firstseen lastseen count from subobjectpath type reason message deploymentconfig controller normal deploymentcreated created new replication controller router for version deployer controller warning failedretry stop retrying couldn t create deployer pod for default router unable to parse requirement found expected deployer controller normal rolloutcancelled rollout for default router cancelled deploymentconfig controller normal deploymentawaitingcancellation deployment of version awaiting cancellation of older running deployments deploymentconfig controller normal deploymentcancelled cancelled deployment router superceded by version deploymentconfig controller normal deploymentcreated created new replication controller router for version deployer controller warning failedcreate error creating deployer pod unable to parse requirement found expected | 1 |
537,707 | 15,733,739,468 | IssuesEvent | 2021-03-29 19:57:56 | JensenJ/EmbargoMC-IssueTracker | https://api.github.com/repos/JensenJ/EmbargoMC-IssueTracker | opened | [FEATURE] Multiple Nation Homes | enhancement low-priority | **Is your feature request related to a problem? Please describe.**
Nation home in other dimensions does not currently work unless a home is set in that dimension.
**Describe the solution you'd like**
One nation home per dimension, with the player being able to specify which home to go to. Usage would be /n home <home>
**Describe alternatives you've considered**
Multiple nation homes that can be set, max 3 per nation or an increasing amount based on number of players in nation.
| 1.0 | [FEATURE] Multiple Nation Homes - **Is your feature request related to a problem? Please describe.**
Nation home in other dimensions does not currently work unless a home is set in that dimension.
**Describe the solution you'd like**
One nation home per dimension, with the player being able to specify which home to go to. Usage would be /n home <home>
**Describe alternatives you've considered**
Multiple nation homes that can be set, max 3 per nation or an increasing amount based on number of players in nation.
| priority | multiple nation homes is your feature request related to a problem please describe nation home in other dimensions does not currently work unless a home is set in that dimension describe the solution you d like one nation home per dimension with the player being able to specify which home to go to usage would be n home describe alternatives you ve considered multiple nation homes that can be set max per nation or an increasing amount based on number of players in nation | 1 |
429,013 | 12,419,801,466 | IssuesEvent | 2020-05-23 08:27:41 | CSBiology/CSBAnnotatorAddIn | https://api.github.com/repos/CSBiology/CSBAnnotatorAddIn | opened | Improve FilePicker functionality | Area: Excel Area: Frontend Priority: Medium Status: In Progress Type: Enhancement | Necessary features:
- [ ] Fill excel cells with file name(s)
- [ ] make file names from list selectable -> insert only selected
- [ ] fill column range with file names, extend range when file name list is larger than range | 1.0 | Improve FilePicker functionality - Necessary features:
- [ ] Fill excel cells with file name(s)
- [ ] make file names from list selectable -> insert only selected
- [ ] fill column range with file names, extend range when file name list is larger than range | priority | improve filepicker functionality necessary features fill excel cells with file name s make file names from list selectable insert only selected fill column range with file names extend range when file name list is larger than range | 1 |
480,110 | 13,823,492,153 | IssuesEvent | 2020-10-13 07:05:21 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Optimise code in onboarding_fragment.xml | Hacktoberfest Priority: Nice-to-have SLoP 2020 SLoP-5 Status: Not started Type: Improvement good first issue | This issue is tracking the optimisation work of `onboarding_fragment.xml`
1. Make sure that the layout uses `ConstraintLayout` and there is minimal nesting of views. If this view already uses `ConstraintLayout` then it's fine.
2. Make sure that all textviews use the style from `styles.xml` file like `Heading1`, `TextFieldLabel`, `Body`, etc. whichever suits best.
While making the above changes the UI should not get affected and the end result should still be same as before (or an optimised version of UI).
If there are multiple files with the same name, update them all.
## References
Mobile Portrait: https://xd.adobe.com/view/e8aa4198-3940-47f9-514a-f41cc54457f6-9e9b/grid/
Mobile Landscape: https://xd.adobe.com/view/ee9e607b-dbd6-4372-48dc-b687d32af3da-98af/grid/
Tablet: https://xd.adobe.com/view/d405de00-a871-4f0f-73a0-f8acef30349b-a234/grid/
Wiki: https://github.com/oppia/oppia-android/wiki/Working-on-UI | 1.0 | Optimise code in onboarding_fragment.xml - This issue is tracking the optimisation work of `onboarding_fragment.xml`
1. Make sure that the layout uses `ConstraintLayout` and there is minimal nesting of views. If this view already uses `ConstraintLayout` then it's fine.
2. Make sure that all textviews use the style from `styles.xml` file like `Heading1`, `TextFieldLabel`, `Body`, etc. whichever suits best.
While making the above changes the UI should not get affected and the end result should still be same as before (or an optimised version of UI).
If there are multiple files with the same name, update them all.
## References
Mobile Portrait: https://xd.adobe.com/view/e8aa4198-3940-47f9-514a-f41cc54457f6-9e9b/grid/
Mobile Landscape: https://xd.adobe.com/view/ee9e607b-dbd6-4372-48dc-b687d32af3da-98af/grid/
Tablet: https://xd.adobe.com/view/d405de00-a871-4f0f-73a0-f8acef30349b-a234/grid/
Wiki: https://github.com/oppia/oppia-android/wiki/Working-on-UI | priority | optimise code in onboarding fragment xml this issue is tracking the optimisation work of onboarding fragment xml make sure that the layout uses constraintlayout and there is minimal nesting of views if this view already uses constraintlayout then it s fine make sure that all textviews use the style from styles xml file like textfieldlabel body etc whichever suits best while making the above changes the ui should not get affected and the end result should still be same as before or an optimised version of ui if there are multiple files with the same name update them all references mobile portrait mobile landscape tablet wiki | 1 |
136,088 | 30,472,868,873 | IssuesEvent | 2023-07-17 14:39:15 | eclipse-theia/theia | https://api.github.com/repos/eclipse-theia/theia | closed | @theia/vscode-builtin-debug-server-ready@0.2.1 does not work | bug help wanted vscode :thinking: needs more info | ### Description
This extension is intended to open a browser window once output meets a specific pattern.
But it doesn't work and there are no any errors on console.
https://www.npmjs.com/package/@theia/vscode-builtin-debug-server-ready | 1.0 | @theia/vscode-builtin-debug-server-ready@0.2.1 does not work - ### Description
This extension is intended to open a browser window once output meets a specific pattern.
But it doesn't work and there are no any errors on console.
https://www.npmjs.com/package/@theia/vscode-builtin-debug-server-ready | non_priority | theia vscode builtin debug server ready does not work description this extension is intended to open a browser window once output meets a specific pattern but it doesn t work and there are no any errors on console | 0 |
158,020 | 13,725,339,036 | IssuesEvent | 2020-10-03 18:05:14 | Interacao-Humano-Computador/2020.1-Prefeitura-Municipal-de-Arco-Iris | https://api.github.com/repos/Interacao-Humano-Computador/2020.1-Prefeitura-Municipal-de-Arco-Iris | closed | Cronograma | correção documentation | **Descrição**
<!-- Descreva de forma clara e concisa qual é o objetivo dessa issue -->
Realizar alterações no cronograma do planejamento.
**Tasks**
<!-- Descreva a seguir que tarefas serão realizadas nessa issue -->
- [x] Detalhar mais o cronograma.
**Informações Adicionais**
<!-- Descreva a seguir qualquer informação adicional que julgar relevante -->
Não esquecer de subir o cronograma com uma resolução melhor. | 1.0 | Cronograma - **Descrição**
<!-- Descreva de forma clara e concisa qual é o objetivo dessa issue -->
Realizar alterações no cronograma do planejamento.
**Tasks**
<!-- Descreva a seguir que tarefas serão realizadas nessa issue -->
- [x] Detalhar mais o cronograma.
**Informações Adicionais**
<!-- Descreva a seguir qualquer informação adicional que julgar relevante -->
Não esquecer de subir o cronograma com uma resolução melhor. | non_priority | cronograma descrição realizar alterações no cronograma do planejamento tasks detalhar mais o cronograma informações adicionais não esquecer de subir o cronograma com uma resolução melhor | 0 |
703,884 | 24,176,291,133 | IssuesEvent | 2022-09-23 02:24:43 | apache/hudi | https://api.github.com/repos/apache/hudi | closed | [SUPPORT] Failed to build hudi 0.12.0 with spark 3.2.2 | priority:critical spark | **Describe the problem you faced**
Failed to build hudi 0.12.0 with spark 3.2.2
**To Reproduce**
Steps to reproduce the behavior:
```shell
mvn clean package -DskipTests -Dscala-2.12 -Dspark3.2 -Dspark32.version=3.2.2
```
**Environment Description**
* Hudi version : 0.12.0
* Spark version : 3.2.2
**Stacktrace**
```
[INFO] --- scala-maven-plugin:3.3.1:compile (scala-compile-first) @ hudi-spark3.2.x_2.12 ---
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/java:-1: info: compiling
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/target/generated-sources/antlr4:-1: info: compiling
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala:-1: info: compiling
[INFO] Compiling 32 source files to /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/target/classes at 1662601520623
[WARNING] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala/org/apache/hudi/Spark32HoodieFileScanRDD.scala:21: warning: imported `HoodieUnsafeRDD' is permanently hidden by definition of trait HoodieUnsafeRDD in package hudi
[WARNING] import org.apache.hudi.HoodieUnsafeRDD
[WARNING] ^
[ERROR] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala/org/apache/spark/sql/execution/datasources/Spark32NestedSchemaPruning.scala:92: error: not enough arguments for method apply: (schema: org.apache.spark.sql.types.StructType, output: org.apache.spark.sql.catalyst.expressions.AttributeSet)org.apache.spark.sql.catalyst.expressions.ProjectionOverSchema in object ProjectionOverSchema.
[ERROR] Unspecified value parameter output.
[ERROR] val projectionOverSchema = ProjectionOverSchema(prunedDataSchema)
[ERROR] ^
[WARNING] one warning found
[ERROR] one error found
```
ProjectionOverSchema.scala:
```scala
//spark 3.2.1
case class ProjectionOverSchema(schema: StructType) {}
//spark 3.2.2 and spark 3.3.0
case class ProjectionOverSchema(schema: StructType, output: AttributeSet) {}
```
Spark32NestedSchemaPruning.scala:
```scala
if (countLeaves(dataSchema) > countLeaves(prunedDataSchema)) {
val prunedRelation = outputRelationBuilder(prunedDataSchema)
val projectionOverSchema = ProjectionOverSchema(prunedDataSchema)
Some(buildNewProjection(projects, normalizedProjects, normalizedFilters,
prunedRelation, projectionOverSchema))
}
```
Spark33NestedSchemaPruning.scala:
```scala
if (countLeaves(dataSchema) > countLeaves(prunedDataSchema)) {
val prunedRelation = outputRelationBuilder(prunedDataSchema)
val projectionOverSchema = ProjectionOverSchema(prunedDataSchema,AttributeSet(output))
Some(buildNewProjection(projects, normalizedProjects, normalizedFilters,
prunedRelation, projectionOverSchema))
}
```
In spark 3.2.1, ProjectionOverSchema has a argument. But in spark 3.2.2 and spark 3.3.0, ProjectionOverSchema has 2 arguments. Spark32NestedSchemaPruning.scala is not compatible with spark 3.2.2. | 1.0 | [SUPPORT] Failed to build hudi 0.12.0 with spark 3.2.2 - **Describe the problem you faced**
Failed to build hudi 0.12.0 with spark 3.2.2
**To Reproduce**
Steps to reproduce the behavior:
```shell
mvn clean package -DskipTests -Dscala-2.12 -Dspark3.2 -Dspark32.version=3.2.2
```
**Environment Description**
* Hudi version : 0.12.0
* Spark version : 3.2.2
**Stacktrace**
```
[INFO] --- scala-maven-plugin:3.3.1:compile (scala-compile-first) @ hudi-spark3.2.x_2.12 ---
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/java:-1: info: compiling
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/target/generated-sources/antlr4:-1: info: compiling
[INFO] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala:-1: info: compiling
[INFO] Compiling 32 source files to /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/target/classes at 1662601520623
[WARNING] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala/org/apache/hudi/Spark32HoodieFileScanRDD.scala:21: warning: imported `HoodieUnsafeRDD' is permanently hidden by definition of trait HoodieUnsafeRDD in package hudi
[WARNING] import org.apache.hudi.HoodieUnsafeRDD
[WARNING] ^
[ERROR] /root/hudi-release-0.12.0/hudi-spark-datasource/hudi-spark3.2.x/src/main/scala/org/apache/spark/sql/execution/datasources/Spark32NestedSchemaPruning.scala:92: error: not enough arguments for method apply: (schema: org.apache.spark.sql.types.StructType, output: org.apache.spark.sql.catalyst.expressions.AttributeSet)org.apache.spark.sql.catalyst.expressions.ProjectionOverSchema in object ProjectionOverSchema.
[ERROR] Unspecified value parameter output.
[ERROR] val projectionOverSchema = ProjectionOverSchema(prunedDataSchema)
[ERROR] ^
[WARNING] one warning found
[ERROR] one error found
```
ProjectionOverSchema.scala:
```scala
//spark 3.2.1
case class ProjectionOverSchema(schema: StructType) {}
//spark 3.2.2 and spark 3.3.0
case class ProjectionOverSchema(schema: StructType, output: AttributeSet) {}
```
Spark32NestedSchemaPruning.scala:
```scala
if (countLeaves(dataSchema) > countLeaves(prunedDataSchema)) {
val prunedRelation = outputRelationBuilder(prunedDataSchema)
val projectionOverSchema = ProjectionOverSchema(prunedDataSchema)
Some(buildNewProjection(projects, normalizedProjects, normalizedFilters,
prunedRelation, projectionOverSchema))
}
```
Spark33NestedSchemaPruning.scala:
```scala
if (countLeaves(dataSchema) > countLeaves(prunedDataSchema)) {
val prunedRelation = outputRelationBuilder(prunedDataSchema)
val projectionOverSchema = ProjectionOverSchema(prunedDataSchema,AttributeSet(output))
Some(buildNewProjection(projects, normalizedProjects, normalizedFilters,
prunedRelation, projectionOverSchema))
}
```
In spark 3.2.1, ProjectionOverSchema has a argument. But in spark 3.2.2 and spark 3.3.0, ProjectionOverSchema has 2 arguments. Spark32NestedSchemaPruning.scala is not compatible with spark 3.2.2. | priority | failed to build hudi with spark describe the problem you faced failed to build hudi with spark to reproduce steps to reproduce the behavior shell mvn clean package dskiptests dscala version environment description hudi version spark version stacktrace scala maven plugin compile scala compile first hudi x root hudi release hudi spark datasource hudi x src main java info compiling root hudi release hudi spark datasource hudi x target generated sources info compiling root hudi release hudi spark datasource hudi x src main scala info compiling compiling source files to root hudi release hudi spark datasource hudi x target classes at root hudi release hudi spark datasource hudi x src main scala org apache hudi scala warning imported hoodieunsaferdd is permanently hidden by definition of trait hoodieunsaferdd in package hudi import org apache hudi hoodieunsaferdd root hudi release hudi spark datasource hudi x src main scala org apache spark sql execution datasources scala error not enough arguments for method apply schema org apache spark sql types structtype output org apache spark sql catalyst expressions attributeset org apache spark sql catalyst expressions projectionoverschema in object projectionoverschema unspecified value parameter output val projectionoverschema projectionoverschema pruneddataschema one warning found one error found projectionoverschema scala scala spark case class projectionoverschema schema structtype spark and spark case class projectionoverschema schema structtype output attributeset scala scala if countleaves dataschema countleaves pruneddataschema val prunedrelation outputrelationbuilder pruneddataschema val projectionoverschema projectionoverschema pruneddataschema some buildnewprojection projects normalizedprojects normalizedfilters prunedrelation projectionoverschema scala scala if countleaves dataschema countleaves pruneddataschema val prunedrelation outputrelationbuilder pruneddataschema val projectionoverschema projectionoverschema pruneddataschema attributeset output some buildnewprojection projects normalizedprojects normalizedfilters prunedrelation projectionoverschema in spark projectionoverschema has a argument but in spark and spark projectionoverschema has arguments scala is not compatible with spark | 1 |
742,822 | 25,871,426,294 | IssuesEvent | 2022-12-14 03:07:01 | jerichosy/CSSWENG-Team-3 | https://api.github.com/repos/jerichosy/CSSWENG-Team-3 | opened | Multiple Submissions of Records | bug stack - frontend low priority | **Summary**
There is no error when submitting multiple times in the cashier
**Steps to Produce:**
1. Go to Review page
2. Click on Submit
3. Click on Submit again
4. After submitting, click submit again
5. Click on submit again"
**Expected Results:**
There should be an error stating that the user can only submit once a day
**Actual Results:**
There was no error and the user can submit multiple times

See 11-SubmitReport, Iteration 1-5 for more details | 1.0 | Multiple Submissions of Records - **Summary**
There is no error when submitting multiple times in the cashier
**Steps to Produce:**
1. Go to Review page
2. Click on Submit
3. Click on Submit again
4. After submitting, click submit again
5. Click on submit again"
**Expected Results:**
There should be an error stating that the user can only submit once a day
**Actual Results:**
There was no error and the user can submit multiple times

See 11-SubmitReport, Iteration 1-5 for more details | priority | multiple submissions of records summary there is no error when submitting multiple times in the cashier steps to produce go to review page click on submit click on submit again after submitting click submit again click on submit again expected results there should be an error stating that the user can only submit once a day actual results there was no error and the user can submit multiple times see submitreport iteration for more details | 1 |
449,465 | 31,846,685,972 | IssuesEvent | 2023-09-14 20:31:36 | Icinga/icinga-powershell-plugins | https://api.github.com/repos/Icinga/icinga-powershell-plugins | closed | Invoke-IcingaCheckUsedPartitionSpace: Threshold problem in default mode (free space) | documentation | Hello,
we have a problem with `Invoke-IcingaCheckUsedPartitionSpace` when we try to check the remaining disk space (mode "free space").
Even without the argument `-CheckUsedSpace` the check currently behaves like mode "used space".
(warning: <=10%, critical: <=5%)

Mode "used space" works fine.
Used versions:

| 1.0 | Invoke-IcingaCheckUsedPartitionSpace: Threshold problem in default mode (free space) - Hello,
we have a problem with `Invoke-IcingaCheckUsedPartitionSpace` when we try to check the remaining disk space (mode "free space").
Even without the argument `-CheckUsedSpace` the check currently behaves like mode "used space".
(warning: <=10%, critical: <=5%)

Mode "used space" works fine.
Used versions:

| non_priority | invoke icingacheckusedpartitionspace threshold problem in default mode free space hello we have a problem with invoke icingacheckusedpartitionspace when we try to check the remaining disk space mode free space even without the argument checkusedspace the check currently behaves like mode used space warning critical mode used space works fine used versions | 0 |
47,320 | 2,974,727,692 | IssuesEvent | 2015-07-15 03:45:21 | cjfields/redmine-test | https://api.github.com/repos/cjfields/redmine-test | opened | ID of match features in bp_search2gff.pl | Category: Bio::Search/Bio::SearchIO Component: URL Priority: Normal Status: New Tracker: Bug | ---
Author Name: **Eric Just** (Eric Just)
Original Redmine Issue: 2376, https://redmine.open-bio.org/issues/2376
Original Date: 2007-10-05
Original Assignee: Bioperl Guts
---
The ID that gets generated for a match feature when the --match
option is set can be problematic. The ID is set to the ID of the query sequence. This
can be problematic if you are representing the query sequence and the
blast hit in the same gff file. When using the resultant gff file for
loading into Chado, it also creates a problem if you have more than
one hit for a given query sequence, for example if you ran two
different analyses that each had a hit for a given query. Would it be
possible to have an option to create a unique ID for match features.
One suggestion could be to create an ID based on the ID of the query +
the id of the hit + the source
currently a hit of the query DDDB0XXXX to DDB00000902 with the source blasntn10052007 would be output like
DDB00000902 blasntn10052007 match 5006 55498 462 + . ID=DDB0XXXXX
I propose a unique ID, for example:
DDB00000902 blasntn10052007 match 5006 55498 462 + . ID=DDB0XXXXX.DDB00000902.blasntn10052007
| 1.0 | ID of match features in bp_search2gff.pl - ---
Author Name: **Eric Just** (Eric Just)
Original Redmine Issue: 2376, https://redmine.open-bio.org/issues/2376
Original Date: 2007-10-05
Original Assignee: Bioperl Guts
---
The ID that gets generated for a match feature when the --match
option is set can be problematic. The ID is set to the ID of the query sequence. This
can be problematic if you are representing the query sequence and the
blast hit in the same gff file. When using the resultant gff file for
loading into Chado, it also creates a problem if you have more than
one hit for a given query sequence, for example if you ran two
different analyses that each had a hit for a given query. Would it be
possible to have an option to create a unique ID for match features.
One suggestion could be to create an ID based on the ID of the query +
the id of the hit + the source
currently a hit of the query DDDB0XXXX to DDB00000902 with the source blasntn10052007 would be output like
DDB00000902 blasntn10052007 match 5006 55498 462 + . ID=DDB0XXXXX
I propose a unique ID, for example:
DDB00000902 blasntn10052007 match 5006 55498 462 + . ID=DDB0XXXXX.DDB00000902.blasntn10052007
| priority | id of match features in bp pl author name eric just eric just original redmine issue original date original assignee bioperl guts the id that gets generated for a match feature when the match option is set can be problematic the id is set to the id of the query sequence this can be problematic if you are representing the query sequence and the blast hit in the same gff file when using the resultant gff file for loading into chado it also creates a problem if you have more than one hit for a given query sequence for example if you ran two different analyses that each had a hit for a given query would it be possible to have an option to create a unique id for match features one suggestion could be to create an id based on the id of the query the id of the hit the source currently a hit of the query to with the source would be output like match id i propose a unique id for example match id | 1 |
225,977 | 17,294,462,688 | IssuesEvent | 2021-07-25 12:44:14 | Kawboy442/food_share_web | https://api.github.com/repos/Kawboy442/food_share_web | closed | 修正: model層のプロパティの命名規則をローワーキャメルケースに統一する | documentation priority: MIddle | カリキュラムの名残で、model層のプロパティ名がスネークケースだったりローワーキャメルケースだったりするので、ローワーキャメルケースに統一する。 | 1.0 | 修正: model層のプロパティの命名規則をローワーキャメルケースに統一する - カリキュラムの名残で、model層のプロパティ名がスネークケースだったりローワーキャメルケースだったりするので、ローワーキャメルケースに統一する。 | non_priority | 修正 model層のプロパティの命名規則をローワーキャメルケースに統一する カリキュラムの名残で、model層のプロパティ名がスネークケースだったりローワーキャメルケースだったりするので、ローワーキャメルケースに統一する。 | 0 |
111,512 | 14,105,090,667 | IssuesEvent | 2020-11-06 12:57:57 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | opened | Pause and resume Blazor applications | design-proposal | ## Summary
Enable the ability to "pause" a blazor application and enable components and services within a Blazor application to persist their state when the application is being "shut down" and enable a mechanism to retrieve and restore that state afterwards.
## Motivation and goals
Today it is very hard to do prerendering without having the content flash on the page due to the fact that a regular component will render multiple times during an asynchronous code path like retrieving data from a service or datase and produce an intermediate render without content that will quickly replace the content on the rendered page to be followed by the actual content afterwards.
That experience is desirable when the application is not prerendering so as to be able to show content quickly to users, but it's jarring when prerendering is enabled. The guidance that we give our users is to compute all the state of the application ahead of time and pass it down to the root component as a parameter, with the idea of avoiding any asynchronous code paths when the state is available.
This has several drawbacks, like the fact that users need to extract complicated logic out of their blazor application to be run outside of the application context and avoid asynchronous code paths on their code which are very easy to introduce.
The idea is that when we prerender the application we "pause" the application and "persist" the state of components that opt-in into it. We marshal that state to the client and when the app "resumes" we make that state available for components that opted-in on to the feature.
This enables application authors to preserve all the state of their components, like data that was retrieved from a database and ensure that they can produce a deterministic render that is equivalent to content they prerendered on the server.
In addition to that, this same mechanism can be used within server-side Blazor to empower customers to "pause" Blazor server applications and resume them at a later point, for example when there is no activity within a circuit for a given period of time or when a server is experiencing heavy load, enabling better resource management.
## In scope
* Provide APIs to preserve and restore component state when an application is "paused" and "resumed".
* Marshall the application state to the client and pass it back to the app when it "restarts".
## Out of scope
* Serializing random object graphs.
* Automatic state management/restore.
* Server side migration scenarios.
* While we call out is possible with this API, it is considered a stretch goals.
## Risks / unknowns
TBD
## Examples
* Marshalling state retrieved from a service or database.
* Marshalling authentication state on the client. | 1.0 | Pause and resume Blazor applications - ## Summary
Enable the ability to "pause" a blazor application and enable components and services within a Blazor application to persist their state when the application is being "shut down" and enable a mechanism to retrieve and restore that state afterwards.
## Motivation and goals
Today it is very hard to do prerendering without having the content flash on the page due to the fact that a regular component will render multiple times during an asynchronous code path like retrieving data from a service or datase and produce an intermediate render without content that will quickly replace the content on the rendered page to be followed by the actual content afterwards.
That experience is desirable when the application is not prerendering so as to be able to show content quickly to users, but it's jarring when prerendering is enabled. The guidance that we give our users is to compute all the state of the application ahead of time and pass it down to the root component as a parameter, with the idea of avoiding any asynchronous code paths when the state is available.
This has several drawbacks, like the fact that users need to extract complicated logic out of their blazor application to be run outside of the application context and avoid asynchronous code paths on their code which are very easy to introduce.
The idea is that when we prerender the application we "pause" the application and "persist" the state of components that opt-in into it. We marshal that state to the client and when the app "resumes" we make that state available for components that opted-in on to the feature.
This enables application authors to preserve all the state of their components, like data that was retrieved from a database and ensure that they can produce a deterministic render that is equivalent to content they prerendered on the server.
In addition to that, this same mechanism can be used within server-side Blazor to empower customers to "pause" Blazor server applications and resume them at a later point, for example when there is no activity within a circuit for a given period of time or when a server is experiencing heavy load, enabling better resource management.
## In scope
* Provide APIs to preserve and restore component state when an application is "paused" and "resumed".
* Marshall the application state to the client and pass it back to the app when it "restarts".
## Out of scope
* Serializing random object graphs.
* Automatic state management/restore.
* Server side migration scenarios.
* While we call out is possible with this API, it is considered a stretch goals.
## Risks / unknowns
TBD
## Examples
* Marshalling state retrieved from a service or database.
* Marshalling authentication state on the client. | non_priority | pause and resume blazor applications summary enable the ability to pause a blazor application and enable components and services within a blazor application to persist their state when the application is being shut down and enable a mechanism to retrieve and restore that state afterwards motivation and goals today it is very hard to do prerendering without having the content flash on the page due to the fact that a regular component will render multiple times during an asynchronous code path like retrieving data from a service or datase and produce an intermediate render without content that will quickly replace the content on the rendered page to be followed by the actual content afterwards that experience is desirable when the application is not prerendering so as to be able to show content quickly to users but it s jarring when prerendering is enabled the guidance that we give our users is to compute all the state of the application ahead of time and pass it down to the root component as a parameter with the idea of avoiding any asynchronous code paths when the state is available this has several drawbacks like the fact that users need to extract complicated logic out of their blazor application to be run outside of the application context and avoid asynchronous code paths on their code which are very easy to introduce the idea is that when we prerender the application we pause the application and persist the state of components that opt in into it we marshal that state to the client and when the app resumes we make that state available for components that opted in on to the feature this enables application authors to preserve all the state of their components like data that was retrieved from a database and ensure that they can produce a deterministic render that is equivalent to content they prerendered on the server in addition to that this same mechanism can be used within server side blazor to empower customers to pause blazor server applications and resume them at a later point for example when there is no activity within a circuit for a given period of time or when a server is experiencing heavy load enabling better resource management in scope provide apis to preserve and restore component state when an application is paused and resumed marshall the application state to the client and pass it back to the app when it restarts out of scope serializing random object graphs automatic state management restore server side migration scenarios while we call out is possible with this api it is considered a stretch goals risks unknowns tbd examples marshalling state retrieved from a service or database marshalling authentication state on the client | 0 |
75,728 | 3,474,659,883 | IssuesEvent | 2015-12-25 01:27:09 | Atrae/geekle | https://api.github.com/repos/Atrae/geekle | closed | 問題追加時に連続で問題を追加することが多いのでreference_idと本のタイトルを引き継ぎたい | 1st priority | - [x] クッキーで付与する仕組みを作る
- [x] クッキーから受け取ってデータを入れる仕組みを作る | 1.0 | 問題追加時に連続で問題を追加することが多いのでreference_idと本のタイトルを引き継ぎたい - - [x] クッキーで付与する仕組みを作る
- [x] クッキーから受け取ってデータを入れる仕組みを作る | priority | 問題追加時に連続で問題を追加することが多いのでreference idと本のタイトルを引き継ぎたい クッキーで付与する仕組みを作る クッキーから受け取ってデータを入れる仕組みを作る | 1 |
88,803 | 3,785,677,008 | IssuesEvent | 2016-03-20 17:26:06 | TranslationWMcs435/TranslationWMcs435 | https://api.github.com/repos/TranslationWMcs435/TranslationWMcs435 | closed | Add interface for managing the output files | High Priority | Add an interface to manage the files outputted by the translation engine. This will make the implementations of the translators cleaner and will create a structure that will be easy to extend to for other UI testing frameworks. | 1.0 | Add interface for managing the output files - Add an interface to manage the files outputted by the translation engine. This will make the implementations of the translators cleaner and will create a structure that will be easy to extend to for other UI testing frameworks. | priority | add interface for managing the output files add an interface to manage the files outputted by the translation engine this will make the implementations of the translators cleaner and will create a structure that will be easy to extend to for other ui testing frameworks | 1 |
79,022 | 3,520,015,040 | IssuesEvent | 2016-01-12 19:05:18 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Uploading zip files stopped working | Component: API Priority: Critical Status: QA | @scolapasta is aware of this but as of a42f628 we can no longer upload files via SWORD.
(Reverting that commit locally fixed the problem.)
To reproduce, for example:
curl -s --insecure --data-binary @scripts/search/data/binary/trees.zip -H 'Content-Disposition: filename=trees.zip' -H 'Content-Type: application/zip' -H 'Packaging: http://purl.org/net/sword/package/SimpleZip' -u 22416513-1a1f-4189-a5b2-3e56c80aec72: https://localhost:8181/dvn/api/data-deposit/v1.1/swordv2/edit-media/study/doi:10.5072/FK2/W7N2EU
Stacktrace:
```
[2015-04-08T21:44:59.722-0400] [glassfish 4.1] [INFO] [] [edu.harvard.iq.dataverse.ingest.IngestServiceBean] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899722] [levelValue: 800] [[
ZipEntry, file: trees/trees.png]]
[2015-04-08T21:44:59.728-0400] [glassfish 4.1] [INFO] [] [edu.harvard.iq.dataverse.ingest.IngestServiceBean] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899728] [levelValue: 800] [[
setting category to trees]]
[2015-04-08T21:44:59.729-0400] [glassfish 4.1] [WARNING] [AS-EJB-00056] [javax.enterprise.ejb.container] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899729] [levelValue: 900] [[
A system exception occurred during an invocation on EJB IngestServiceBean, method: public java.util.List edu.harvard.iq.dataverse.ingest.IngestServiceBean.createDataFiles(edu.harvard.iq.dataverse.DatasetVersion,java.io.InputStream,java.lang.String,java.lang.String) throws java.io.IOException]]
[2015-04-08T21:44:59.729-0400] [glassfish 4.1] [WARNING] [] [javax.enterprise.ejb.container] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899729] [levelValue: 900] [[
javax.ejb.EJBException
at com.sun.ejb.containers.EJBContainerTransactionManager.processSystemException(EJBContainerTransactionManager.java:748)
at com.sun.ejb.containers.EJBContainerTransactionManager.completeNewTx(EJBContainerTransactionManager.java:698)
at com.sun.ejb.containers.EJBContainerTransactionManager.postInvokeTx(EJBContainerTransactionManager.java:503)
at com.sun.ejb.containers.BaseContainer.postInvokeTx(BaseContainer.java:4566)
at com.sun.ejb.containers.BaseContainer.postInvoke(BaseContainer.java:2074)
at com.sun.ejb.containers.BaseContainer.postInvoke(BaseContainer.java:2044)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandler.invoke(EJBLocalObjectInvocationHandler.java:220)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandlerDelegate.invoke(EJBLocalObjectInvocationHandlerDelegate.java:88)
at com.sun.proxy.$Proxy462.createDataFiles(Unknown Source)
at edu.harvard.iq.dataverse.ingest.__EJB31_Generated__IngestServiceBean__Intf____Bean__.createDataFiles(Unknown Source)
at edu.harvard.iq.dataverse.api.datadeposit.MediaResourceManagerImpl.replaceOrAddFiles(MediaResourceManagerImpl.java:249)
at edu.harvard.iq.dataverse.api.datadeposit.MediaResourceManagerImpl.addResource(MediaResourceManagerImpl.java:181)
at org.swordapp.server.MediaResourceAPI.post(MediaResourceAPI.java:272)
at edu.harvard.iq.dataverse.api.datadeposit.SWORDv2MediaResourceServlet.doPost(SWORDv2MediaResourceServlet.java:44)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.apache.catalina.core.StandardWrapper.service(StandardWrapper.java:1682)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:344)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.glassfish.tyrus.servlet.TyrusServletFilter.doFilter(TyrusServletFilter.java:295)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:256)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.ocpsoft.rewrite.servlet.RewriteFilter.doFilter(RewriteFilter.java:205)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:256)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:316)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:160)
at org.apache.catalina.core.StandardPipeline.doInvoke(StandardPipeline.java:734)
at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:673)
at com.sun.enterprise.web.WebPipeline.invoke(WebPipeline.java:99)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:174)
at org.apache.catalina.connector.CoyoteAdapter.doService(CoyoteAdapter.java:415)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:282)
at com.sun.enterprise.v3.services.impl.ContainerMapper$HttpHandlerCallable.call(ContainerMapper.java:459)
at com.sun.enterprise.v3.services.impl.ContainerMapper.service(ContainerMapper.java:167)
at org.glassfish.grizzly.http.server.HttpHandler.runService(HttpHandler.java:201)
at org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:175)
at org.glassfish.grizzly.http.server.HttpServerFilter.handleRead(HttpServerFilter.java:235)
at org.glassfish.grizzly.filterchain.ExecutorResolver$9.execute(ExecutorResolver.java:119)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeFilter(DefaultFilterChain.java:284)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeChainPart(DefaultFilterChain.java:201)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.execute(DefaultFilterChain.java:133)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.process(DefaultFilterChain.java:112)
at org.glassfish.grizzly.ProcessorExecutor.execute(ProcessorExecutor.java:77)
at org.glassfish.grizzly.nio.transport.TCPNIOTransport.fireIOEvent(TCPNIOTransport.java:561)
at org.glassfish.grizzly.strategies.AbstractIOStrategy.fireIOEvent(AbstractIOStrategy.java:112)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.run0(WorkerThreadIOStrategy.java:117)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.access$100(WorkerThreadIOStrategy.java:56)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy$WorkerThreadRunnable.run(WorkerThreadIOStrategy.java:137)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:565)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:545)
at java.lang.Thread.run(Thread.java:744)
Caused by: java.lang.NullPointerException
at edu.harvard.iq.dataverse.DataFile.getLatestFileMetadata(DataFile.java:291)
at edu.harvard.iq.dataverse.DataFile.getFileMetadata(DataFile.java:282)
at edu.harvard.iq.dataverse.ingest.IngestServiceBean.createDataFiles(IngestServiceBean.java:420)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.glassfish.ejb.security.application.EJBSecurityManager.runMethod(EJBSecurityManager.java:1081)
at org.glassfish.ejb.security.application.EJBSecurityManager.invoke(EJBSecurityManager.java:1153)
at com.sun.ejb.containers.BaseContainer.invokeBeanMethod(BaseContainer.java:4786)
at com.sun.ejb.EjbInvocation.invokeBeanMethod(EjbInvocation.java:656)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.EjbInvocation.proceed(EjbInvocation.java:608)
at org.jboss.weld.ejb.AbstractEJBRequestScopeActivationInterceptor.aroundInvoke(AbstractEJBRequestScopeActivationInterceptor.java:46)
at org.jboss.weld.ejb.SessionBeanInterceptor.aroundInvoke(SessionBeanInterceptor.java:52)
at sun.reflect.GeneratedMethodAccessor92.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.sun.ejb.containers.interceptors.AroundInvokeInterceptor.intercept(InterceptorManager.java:883)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.EjbInvocation.proceed(EjbInvocation.java:608)
at com.sun.ejb.containers.interceptors.SystemInterceptorProxy.doCall(SystemInterceptorProxy.java:163)
at com.sun.ejb.containers.interceptors.SystemInterceptorProxy.aroundInvoke(SystemInterceptorProxy.java:140)
at sun.reflect.GeneratedMethodAccessor102.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.sun.ejb.containers.interceptors.AroundInvokeInterceptor.intercept(InterceptorManager.java:883)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.containers.interceptors.InterceptorManager.intercept(InterceptorManager.java:369)
at com.sun.ejb.containers.BaseContainer.__intercept(BaseContainer.java:4758)
at com.sun.ejb.containers.BaseContainer.intercept(BaseContainer.java:4746)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandler.invoke(EJBLocalObjectInvocationHandler.java:212)
... 45 more
]]
``` | 1.0 | Uploading zip files stopped working - @scolapasta is aware of this but as of a42f628 we can no longer upload files via SWORD.
(Reverting that commit locally fixed the problem.)
To reproduce, for example:
curl -s --insecure --data-binary @scripts/search/data/binary/trees.zip -H 'Content-Disposition: filename=trees.zip' -H 'Content-Type: application/zip' -H 'Packaging: http://purl.org/net/sword/package/SimpleZip' -u 22416513-1a1f-4189-a5b2-3e56c80aec72: https://localhost:8181/dvn/api/data-deposit/v1.1/swordv2/edit-media/study/doi:10.5072/FK2/W7N2EU
Stacktrace:
```
[2015-04-08T21:44:59.722-0400] [glassfish 4.1] [INFO] [] [edu.harvard.iq.dataverse.ingest.IngestServiceBean] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899722] [levelValue: 800] [[
ZipEntry, file: trees/trees.png]]
[2015-04-08T21:44:59.728-0400] [glassfish 4.1] [INFO] [] [edu.harvard.iq.dataverse.ingest.IngestServiceBean] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899728] [levelValue: 800] [[
setting category to trees]]
[2015-04-08T21:44:59.729-0400] [glassfish 4.1] [WARNING] [AS-EJB-00056] [javax.enterprise.ejb.container] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899729] [levelValue: 900] [[
A system exception occurred during an invocation on EJB IngestServiceBean, method: public java.util.List edu.harvard.iq.dataverse.ingest.IngestServiceBean.createDataFiles(edu.harvard.iq.dataverse.DatasetVersion,java.io.InputStream,java.lang.String,java.lang.String) throws java.io.IOException]]
[2015-04-08T21:44:59.729-0400] [glassfish 4.1] [WARNING] [] [javax.enterprise.ejb.container] [tid: _ThreadID=34 _ThreadName=http-listener-2(3)] [timeMillis: 1428543899729] [levelValue: 900] [[
javax.ejb.EJBException
at com.sun.ejb.containers.EJBContainerTransactionManager.processSystemException(EJBContainerTransactionManager.java:748)
at com.sun.ejb.containers.EJBContainerTransactionManager.completeNewTx(EJBContainerTransactionManager.java:698)
at com.sun.ejb.containers.EJBContainerTransactionManager.postInvokeTx(EJBContainerTransactionManager.java:503)
at com.sun.ejb.containers.BaseContainer.postInvokeTx(BaseContainer.java:4566)
at com.sun.ejb.containers.BaseContainer.postInvoke(BaseContainer.java:2074)
at com.sun.ejb.containers.BaseContainer.postInvoke(BaseContainer.java:2044)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandler.invoke(EJBLocalObjectInvocationHandler.java:220)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandlerDelegate.invoke(EJBLocalObjectInvocationHandlerDelegate.java:88)
at com.sun.proxy.$Proxy462.createDataFiles(Unknown Source)
at edu.harvard.iq.dataverse.ingest.__EJB31_Generated__IngestServiceBean__Intf____Bean__.createDataFiles(Unknown Source)
at edu.harvard.iq.dataverse.api.datadeposit.MediaResourceManagerImpl.replaceOrAddFiles(MediaResourceManagerImpl.java:249)
at edu.harvard.iq.dataverse.api.datadeposit.MediaResourceManagerImpl.addResource(MediaResourceManagerImpl.java:181)
at org.swordapp.server.MediaResourceAPI.post(MediaResourceAPI.java:272)
at edu.harvard.iq.dataverse.api.datadeposit.SWORDv2MediaResourceServlet.doPost(SWORDv2MediaResourceServlet.java:44)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.apache.catalina.core.StandardWrapper.service(StandardWrapper.java:1682)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:344)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.glassfish.tyrus.servlet.TyrusServletFilter.doFilter(TyrusServletFilter.java:295)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:256)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.ocpsoft.rewrite.servlet.RewriteFilter.doFilter(RewriteFilter.java:205)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:256)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:214)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:316)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:160)
at org.apache.catalina.core.StandardPipeline.doInvoke(StandardPipeline.java:734)
at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:673)
at com.sun.enterprise.web.WebPipeline.invoke(WebPipeline.java:99)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:174)
at org.apache.catalina.connector.CoyoteAdapter.doService(CoyoteAdapter.java:415)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:282)
at com.sun.enterprise.v3.services.impl.ContainerMapper$HttpHandlerCallable.call(ContainerMapper.java:459)
at com.sun.enterprise.v3.services.impl.ContainerMapper.service(ContainerMapper.java:167)
at org.glassfish.grizzly.http.server.HttpHandler.runService(HttpHandler.java:201)
at org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:175)
at org.glassfish.grizzly.http.server.HttpServerFilter.handleRead(HttpServerFilter.java:235)
at org.glassfish.grizzly.filterchain.ExecutorResolver$9.execute(ExecutorResolver.java:119)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeFilter(DefaultFilterChain.java:284)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeChainPart(DefaultFilterChain.java:201)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.execute(DefaultFilterChain.java:133)
at org.glassfish.grizzly.filterchain.DefaultFilterChain.process(DefaultFilterChain.java:112)
at org.glassfish.grizzly.ProcessorExecutor.execute(ProcessorExecutor.java:77)
at org.glassfish.grizzly.nio.transport.TCPNIOTransport.fireIOEvent(TCPNIOTransport.java:561)
at org.glassfish.grizzly.strategies.AbstractIOStrategy.fireIOEvent(AbstractIOStrategy.java:112)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.run0(WorkerThreadIOStrategy.java:117)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.access$100(WorkerThreadIOStrategy.java:56)
at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy$WorkerThreadRunnable.run(WorkerThreadIOStrategy.java:137)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:565)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:545)
at java.lang.Thread.run(Thread.java:744)
Caused by: java.lang.NullPointerException
at edu.harvard.iq.dataverse.DataFile.getLatestFileMetadata(DataFile.java:291)
at edu.harvard.iq.dataverse.DataFile.getFileMetadata(DataFile.java:282)
at edu.harvard.iq.dataverse.ingest.IngestServiceBean.createDataFiles(IngestServiceBean.java:420)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.glassfish.ejb.security.application.EJBSecurityManager.runMethod(EJBSecurityManager.java:1081)
at org.glassfish.ejb.security.application.EJBSecurityManager.invoke(EJBSecurityManager.java:1153)
at com.sun.ejb.containers.BaseContainer.invokeBeanMethod(BaseContainer.java:4786)
at com.sun.ejb.EjbInvocation.invokeBeanMethod(EjbInvocation.java:656)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.EjbInvocation.proceed(EjbInvocation.java:608)
at org.jboss.weld.ejb.AbstractEJBRequestScopeActivationInterceptor.aroundInvoke(AbstractEJBRequestScopeActivationInterceptor.java:46)
at org.jboss.weld.ejb.SessionBeanInterceptor.aroundInvoke(SessionBeanInterceptor.java:52)
at sun.reflect.GeneratedMethodAccessor92.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.sun.ejb.containers.interceptors.AroundInvokeInterceptor.intercept(InterceptorManager.java:883)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.EjbInvocation.proceed(EjbInvocation.java:608)
at com.sun.ejb.containers.interceptors.SystemInterceptorProxy.doCall(SystemInterceptorProxy.java:163)
at com.sun.ejb.containers.interceptors.SystemInterceptorProxy.aroundInvoke(SystemInterceptorProxy.java:140)
at sun.reflect.GeneratedMethodAccessor102.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.sun.ejb.containers.interceptors.AroundInvokeInterceptor.intercept(InterceptorManager.java:883)
at com.sun.ejb.containers.interceptors.AroundInvokeChainImpl.invokeNext(InterceptorManager.java:822)
at com.sun.ejb.containers.interceptors.InterceptorManager.intercept(InterceptorManager.java:369)
at com.sun.ejb.containers.BaseContainer.__intercept(BaseContainer.java:4758)
at com.sun.ejb.containers.BaseContainer.intercept(BaseContainer.java:4746)
at com.sun.ejb.containers.EJBLocalObjectInvocationHandler.invoke(EJBLocalObjectInvocationHandler.java:212)
... 45 more
]]
``` | priority | uploading zip files stopped working scolapasta is aware of this but as of we can no longer upload files via sword reverting that commit locally fixed the problem to reproduce for example curl s insecure data binary scripts search data binary trees zip h content disposition filename trees zip h content type application zip h packaging u stacktrace zipentry file trees trees png setting category to trees a system exception occurred during an invocation on ejb ingestservicebean method public java util list edu harvard iq dataverse ingest ingestservicebean createdatafiles edu harvard iq dataverse datasetversion java io inputstream java lang string java lang string throws java io ioexception javax ejb ejbexception at com sun ejb containers ejbcontainertransactionmanager processsystemexception ejbcontainertransactionmanager java at com sun ejb containers ejbcontainertransactionmanager completenewtx ejbcontainertransactionmanager java at com sun ejb containers ejbcontainertransactionmanager postinvoketx ejbcontainertransactionmanager java at com sun ejb containers basecontainer postinvoketx basecontainer java at com sun ejb containers basecontainer postinvoke basecontainer java at com sun ejb containers basecontainer postinvoke basecontainer java at com sun ejb containers ejblocalobjectinvocationhandler invoke ejblocalobjectinvocationhandler java at com sun ejb containers ejblocalobjectinvocationhandlerdelegate invoke ejblocalobjectinvocationhandlerdelegate java at com sun proxy createdatafiles unknown source at edu harvard iq dataverse ingest generated ingestservicebean intf bean createdatafiles unknown source at edu harvard iq dataverse api datadeposit mediaresourcemanagerimpl replaceoraddfiles mediaresourcemanagerimpl java at edu harvard iq dataverse api datadeposit mediaresourcemanagerimpl addresource mediaresourcemanagerimpl java at org swordapp server mediaresourceapi post mediaresourceapi java at edu harvard iq dataverse api datadeposit dopost java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at org apache catalina core standardwrapper service standardwrapper java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org glassfish tyrus servlet tyrusservletfilter dofilter tyrusservletfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org ocpsoft rewrite servlet rewritefilter dofilter rewritefilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina core standardpipeline doinvoke standardpipeline java at org apache catalina core standardpipeline invoke standardpipeline java at com sun enterprise web webpipeline invoke webpipeline java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina connector coyoteadapter doservice coyoteadapter java at org apache catalina connector coyoteadapter service coyoteadapter java at com sun enterprise services impl containermapper httphandlercallable call containermapper java at com sun enterprise services impl containermapper service containermapper java at org glassfish grizzly http server httphandler runservice httphandler java at org glassfish grizzly http server httphandler dohandle httphandler java at org glassfish grizzly http server httpserverfilter handleread httpserverfilter java at org glassfish grizzly filterchain executorresolver execute executorresolver java at org glassfish grizzly filterchain defaultfilterchain executefilter defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain executechainpart defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain execute defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain process defaultfilterchain java at org glassfish grizzly processorexecutor execute processorexecutor java at org glassfish grizzly nio transport tcpniotransport fireioevent tcpniotransport java at org glassfish grizzly strategies abstractiostrategy fireioevent abstractiostrategy java at org glassfish grizzly strategies workerthreadiostrategy workerthreadiostrategy java at org glassfish grizzly strategies workerthreadiostrategy access workerthreadiostrategy java at org glassfish grizzly strategies workerthreadiostrategy workerthreadrunnable run workerthreadiostrategy java at org glassfish grizzly threadpool abstractthreadpool worker dowork abstractthreadpool java at org glassfish grizzly threadpool abstractthreadpool worker run abstractthreadpool java at java lang thread run thread java caused by java lang nullpointerexception at edu harvard iq dataverse datafile getlatestfilemetadata datafile java at edu harvard iq dataverse datafile getfilemetadata datafile java at edu harvard iq dataverse ingest ingestservicebean createdatafiles ingestservicebean java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org glassfish ejb security application ejbsecuritymanager runmethod ejbsecuritymanager java at org glassfish ejb security application ejbsecuritymanager invoke ejbsecuritymanager java at com sun ejb containers basecontainer invokebeanmethod basecontainer java at com sun ejb ejbinvocation invokebeanmethod ejbinvocation java at com sun ejb containers interceptors aroundinvokechainimpl invokenext interceptormanager java at com sun ejb ejbinvocation proceed ejbinvocation java at org jboss weld ejb abstractejbrequestscopeactivationinterceptor aroundinvoke abstractejbrequestscopeactivationinterceptor java at org jboss weld ejb sessionbeaninterceptor aroundinvoke sessionbeaninterceptor java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com sun ejb containers interceptors aroundinvokeinterceptor intercept interceptormanager java at com sun ejb containers interceptors aroundinvokechainimpl invokenext interceptormanager java at com sun ejb ejbinvocation proceed ejbinvocation java at com sun ejb containers interceptors systeminterceptorproxy docall systeminterceptorproxy java at com sun ejb containers interceptors systeminterceptorproxy aroundinvoke systeminterceptorproxy java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com sun ejb containers interceptors aroundinvokeinterceptor intercept interceptormanager java at com sun ejb containers interceptors aroundinvokechainimpl invokenext interceptormanager java at com sun ejb containers interceptors interceptormanager intercept interceptormanager java at com sun ejb containers basecontainer intercept basecontainer java at com sun ejb containers basecontainer intercept basecontainer java at com sun ejb containers ejblocalobjectinvocationhandler invoke ejblocalobjectinvocationhandler java more | 1 |
149,361 | 5,717,039,393 | IssuesEvent | 2017-04-19 16:19:52 | GoogleCloudPlatform/google-cloud-node | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-node | closed | automatic batching interface for cloud logging client | api: logging priority: p1 status: acknowledged type: enhancement | The logging API support batching mode manually, where you can put multiple entry in the log.write()
`
log.write([entry, secondEntry]);
`
Is that possible to implement an batching interface to support this automatically?
for example:
```
log.add(entry);
...
log.add(another_entry);
```
In the background, logging client does periodic flush internally sending all buffered logs to cloud logging automatically. The flush is triggered when enough entries are buffered and/or at a fixed interval.
This would make the API better performance wise by making less APIs calls to cloud logging?
| 1.0 | automatic batching interface for cloud logging client - The logging API support batching mode manually, where you can put multiple entry in the log.write()
`
log.write([entry, secondEntry]);
`
Is that possible to implement an batching interface to support this automatically?
for example:
```
log.add(entry);
...
log.add(another_entry);
```
In the background, logging client does periodic flush internally sending all buffered logs to cloud logging automatically. The flush is triggered when enough entries are buffered and/or at a fixed interval.
This would make the API better performance wise by making less APIs calls to cloud logging?
| priority | automatic batching interface for cloud logging client the logging api support batching mode manually where you can put multiple entry in the log write log write is that possible to implement an batching interface to support this automatically for example log add entry log add another entry in the background logging client does periodic flush internally sending all buffered logs to cloud logging automatically the flush is triggered when enough entries are buffered and or at a fixed interval this would make the api better performance wise by making less apis calls to cloud logging | 1 |
81,278 | 3,588,351,599 | IssuesEvent | 2016-01-30 23:28:45 | bigbluebutton/bigbluebutton | https://api.github.com/repos/bigbluebutton/bigbluebutton | closed | Explore ways to reduce the occurrences of disconnects on wireless networks | Client Normal Priority Research Stability | Originally reported on Google Code with ID 1728
```
Currently, if BigBlueButton detects its network connection has dropped, it immediately
disconnects the user and closes the session. This ensures the user does not believe
the client is still connected when, in fact, it's no longer receiving updates from
the server.
The initial disconnect occurs from the red5 server after it fails to get a response
from a heartbeat check for the client. If not response (with a given threshold) is
received, red5 disconnects the client, which triggers the client to close.
Explore if it's possible to adjust the thresholds and lower the frequency of disconnects
on a wireless network without introducing performance (or delays) into the connection
logic.
```
Reported by `ffdixon` on 2014-02-13 01:46:25
| 1.0 | Explore ways to reduce the occurrences of disconnects on wireless networks - Originally reported on Google Code with ID 1728
```
Currently, if BigBlueButton detects its network connection has dropped, it immediately
disconnects the user and closes the session. This ensures the user does not believe
the client is still connected when, in fact, it's no longer receiving updates from
the server.
The initial disconnect occurs from the red5 server after it fails to get a response
from a heartbeat check for the client. If not response (with a given threshold) is
received, red5 disconnects the client, which triggers the client to close.
Explore if it's possible to adjust the thresholds and lower the frequency of disconnects
on a wireless network without introducing performance (or delays) into the connection
logic.
```
Reported by `ffdixon` on 2014-02-13 01:46:25
| priority | explore ways to reduce the occurrences of disconnects on wireless networks originally reported on google code with id currently if bigbluebutton detects its network connection has dropped it immediately disconnects the user and closes the session this ensures the user does not believe the client is still connected when in fact it s no longer receiving updates from the server the initial disconnect occurs from the server after it fails to get a response from a heartbeat check for the client if not response with a given threshold is received disconnects the client which triggers the client to close explore if it s possible to adjust the thresholds and lower the frequency of disconnects on a wireless network without introducing performance or delays into the connection logic reported by ffdixon on | 1 |
329,469 | 28,244,647,100 | IssuesEvent | 2023-04-06 09:47:28 | openBackhaul/AdministratorAdministration | https://api.github.com/repos/openBackhaul/AdministratorAdministration | closed | release-number pattern update | testsuite_to_be_changed | Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'.
Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator.
Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder
This scenario "multiple digit in a placeholder" can b:e added to following services:
Service Layer - Acceptance :: Attribute correctness :: release-number checked?
- [x] /v1/bequeath-your-data-and-die
- [x] /v1/regard-application
- [x] /v1/disregard-application | 1.0 | release-number pattern update - Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'.
Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator.
Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder
This scenario "multiple digit in a placeholder" can b:e added to following services:
Service Layer - Acceptance :: Attribute correctness :: release-number checked?
- [x] /v1/bequeath-your-data-and-die
- [x] /v1/regard-application
- [x] /v1/disregard-application | non_priority | release number pattern update pattern of release number has been updated to already testcases are available to check for too short release number too long release number letters in release number sign in release number incorrect separator additionally a scenario can be added to test whether in each placeholder for a number only two one or digits are allowed in earlier release number more than two digits are allowed in a placeholder this scenario multiple digit in a placeholder can b e added to following services service layer acceptance attribute correctness release number checked bequeath your data and die regard application disregard application | 0 |
12,168 | 3,256,248,635 | IssuesEvent | 2015-10-20 13:01:48 | idies/SkyServer | https://api.github.com/repos/idies/SkyServer | closed | Invalid rectangular search values in IQS result in unhelpful, mishandled error message | errorhandling needstest sciserver_v1_3_0 | If you enter invalid values for the rectangular search option in the Imaging Query Form (IQS), e.g., dec values of 0.2 and -0.2 (second value less than first), the error is not properly handled and results in the following cryptic error message:
SQL returned the following error:
Failed to execute a query: Incorrect syntax near 'Illegal'. Incorrect syntax near 'pERROR:'. | 1.0 | Invalid rectangular search values in IQS result in unhelpful, mishandled error message - If you enter invalid values for the rectangular search option in the Imaging Query Form (IQS), e.g., dec values of 0.2 and -0.2 (second value less than first), the error is not properly handled and results in the following cryptic error message:
SQL returned the following error:
Failed to execute a query: Incorrect syntax near 'Illegal'. Incorrect syntax near 'pERROR:'. | non_priority | invalid rectangular search values in iqs result in unhelpful mishandled error message if you enter invalid values for the rectangular search option in the imaging query form iqs e g dec values of and second value less than first the error is not properly handled and results in the following cryptic error message sql returned the following error failed to execute a query incorrect syntax near illegal incorrect syntax near perror | 0 |
262,583 | 22,919,128,696 | IssuesEvent | 2022-07-17 12:01:26 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed | C-test-failure O-robot O-roachtest branch-master T-sql-queries | roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5771178?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5771178?buildTab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on master @ [e4cafeb8b1d586d091fb98e3e570650d7eeea294](https://github.com/cockroachdb/cockroach/commits/e4cafeb8b1d586d091fb98e3e570650d7eeea294):
```
test artifacts and logs in: /artifacts/sqlsmith/setup=rand-tables/setting=no-mutations/run_1
sqlsmith.go:265,sqlsmith.go:325,test_runner.go:896: error: pq: internal error: lookup for ComparisonExpr ((col_202437)[void] IS DISTINCT FROM (NULL)[unknown])[bool]'s CmpOp failed
stmt:
WITH
with_39168 (col_202437)
AS (
SELECT
*
FROM
(VALUES ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID))
AS tab_95501 (col_202437)
UNION SELECT * FROM (VALUES ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID)) AS tab_95502 (col_202438)
),
with_39169 (col_202439)
AS (
SELECT
*
FROM
(VALUES ('#S':::STRING:::NAME), ('':::STRING:::NAME), (NULL), (NULL)) AS tab_95503 (col_202439)
)
SELECT
COALESCE(cte_ref_11299.col_202437, '':::VOID) AS col_202440
FROM
with_39168 AS cte_ref_11299
WHERE
true
ORDER BY
cte_ref_11299.col_202437, cte_ref_11299.col_202437 DESC
LIMIT
16:::INT8;
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-17718 | 2.0 | roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed - roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5771178?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5771178?buildTab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on master @ [e4cafeb8b1d586d091fb98e3e570650d7eeea294](https://github.com/cockroachdb/cockroach/commits/e4cafeb8b1d586d091fb98e3e570650d7eeea294):
```
test artifacts and logs in: /artifacts/sqlsmith/setup=rand-tables/setting=no-mutations/run_1
sqlsmith.go:265,sqlsmith.go:325,test_runner.go:896: error: pq: internal error: lookup for ComparisonExpr ((col_202437)[void] IS DISTINCT FROM (NULL)[unknown])[bool]'s CmpOp failed
stmt:
WITH
with_39168 (col_202437)
AS (
SELECT
*
FROM
(VALUES ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID))
AS tab_95501 (col_202437)
UNION SELECT * FROM (VALUES ('':::VOID), ('':::VOID), ('':::VOID), ('':::VOID)) AS tab_95502 (col_202438)
),
with_39169 (col_202439)
AS (
SELECT
*
FROM
(VALUES ('#S':::STRING:::NAME), ('':::STRING:::NAME), (NULL), (NULL)) AS tab_95503 (col_202439)
)
SELECT
COALESCE(cte_ref_11299.col_202437, '':::VOID) AS col_202440
FROM
with_39168 AS cte_ref_11299
WHERE
true
ORDER BY
cte_ref_11299.col_202437, cte_ref_11299.col_202437 DESC
LIMIT
16:::INT8;
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-17718 | non_priority | roachtest sqlsmith setup rand tables setting no mutations failed roachtest sqlsmith setup rand tables setting no mutations with on master test artifacts and logs in artifacts sqlsmith setup rand tables setting no mutations run sqlsmith go sqlsmith go test runner go error pq internal error lookup for comparisonexpr col is distinct from null s cmpop failed stmt with with col as select from values void void void void void void as tab col union select from values void void void void as tab col with col as select from values s string name string name null null as tab col select coalesce cte ref col void as col from with as cte ref where true order by cte ref col cte ref col desc limit parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb sql queries jira issue crdb | 0 |
474,220 | 13,654,609,623 | IssuesEvent | 2020-09-27 18:14:30 | JohnVicke/pumaht20 | https://api.github.com/repos/JohnVicke/pumaht20 | opened | [Component]: Card modal | new feature priority:med | Create a card modal with adjustable or fixed width property. Should be able to pass content to the component via props and children. | 1.0 | [Component]: Card modal - Create a card modal with adjustable or fixed width property. Should be able to pass content to the component via props and children. | priority | card modal create a card modal with adjustable or fixed width property should be able to pass content to the component via props and children | 1 |
446,994 | 12,881,505,676 | IssuesEvent | 2020-07-12 12:24:54 | code4romania/stam-acasa | https://api.github.com/repos/code4romania/stam-acasa | closed | Edit content in Despre page | content high-priority | **Change this text**
”Despre proiect
Starea de alertă poate fi dificil de gestionat. Avem din nou voie să ieșim din casă, dar urmând proceduri și reguli foarte clare. Riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și, de asemenea, este util să ne menținem un jurnal al ieșirilor
Telefoanele de urgență sunt împovărate de numărul mare de apeluri, făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală. Toți suntem în aceeași situație. Jurnal Medical este o soluție digitală cu ajutorul căreia, împreună, reducem suprasolicitarea numerelor de urgență, colectăm rapid informații de la o populație foarte mare, ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță.
Jurnal Medical este un proiect dezvoltat în parteneriat cu Guvernul României prin Autoritatea pentru Digitalizarea României, cu colaborarea Ministerului Sănătății, de către organizația Code for Romania, în cadrul programului Code for Romania Task Force. Proiectul este administrat de Autoritatea pentru Digitalizarea României.”
**with**
” Despre proiect
Starea de alertă poate fi dificil de gestionat. Avem voie să ieșim din casă, dar urmând proceduri și reguli foarte clare. Riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și, de asemenea, este util să ne menținem un jurnal al ieșirilor
Telefoanele de urgență sunt împovărate de numărul mare de apeluri, făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală. Toți suntem în aceeași situație. Jurnal Medical este o soluție digitală cu ajutorul căreia, împreună, reducem suprasolicitarea numerelor de urgență, colectăm rapid informații de la o populație foarte mare, ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță.
Jurnal Medical este o soluție digitală ale cărei principale beneficii, pentru cetățenii României și autoritățile sanitare deopotrivă, sunt:
1. atenuarea suprasolicitării numerelor de urgență;
2. colectarea rapidă, constantă și în timp real a informațiilor relevante; de la comunități mari de cetățeni din toată țara;
3. monitorizarea consistentă a stării de sănătate pentru utilizatorii platformei și pentru cei dragi lor, astfel încât să rămânem cu toții în siguranță.
Jurnal Medical este o aplicație web în care fiecare cetățean român, indiferent de localitate, se poate înscrie pentru a-și monitoriza starea de sănătate. Contul de utilizator poate fi creat prin intermediul site-ului. După înregistrare, utilizatorul va avea posibilitatea de a completa datele aferente profilului său, pe cele ale membrilor familiei sau ale altor persoane pe care le are în grijă, urmând ca apoi să completeze formularele recurente de monitorizare.
Datele completate de utilizatorii aplicației sunt monitorizate de Institutul Național de Sănătate Publică și de Direcțiile de Sănătate Publică din toată țara, astfel încât potențialele focare să poată fi monitorizate și gestionate, și pentru ca noi toți să ne putem proteja mai bine de pericol.
Jurnal Medical este un proiect dezvoltat prin parteneriat de Guvernul României, prin Autoritatea pentru Digitalizarea României, cu participarea Ministerului Sănătății și de către organizația Code for Romania, în cadrul programului Code for Romania Task Force. Platforma este administrată de Autoritatea pentru Digitalizarea României. Pentru orice intrebare, ne puteți contacta la adresa jurnalmedical@adr.gov.ro.” | 1.0 | Edit content in Despre page - **Change this text**
”Despre proiect
Starea de alertă poate fi dificil de gestionat. Avem din nou voie să ieșim din casă, dar urmând proceduri și reguli foarte clare. Riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și, de asemenea, este util să ne menținem un jurnal al ieșirilor
Telefoanele de urgență sunt împovărate de numărul mare de apeluri, făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală. Toți suntem în aceeași situație. Jurnal Medical este o soluție digitală cu ajutorul căreia, împreună, reducem suprasolicitarea numerelor de urgență, colectăm rapid informații de la o populație foarte mare, ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță.
Jurnal Medical este un proiect dezvoltat în parteneriat cu Guvernul României prin Autoritatea pentru Digitalizarea României, cu colaborarea Ministerului Sănătății, de către organizația Code for Romania, în cadrul programului Code for Romania Task Force. Proiectul este administrat de Autoritatea pentru Digitalizarea României.”
**with**
” Despre proiect
Starea de alertă poate fi dificil de gestionat. Avem voie să ieșim din casă, dar urmând proceduri și reguli foarte clare. Riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și, de asemenea, este util să ne menținem un jurnal al ieșirilor
Telefoanele de urgență sunt împovărate de numărul mare de apeluri, făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală. Toți suntem în aceeași situație. Jurnal Medical este o soluție digitală cu ajutorul căreia, împreună, reducem suprasolicitarea numerelor de urgență, colectăm rapid informații de la o populație foarte mare, ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță.
Jurnal Medical este o soluție digitală ale cărei principale beneficii, pentru cetățenii României și autoritățile sanitare deopotrivă, sunt:
1. atenuarea suprasolicitării numerelor de urgență;
2. colectarea rapidă, constantă și în timp real a informațiilor relevante; de la comunități mari de cetățeni din toată țara;
3. monitorizarea consistentă a stării de sănătate pentru utilizatorii platformei și pentru cei dragi lor, astfel încât să rămânem cu toții în siguranță.
Jurnal Medical este o aplicație web în care fiecare cetățean român, indiferent de localitate, se poate înscrie pentru a-și monitoriza starea de sănătate. Contul de utilizator poate fi creat prin intermediul site-ului. După înregistrare, utilizatorul va avea posibilitatea de a completa datele aferente profilului său, pe cele ale membrilor familiei sau ale altor persoane pe care le are în grijă, urmând ca apoi să completeze formularele recurente de monitorizare.
Datele completate de utilizatorii aplicației sunt monitorizate de Institutul Național de Sănătate Publică și de Direcțiile de Sănătate Publică din toată țara, astfel încât potențialele focare să poată fi monitorizate și gestionate, și pentru ca noi toți să ne putem proteja mai bine de pericol.
Jurnal Medical este un proiect dezvoltat prin parteneriat de Guvernul României, prin Autoritatea pentru Digitalizarea României, cu participarea Ministerului Sănătății și de către organizația Code for Romania, în cadrul programului Code for Romania Task Force. Platforma este administrată de Autoritatea pentru Digitalizarea României. Pentru orice intrebare, ne puteți contacta la adresa jurnalmedical@adr.gov.ro.” | priority | edit content in despre page change this text ”despre proiect starea de alertă poate fi dificil de gestionat avem din nou voie să ieșim din casă dar urmând proceduri și reguli foarte clare riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și de asemenea este util să ne menținem un jurnal al ieșirilor telefoanele de urgență sunt împovărate de numărul mare de apeluri făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală toți suntem în aceeași situație jurnal medical este o soluție digitală cu ajutorul căreia împreună reducem suprasolicitarea numerelor de urgență colectăm rapid informații de la o populație foarte mare ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță jurnal medical este un proiect dezvoltat în parteneriat cu guvernul româniei prin autoritatea pentru digitalizarea româniei cu colaborarea ministerului sănătății de către organizația code for romania în cadrul programului code for romania task force proiectul este administrat de autoritatea pentru digitalizarea româniei ” with ” despre proiect starea de alertă poate fi dificil de gestionat avem voie să ieșim din casă dar urmând proceduri și reguli foarte clare riscul de a contracta virusul este foarte ridicat și este important să putem să ne monitorizăm starea cu ușurință și de asemenea este util să ne menținem un jurnal al ieșirilor telefoanele de urgență sunt împovărate de numărul mare de apeluri făcând astfel evaluarea situațiilor medicale la nivel național o problemă reală toți suntem în aceeași situație jurnal medical este o soluție digitală cu ajutorul căreia împreună reducem suprasolicitarea numerelor de urgență colectăm rapid informații de la o populație foarte mare ne monitorizăm starea de sănătate pentru noi și pentru cei dragi și rămânem cu toții în siguranță jurnal medical este o soluție digitală ale cărei principale beneficii pentru cetățenii româniei și autoritățile sanitare deopotrivă sunt atenuarea suprasolicitării numerelor de urgență colectarea rapidă constantă și în timp real a informațiilor relevante de la comunități mari de cetățeni din toată țara monitorizarea consistentă a stării de sănătate pentru utilizatorii platformei și pentru cei dragi lor astfel încât să rămânem cu toții în siguranță jurnal medical este o aplicație web în care fiecare cetățean român indiferent de localitate se poate înscrie pentru a și monitoriza starea de sănătate contul de utilizator poate fi creat prin intermediul site ului după înregistrare utilizatorul va avea posibilitatea de a completa datele aferente profilului său pe cele ale membrilor familiei sau ale altor persoane pe care le are în grijă urmând ca apoi să completeze formularele recurente de monitorizare datele completate de utilizatorii aplicației sunt monitorizate de institutul național de sănătate publică și de direcțiile de sănătate publică din toată țara astfel încât potențialele focare să poată fi monitorizate și gestionate și pentru ca noi toți să ne putem proteja mai bine de pericol jurnal medical este un proiect dezvoltat prin parteneriat de guvernul româniei prin autoritatea pentru digitalizarea româniei cu participarea ministerului sănătății și de către organizația code for romania în cadrul programului code for romania task force platforma este administrată de autoritatea pentru digitalizarea româniei pentru orice intrebare ne puteți contacta la adresa jurnalmedical adr gov ro ” | 1 |
199,632 | 6,992,726,771 | IssuesEvent | 2017-12-15 08:25:50 | xwikisas/application-meeting | https://api.github.com/repos/xwikisas/application-meeting | closed | Update Meeting.WebHome to use the #getMissingLicenseMessage macro | Priority: Minor Type: Improvement | We should use the #getMissingLicenseMessage macro instead of the current implementation: https://github.com/xwikisas/application-meeting/blob/de9554f385f90f9bd01cb05d62d01a1749180095/ui/src/main/resources/Meeting/WebHome.xml#L46
The macro also makes sure the link always points to the Main Wiki.
Example usage of the macro here: https://github.com/xwikisas/application-ideas/blob/0aa38048b83ba9e113cdbe7389a1419a1cb7f37f/src/main/resources/Ideas/WebHome.xml#L192 | 1.0 | Update Meeting.WebHome to use the #getMissingLicenseMessage macro - We should use the #getMissingLicenseMessage macro instead of the current implementation: https://github.com/xwikisas/application-meeting/blob/de9554f385f90f9bd01cb05d62d01a1749180095/ui/src/main/resources/Meeting/WebHome.xml#L46
The macro also makes sure the link always points to the Main Wiki.
Example usage of the macro here: https://github.com/xwikisas/application-ideas/blob/0aa38048b83ba9e113cdbe7389a1419a1cb7f37f/src/main/resources/Ideas/WebHome.xml#L192 | priority | update meeting webhome to use the getmissinglicensemessage macro we should use the getmissinglicensemessage macro instead of the current implementation the macro also makes sure the link always points to the main wiki example usage of the macro here | 1 |
382,973 | 11,341,172,036 | IssuesEvent | 2020-01-23 08:44:27 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | Change Copy for "Allow conversion tracking" in Brave Ads "Ads Settings" | enhancement feature/ads priority/P3 | Change Copy for "Allow conversion tracking" in Brave Ads "Ads Settings" from:
<img width="640" alt="Screenshot 2020-01-23 at 08 43 22" src="https://user-images.githubusercontent.com/1188719/72968955-7c7e0100-3dbc-11ea-8b08-281678cb6929.png">
to:
Allow ad conversion detection
This will anonymously let Brave know that there was a response to an ad on this browser. Enabling this will increase earnings. | 1.0 | Change Copy for "Allow conversion tracking" in Brave Ads "Ads Settings" - Change Copy for "Allow conversion tracking" in Brave Ads "Ads Settings" from:
<img width="640" alt="Screenshot 2020-01-23 at 08 43 22" src="https://user-images.githubusercontent.com/1188719/72968955-7c7e0100-3dbc-11ea-8b08-281678cb6929.png">
to:
Allow ad conversion detection
This will anonymously let Brave know that there was a response to an ad on this browser. Enabling this will increase earnings. | priority | change copy for allow conversion tracking in brave ads ads settings change copy for allow conversion tracking in brave ads ads settings from img width alt screenshot at src to allow ad conversion detection this will anonymously let brave know that there was a response to an ad on this browser enabling this will increase earnings | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.