Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,406
| 3,201,896,449
|
IssuesEvent
|
2015-10-02 10:34:47
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
Fix for GSQL backend to enable multiple dnssec storage backends
|
auth defect
|
Due to mistake in GSQL backend, one cannot use more than one storage for DNSSEC data. The patch included fixes this by checking whether the backend in question actually serves the domain, causing Ueberbackend to find correct backend for the domain. This allows to have multiple SQL backends with DNSSEC enabled.
|
1.0
|
Fix for GSQL backend to enable multiple dnssec storage backends - Due to mistake in GSQL backend, one cannot use more than one storage for DNSSEC data. The patch included fixes this by checking whether the backend in question actually serves the domain, causing Ueberbackend to find correct backend for the domain. This allows to have multiple SQL backends with DNSSEC enabled.
|
defect
|
fix for gsql backend to enable multiple dnssec storage backends due to mistake in gsql backend one cannot use more than one storage for dnssec data the patch included fixes this by checking whether the backend in question actually serves the domain causing ueberbackend to find correct backend for the domain this allows to have multiple sql backends with dnssec enabled
| 1
|
68,828
| 21,918,348,280
|
IssuesEvent
|
2022-05-22 07:07:30
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Other user's name wrapped on bubble message layout
|
T-Defect S-Tolerable A-Appearance A-Message-Bubbles O-Uncommon
|
### Steps to reproduce
1. Create a test user
2. Set a long user name to the user
3. Invite the user to a test room
4. Send a message from the user
5. Open the room with the main account
### Outcome
#### What did you expect?
The user name should be hidden with ellipsis, following other cases on UI.
#### What happened instead?
The user name is wrapped.

### Operating system
Debian
### Browser information
Firefox
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Other user's name wrapped on bubble message layout - ### Steps to reproduce
1. Create a test user
2. Set a long user name to the user
3. Invite the user to a test room
4. Send a message from the user
5. Open the room with the main account
### Outcome
#### What did you expect?
The user name should be hidden with ellipsis, following other cases on UI.
#### What happened instead?
The user name is wrapped.

### Operating system
Debian
### Browser information
Firefox
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
other user s name wrapped on bubble message layout steps to reproduce create a test user set a long user name to the user invite the user to a test room send a message from the user open the room with the main account outcome what did you expect the user name should be hidden with ellipsis following other cases on ui what happened instead the user name is wrapped operating system debian browser information firefox url for webapp localhost application version develop branch homeserver no response will you send logs no
| 1
|
122,554
| 26,140,655,617
|
IssuesEvent
|
2022-12-29 17:56:14
|
gdscashesi/ashesi-hackers-league
|
https://api.github.com/repos/gdscashesi/ashesi-hackers-league
|
closed
|
Update Table with points
|
enhancement code
|
This page will let us update the table with points for the individual teams
page: **/admin/update**
We will re-use the table component, and add an additional column : **action**
This column will contain **3 counter buttons** for each of algo, scripting, SQL for updating a specific row
|rank | name |algorithms |scripting |sql |total|action
|--|--|--|--|--|--|--|
|1|Berekusu Warriors|21|45|3|69| +- +- +-|
|2|Ashesi Ninjas|27|9|6|42| +- +- +- |
|3|Crusading Coders|18|1|0|19| +- +- +-|
but here comes the **hard part**
- [ ] we should be able to look at the before and after states of the table (before updating, after updating because the ranking will change)
but for now, implement just the action column and make sure its working. Let me know when this is done before you start the **hard part**
If you have a better/interesting way we can do this. Let me know ASAP. Otherwise, start work!
|
1.0
|
Update Table with points - This page will let us update the table with points for the individual teams
page: **/admin/update**
We will re-use the table component, and add an additional column : **action**
This column will contain **3 counter buttons** for each of algo, scripting, SQL for updating a specific row
|rank | name |algorithms |scripting |sql |total|action
|--|--|--|--|--|--|--|
|1|Berekusu Warriors|21|45|3|69| +- +- +-|
|2|Ashesi Ninjas|27|9|6|42| +- +- +- |
|3|Crusading Coders|18|1|0|19| +- +- +-|
but here comes the **hard part**
- [ ] we should be able to look at the before and after states of the table (before updating, after updating because the ranking will change)
but for now, implement just the action column and make sure its working. Let me know when this is done before you start the **hard part**
If you have a better/interesting way we can do this. Let me know ASAP. Otherwise, start work!
|
non_defect
|
update table with points this page will let us update the table with points for the individual teams page admin update we will re use the table component and add an additional column action this column will contain counter buttons for each of algo scripting sql for updating a specific row rank name algorithms scripting sql total action berekusu warriors ashesi ninjas crusading coders but here comes the hard part we should be able to look at the before and after states of the table before updating after updating because the ranking will change but for now implement just the action column and make sure its working let me know when this is done before you start the hard part if you have a better interesting way we can do this let me know asap otherwise start work
| 0
|
165,766
| 20,619,777,231
|
IssuesEvent
|
2022-03-07 16:21:00
|
Azure/AKS
|
https://api.github.com/repos/Azure/AKS
|
opened
|
CVE-2022-0847: Linux kernel: overwriting read-only files
|
security
|
A new vulnerability in the Linux kernel since
version 5.8 commit f6dd975583bd ("pipe: merge anon_pipe_buf*_ops") due
to uninitialized variables. It enables anybody to write arbitrary
data to arbitrary files, even if the file is O_RDONLY, immutable or on
a MS_RDONLY filesystem. It can be used to inject code into arbitrary
processes.
It is similar to CVE-2016-5195 "Dirty Cow", but is easier to exploit.
The vulnerability was fixed in Linux 5.16.11, 5.15.25 and 5.10.102.
https://www.openwall.com/lists/oss-security/2022/03/07/1
# AKS Information:
AKS is working on an updated Node Image with the required updated kernels. This issue will be updated once a build has been commited.
|
True
|
CVE-2022-0847: Linux kernel: overwriting read-only files - A new vulnerability in the Linux kernel since
version 5.8 commit f6dd975583bd ("pipe: merge anon_pipe_buf*_ops") due
to uninitialized variables. It enables anybody to write arbitrary
data to arbitrary files, even if the file is O_RDONLY, immutable or on
a MS_RDONLY filesystem. It can be used to inject code into arbitrary
processes.
It is similar to CVE-2016-5195 "Dirty Cow", but is easier to exploit.
The vulnerability was fixed in Linux 5.16.11, 5.15.25 and 5.10.102.
https://www.openwall.com/lists/oss-security/2022/03/07/1
# AKS Information:
AKS is working on an updated Node Image with the required updated kernels. This issue will be updated once a build has been commited.
|
non_defect
|
cve linux kernel overwriting read only files a new vulnerability in the linux kernel since version commit pipe merge anon pipe buf ops due to uninitialized variables it enables anybody to write arbitrary data to arbitrary files even if the file is o rdonly immutable or on a ms rdonly filesystem it can be used to inject code into arbitrary processes it is similar to cve dirty cow but is easier to exploit the vulnerability was fixed in linux and aks information aks is working on an updated node image with the required updated kernels this issue will be updated once a build has been commited
| 0
|
135,740
| 12,690,899,978
|
IssuesEvent
|
2020-06-21 14:28:40
|
kubernetes/kubeadm
|
https://api.github.com/repos/kubernetes/kubeadm
|
closed
|
update kubeadm-kubelet documentation to reflect CNI package change
|
kind/documentation lifecycle/active priority/important-longterm
|
the kubernetes-cni contents (plugins) will now be bundled as part of the kubelet package:
https://github.com/kubernetes/release/pull/1330
the kubernetes-cni package will be removed.
the action item here is to remove:
> kubernetes-cni
from this list:
https://kubernetes.io/docs/setup/production-environment/tools/kubeadm/kubelet-integration/#kubernetes-binaries-and-package-contents
and update the item:
> kubelet
to have the description:
> Installs the kubelet binary in /usr/bin and CNI binaries in /opt/cni/bin.
file location:
https://github.com/kubernetes/website/blob/master/content/en/docs/setup/production-environment/tools/kubeadm/kubelet-integration.md
/priority important-longterm
/kind documentation
/help
/good-first-issue
|
1.0
|
update kubeadm-kubelet documentation to reflect CNI package change - the kubernetes-cni contents (plugins) will now be bundled as part of the kubelet package:
https://github.com/kubernetes/release/pull/1330
the kubernetes-cni package will be removed.
the action item here is to remove:
> kubernetes-cni
from this list:
https://kubernetes.io/docs/setup/production-environment/tools/kubeadm/kubelet-integration/#kubernetes-binaries-and-package-contents
and update the item:
> kubelet
to have the description:
> Installs the kubelet binary in /usr/bin and CNI binaries in /opt/cni/bin.
file location:
https://github.com/kubernetes/website/blob/master/content/en/docs/setup/production-environment/tools/kubeadm/kubelet-integration.md
/priority important-longterm
/kind documentation
/help
/good-first-issue
|
non_defect
|
update kubeadm kubelet documentation to reflect cni package change the kubernetes cni contents plugins will now be bundled as part of the kubelet package the kubernetes cni package will be removed the action item here is to remove kubernetes cni from this list and update the item kubelet to have the description installs the kubelet binary in usr bin and cni binaries in opt cni bin file location priority important longterm kind documentation help good first issue
| 0
|
242,993
| 18,675,300,302
|
IssuesEvent
|
2021-10-31 13:06:11
|
AY2122S1-CS2103T-T17-3/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-T17-3/tp
|
closed
|
[PE-D] UG examples for filter is wrong
|
documentation
|

Note the examples are for `search` not filter
<!--session: 1635495170210-7d18469e-ff21-47c2-9921-cf910f65f10c-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.DocumentationBug` `severity.Low`
original: mokdarren/ped#4
|
1.0
|
[PE-D] UG examples for filter is wrong - 
Note the examples are for `search` not filter
<!--session: 1635495170210-7d18469e-ff21-47c2-9921-cf910f65f10c-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.DocumentationBug` `severity.Low`
original: mokdarren/ped#4
|
non_defect
|
ug examples for filter is wrong note the examples are for search not filter labels type documentationbug severity low original mokdarren ped
| 0
|
78,770
| 27,751,929,167
|
IssuesEvent
|
2023-03-15 21:31:45
|
jurgendl/hql-builder
|
https://api.github.com/repos/jurgendl/hql-builder
|
closed
|
exception after syntax error on EDT
|
Type-Defect auto-migrated Priority-Medium no-issue-activity
|
```
java.lang.ArrayIndexOutOfBoundsException: 0 >= 0
at java.util.Vector.elementAt(Vector.java:427)
at javax.swing.text.DefaultHighlighter.paintLayeredHighlights(DefaultHighlighter.java:268)
at javax.swing.text.PlainView.paint(PlainView.java:283)
at javax.swing.plaf.basic.BasicTextUI$RootView.paint(BasicTextUI.java:1422)
at javax.swing.plaf.basic.BasicTextUI.paintSafely(BasicTextUI.java:722)
at javax.swing.plaf.basic.BasicTextUI.paint(BasicTextUI.java:869)
at javax.swing.plaf.basic.BasicTextUI.update(BasicTextUI.java:848)
at javax.swing.JComponent.paintComponent(JComponent.java:752)
at javax.swing.JComponent.paint(JComponent.java:1029)
at javax.swing.JComponent.paintToOffscreen(JComponent.java:5124)
at javax.swing.RepaintManager$PaintManager.paintDoubleBuffered(RepaintManager.java:1491)
at javax.swing.RepaintManager$PaintManager.paint(RepaintManager.java:1422)
at javax.swing.RepaintManager.paint(RepaintManager.java:1225)
at javax.swing.JComponent._paintImmediately(JComponent.java:5072)
at javax.swing.JComponent.paintImmediately(JComponent.java:4882)
at javax.swing.RepaintManager.paintDirtyRegions(RepaintManager.java:786)
at javax.swing.RepaintManager.paintDirtyRegions(RepaintManager.java:714)
at javax.swing.RepaintManager.prePaintDirtyRegions(RepaintManager.java:694)
at javax.swing.RepaintManager.access$700(RepaintManager.java:41)
at javax.swing.RepaintManager$ProcessingRunnable.run(RepaintManager.java:1636)
at java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:209)
at java.awt.EventQueue.dispatchEventImpl(EventQueue.java:646)
at java.awt.EventQueue.access$000(EventQueue.java:84)
at java.awt.EventQueue$1.run(EventQueue.java:607)
at java.awt.EventQueue$1.run(EventQueue.java:605)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.AccessControlContext$1.doIntersectionPrivilege(AccessControlContext.java:87)
at java.awt.EventQueue.dispatchEvent(EventQueue.java:616)
at java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:269)
at java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:184)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:174)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:169)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:161)
at java.awt.EventDispatchThread.run(EventDispatchThread.java:122)
```
Original issue reported on code.google.com by `jurgen.d...@gmail.com` on 12 Jul 2012 at 3:10
|
1.0
|
exception after syntax error on EDT - ```
java.lang.ArrayIndexOutOfBoundsException: 0 >= 0
at java.util.Vector.elementAt(Vector.java:427)
at javax.swing.text.DefaultHighlighter.paintLayeredHighlights(DefaultHighlighter.java:268)
at javax.swing.text.PlainView.paint(PlainView.java:283)
at javax.swing.plaf.basic.BasicTextUI$RootView.paint(BasicTextUI.java:1422)
at javax.swing.plaf.basic.BasicTextUI.paintSafely(BasicTextUI.java:722)
at javax.swing.plaf.basic.BasicTextUI.paint(BasicTextUI.java:869)
at javax.swing.plaf.basic.BasicTextUI.update(BasicTextUI.java:848)
at javax.swing.JComponent.paintComponent(JComponent.java:752)
at javax.swing.JComponent.paint(JComponent.java:1029)
at javax.swing.JComponent.paintToOffscreen(JComponent.java:5124)
at javax.swing.RepaintManager$PaintManager.paintDoubleBuffered(RepaintManager.java:1491)
at javax.swing.RepaintManager$PaintManager.paint(RepaintManager.java:1422)
at javax.swing.RepaintManager.paint(RepaintManager.java:1225)
at javax.swing.JComponent._paintImmediately(JComponent.java:5072)
at javax.swing.JComponent.paintImmediately(JComponent.java:4882)
at javax.swing.RepaintManager.paintDirtyRegions(RepaintManager.java:786)
at javax.swing.RepaintManager.paintDirtyRegions(RepaintManager.java:714)
at javax.swing.RepaintManager.prePaintDirtyRegions(RepaintManager.java:694)
at javax.swing.RepaintManager.access$700(RepaintManager.java:41)
at javax.swing.RepaintManager$ProcessingRunnable.run(RepaintManager.java:1636)
at java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:209)
at java.awt.EventQueue.dispatchEventImpl(EventQueue.java:646)
at java.awt.EventQueue.access$000(EventQueue.java:84)
at java.awt.EventQueue$1.run(EventQueue.java:607)
at java.awt.EventQueue$1.run(EventQueue.java:605)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.AccessControlContext$1.doIntersectionPrivilege(AccessControlContext.java:87)
at java.awt.EventQueue.dispatchEvent(EventQueue.java:616)
at java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:269)
at java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:184)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:174)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:169)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:161)
at java.awt.EventDispatchThread.run(EventDispatchThread.java:122)
```
Original issue reported on code.google.com by `jurgen.d...@gmail.com` on 12 Jul 2012 at 3:10
|
defect
|
exception after syntax error on edt java lang arrayindexoutofboundsexception at java util vector elementat vector java at javax swing text defaulthighlighter paintlayeredhighlights defaulthighlighter java at javax swing text plainview paint plainview java at javax swing plaf basic basictextui rootview paint basictextui java at javax swing plaf basic basictextui paintsafely basictextui java at javax swing plaf basic basictextui paint basictextui java at javax swing plaf basic basictextui update basictextui java at javax swing jcomponent paintcomponent jcomponent java at javax swing jcomponent paint jcomponent java at javax swing jcomponent painttooffscreen jcomponent java at javax swing repaintmanager paintmanager paintdoublebuffered repaintmanager java at javax swing repaintmanager paintmanager paint repaintmanager java at javax swing repaintmanager paint repaintmanager java at javax swing jcomponent paintimmediately jcomponent java at javax swing jcomponent paintimmediately jcomponent java at javax swing repaintmanager paintdirtyregions repaintmanager java at javax swing repaintmanager paintdirtyregions repaintmanager java at javax swing repaintmanager prepaintdirtyregions repaintmanager java at javax swing repaintmanager access repaintmanager java at javax swing repaintmanager processingrunnable run repaintmanager java at java awt event invocationevent dispatch invocationevent java at java awt eventqueue dispatcheventimpl eventqueue java at java awt eventqueue access eventqueue java at java awt eventqueue run eventqueue java at java awt eventqueue run eventqueue java at java security accesscontroller doprivileged native method at java security accesscontrolcontext dointersectionprivilege accesscontrolcontext java at java awt eventqueue dispatchevent eventqueue java at java awt eventdispatchthread pumponeeventforfilters eventdispatchthread java at java awt eventdispatchthread pumpeventsforfilter eventdispatchthread java at java awt eventdispatchthread pumpeventsforhierarchy eventdispatchthread java at java awt eventdispatchthread pumpevents eventdispatchthread java at java awt eventdispatchthread pumpevents eventdispatchthread java at java awt eventdispatchthread run eventdispatchthread java original issue reported on code google com by jurgen d gmail com on jul at
| 1
|
165,973
| 14,016,642,009
|
IssuesEvent
|
2020-10-29 14:45:21
|
TUM-IDP-WS-20/doc
|
https://api.github.com/repos/TUM-IDP-WS-20/doc
|
closed
|
Research on Topic modelling techniques
|
documentation
|
Parent: #1
---
- [x] Find all topic modelling techniques
- [x] Compare and document powerful features of each technique
- [x] SVD
- [x] LSA
- [x] LDA
- [x] ...
|
1.0
|
Research on Topic modelling techniques - Parent: #1
---
- [x] Find all topic modelling techniques
- [x] Compare and document powerful features of each technique
- [x] SVD
- [x] LSA
- [x] LDA
- [x] ...
|
non_defect
|
research on topic modelling techniques parent find all topic modelling techniques compare and document powerful features of each technique svd lsa lda
| 0
|
27,675
| 2,695,028,601
|
IssuesEvent
|
2015-04-02 00:26:00
|
cs2103jan2015-t10-2j/main
|
https://api.github.com/repos/cs2103jan2015-t10-2j/main
|
closed
|
allow the user to skip certain arguments in the add/alter functions.
|
priority.high type.task
|
For example, if the user wants to have no description, no date, etc.
This will also help to solve #15
|
1.0
|
allow the user to skip certain arguments in the add/alter functions. - For example, if the user wants to have no description, no date, etc.
This will also help to solve #15
|
non_defect
|
allow the user to skip certain arguments in the add alter functions for example if the user wants to have no description no date etc this will also help to solve
| 0
|
21,306
| 3,487,401,734
|
IssuesEvent
|
2016-01-01 21:34:00
|
Ryzhehvost/keyla
|
https://api.github.com/repos/Ryzhehvost/keyla
|
closed
|
x64 doesn't start
|
auto-migrated Can't reproduce Priority-Medium Type-Defect
|
```
Установлена версия x64 0.1.9
При запуске ошибка
---------------------------
keyla.exe - Ошибка приложения
---------------------------
Ошибка при запуске приложения (0xc000007b). Для
выхода из приложения нажмите кнопку "ОК".
---------------------------
ОК
---------------------------
Система Win 7 Pro x64 Ru SP1
В системе установлен Punto Switcher, но он
выгружен из памяти.
Сейчас попробую перезагрузиться и
проверить снова.
```
Original issue reported on code.google.com by `hobbit.mage` on 10 Nov 2012 at 12:20
|
1.0
|
x64 doesn't start - ```
Установлена версия x64 0.1.9
При запуске ошибка
---------------------------
keyla.exe - Ошибка приложения
---------------------------
Ошибка при запуске приложения (0xc000007b). Для
выхода из приложения нажмите кнопку "ОК".
---------------------------
ОК
---------------------------
Система Win 7 Pro x64 Ru SP1
В системе установлен Punto Switcher, но он
выгружен из памяти.
Сейчас попробую перезагрузиться и
проверить снова.
```
Original issue reported on code.google.com by `hobbit.mage` on 10 Nov 2012 at 12:20
|
defect
|
doesn t start установлена версия при запуске ошибка keyla exe ошибка приложения ошибка при запуске приложения для выхода из приложения нажмите кнопку ок ок система win pro ru в системе установлен punto switcher но он выгружен из памяти сейчас попробую перезагрузиться и проверить снова original issue reported on code google com by hobbit mage on nov at
| 1
|
792,410
| 27,959,161,502
|
IssuesEvent
|
2023-03-24 14:29:32
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
play.google.com - see bug description
|
status-needsinfo priority-critical browser-focus-geckoview engine-gecko
|
<!-- @browser: Firefox Mobile 101.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:101.0) Gecko/101.0 Firefox/101.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://play.google.com/store/apps/details?id=com.snapchat.android
**Browser / Version**: Firefox Mobile 101.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: advertising worst
**Steps to Reproduce**:
I m disturbe with these erritate
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/7727b414-eec4-4b5e-bc0e-e86489696131.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220608170832</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/3/acddf479-9640-420c-9ec9-f76db4a5f312)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
play.google.com - see bug description - <!-- @browser: Firefox Mobile 101.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:101.0) Gecko/101.0 Firefox/101.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://play.google.com/store/apps/details?id=com.snapchat.android
**Browser / Version**: Firefox Mobile 101.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: advertising worst
**Steps to Reproduce**:
I m disturbe with these erritate
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/7727b414-eec4-4b5e-bc0e-e86489696131.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220608170832</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/3/acddf479-9640-420c-9ec9-f76db4a5f312)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
play google com see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description advertising worst steps to reproduce i m disturbe with these erritate view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
53,742
| 13,262,217,162
|
IssuesEvent
|
2020-08-20 21:19:49
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[core-removal] move documentation to sphinx (Trac #1988)
|
Migrated from Trac analysis defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1988">https://code.icecube.wisc.edu/projects/icecube/ticket/1988</a>, reported by kjmeagher</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2017-04-26T19:06:22",
"component": "analysis",
"summary": "[core-removal] move documentation to sphinx",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[core-removal] move documentation to sphinx (Trac #1988) -
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1988">https://code.icecube.wisc.edu/projects/icecube/ticket/1988</a>, reported by kjmeagher</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2017-04-26T19:06:22",
"component": "analysis",
"summary": "[core-removal] move documentation to sphinx",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
defect
|
move documentation to sphinx trac migrated from json status closed changetime ts description reporter kjmeagher cc resolution fixed time component analysis summary move documentation to sphinx priority normal keywords milestone owner type defect
| 1
|
84,218
| 10,361,442,377
|
IssuesEvent
|
2019-09-06 10:01:41
|
Sharaal/sql-pg
|
https://api.github.com/repos/Sharaal/sql-pg
|
closed
|
Refactor the README.md
|
documentation
|
Currently the README.md is good, but there are some things which can be made a lot better like described here very good:
https://dev.to/carlillo/how-to-reach-your-goals-1000-github-stars-in-the-first-open-source-software-337h
|
1.0
|
Refactor the README.md - Currently the README.md is good, but there are some things which can be made a lot better like described here very good:
https://dev.to/carlillo/how-to-reach-your-goals-1000-github-stars-in-the-first-open-source-software-337h
|
non_defect
|
refactor the readme md currently the readme md is good but there are some things which can be made a lot better like described here very good
| 0
|
240,798
| 20,073,997,900
|
IssuesEvent
|
2022-02-04 10:34:08
|
tracim/tracim
|
https://api.github.com/repos/tracim/tracim
|
closed
|
Bug: Activation/Deactivation of agenda is not properly handled.
|
manually tested backend add to changelog p0
|
## Description and expectations
Currently switching the button for activation or disabling of workspace agenda does not have any consequences for the user calendars list.
To make things work, we do need to add/delete symlink during workspace_modification if the agenda enabled/disabled state as changed.
The fix doesn't need to fix existing incorrect case for existing agenda as switching two time the state, will restore the correct state.
<!-- Optionally, if you know how to get them through the browser's developer tools, please include console logs written during the bug occurence. You can also include HTTP responses which have a 4xx or 5xx error code. -->
### Version information
<!-- Optional - describe the environment which reveals the bug: -->
- Tracim version: v4.0.5 build 067
## Diagnostic
change should be made in this event handler:
https://github.com/tracim/tracim/blob/a0e9746fde5a4c45b4e0f0bfa2caf9522b8c4e21/backend/tracim_backend/applications/agenda/lib.py#L599
<!-- Explain what may cause the bug. -->
## Prerequisites
<!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. -->
<!-- ## Required sections, if relevant ## -->
<!-- - To be discussed before development -->
<!-- - Interface -->
<!-- - Translations -->
<!-- - Workaround -->
<!-- - Extra information -->
## Implemented solution
https://github.com/tracim/tracim/pull/5323
- handle properly the activation/deactivation case of agenda in hooks.
- refactor caldav sync code to work more in a "set to this state" logic than "do this" logic.
- fix issues during usage of tracimcli user delete command (agenda not properly deleted)
- correct error code when calendar is disabled for better support of DAVx⁵.
|
1.0
|
Bug: Activation/Deactivation of agenda is not properly handled. - ## Description and expectations
Currently switching the button for activation or disabling of workspace agenda does not have any consequences for the user calendars list.
To make things work, we do need to add/delete symlink during workspace_modification if the agenda enabled/disabled state as changed.
The fix doesn't need to fix existing incorrect case for existing agenda as switching two time the state, will restore the correct state.
<!-- Optionally, if you know how to get them through the browser's developer tools, please include console logs written during the bug occurence. You can also include HTTP responses which have a 4xx or 5xx error code. -->
### Version information
<!-- Optional - describe the environment which reveals the bug: -->
- Tracim version: v4.0.5 build 067
## Diagnostic
change should be made in this event handler:
https://github.com/tracim/tracim/blob/a0e9746fde5a4c45b4e0f0bfa2caf9522b8c4e21/backend/tracim_backend/applications/agenda/lib.py#L599
<!-- Explain what may cause the bug. -->
## Prerequisites
<!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. -->
<!-- ## Required sections, if relevant ## -->
<!-- - To be discussed before development -->
<!-- - Interface -->
<!-- - Translations -->
<!-- - Workaround -->
<!-- - Extra information -->
## Implemented solution
https://github.com/tracim/tracim/pull/5323
- handle properly the activation/deactivation case of agenda in hooks.
- refactor caldav sync code to work more in a "set to this state" logic than "do this" logic.
- fix issues during usage of tracimcli user delete command (agenda not properly deleted)
- correct error code when calendar is disabled for better support of DAVx⁵.
|
non_defect
|
bug activation deactivation of agenda is not properly handled description and expectations currently switching the button for activation or disabling of workspace agenda does not have any consequences for the user calendars list to make things work we do need to add delete symlink during workspace modification if the agenda enabled disabled state as changed the fix doesn t need to fix existing incorrect case for existing agenda as switching two time the state will restore the correct state version information tracim version build diagnostic change should be made in this event handler prerequisites implemented solution handle properly the activation deactivation case of agenda in hooks refactor caldav sync code to work more in a set to this state logic than do this logic fix issues during usage of tracimcli user delete command agenda not properly deleted correct error code when calendar is disabled for better support of davx⁵
| 0
|
71,333
| 23,545,368,080
|
IssuesEvent
|
2022-08-21 02:47:45
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
OauthTest::testRsaSigningString() fails with OpenSSL 3 on Ubuntu 22.04
|
defect
|
### Description
```
1) Cake\Test\TestCase\Http\Client\Auth\OauthTest::testRsaSigningString
RuntimeException: openssl error: error:0A000126:SSL routines::unexpected eof while readingerror:0A000197:SSL routines::shutdown while in init
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:388
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:223
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:72
/home/runner/work/cakephp/cakephp/tests/TestCase/Http/Client/Auth/OauthTest.php:378
phpvfscomposer:///home/runner/work/cakephp/cakephp/vendor/phpunit/phpunit/phpunit:97
```
### CakePHP Version
4.4
### PHP Version
_No response_
|
1.0
|
OauthTest::testRsaSigningString() fails with OpenSSL 3 on Ubuntu 22.04 - ### Description
```
1) Cake\Test\TestCase\Http\Client\Auth\OauthTest::testRsaSigningString
RuntimeException: openssl error: error:0A000126:SSL routines::unexpected eof while readingerror:0A000197:SSL routines::shutdown while in init
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:388
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:223
/home/runner/work/cakephp/cakephp/src/Http/Client/Auth/Oauth.php:72
/home/runner/work/cakephp/cakephp/tests/TestCase/Http/Client/Auth/OauthTest.php:378
phpvfscomposer:///home/runner/work/cakephp/cakephp/vendor/phpunit/phpunit/phpunit:97
```
### CakePHP Version
4.4
### PHP Version
_No response_
|
defect
|
oauthtest testrsasigningstring fails with openssl on ubuntu description cake test testcase http client auth oauthtest testrsasigningstring runtimeexception openssl error error ssl routines unexpected eof while readingerror ssl routines shutdown while in init home runner work cakephp cakephp src http client auth oauth php home runner work cakephp cakephp src http client auth oauth php home runner work cakephp cakephp src http client auth oauth php home runner work cakephp cakephp tests testcase http client auth oauthtest php phpvfscomposer home runner work cakephp cakephp vendor phpunit phpunit phpunit cakephp version php version no response
| 1
|
57,841
| 16,100,979,469
|
IssuesEvent
|
2021-04-27 09:14:55
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Invites(?) cause console to flood with "Room does not have an m.room.create event" warnings
|
T-Defect
|
With spaces enabled on today's nightly, my console has nothing but this in it:
```
10:11:57.373 rageshake.js?432e:65 Room !whereverZflxV:matrix.org does not have an m.room.create event
consoleObj.<computed> @ rageshake.js?432e:65
eval @ logger.ts?6b0b:50
Room.getType @ room.js?6146:1966
Room.isSpaceRoom @ room.js?6146:1977
isRoomVisible @ VisibilityProvider.ts?84e8:53
get globalState @ RoomNotificationStateStore.ts?54ff:51
updateStatusIndicator @ MatrixChat.tsx?b3f7:1899
eval @ MatrixChat.tsx?b3f7:1407
emit @ events.js?faa1:158
SyncApi._updateSyncState @ sync.js?cdb6:1695
SyncApi._sync @ sync.js?cdb6:803
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:762
|
1.0
|
Invites(?) cause console to flood with "Room does not have an m.room.create event" warnings - With spaces enabled on today's nightly, my console has nothing but this in it:
```
10:11:57.373 rageshake.js?432e:65 Room !whereverZflxV:matrix.org does not have an m.room.create event
consoleObj.<computed> @ rageshake.js?432e:65
eval @ logger.ts?6b0b:50
Room.getType @ room.js?6146:1966
Room.isSpaceRoom @ room.js?6146:1977
isRoomVisible @ VisibilityProvider.ts?84e8:53
get globalState @ RoomNotificationStateStore.ts?54ff:51
updateStatusIndicator @ MatrixChat.tsx?b3f7:1899
eval @ MatrixChat.tsx?b3f7:1407
emit @ events.js?faa1:158
SyncApi._updateSyncState @ sync.js?cdb6:1695
SyncApi._sync @ sync.js?cdb6:803
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:744
SyncApi._sync @ sync.js?cdb6:820
async function (async)
SyncApi._sync @ sync.js?cdb6:762
|
defect
|
invites cause console to flood with room does not have an m room create event warnings with spaces enabled on today s nightly my console has nothing but this in it rageshake js room whereverzflxv matrix org does not have an m room create event consoleobj rageshake js eval logger ts room gettype room js room isspaceroom room js isroomvisible visibilityprovider ts get globalstate roomnotificationstatestore ts updatestatusindicator matrixchat tsx eval matrixchat tsx emit events js syncapi updatesyncstate sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js syncapi sync sync js async function async syncapi sync sync js
| 1
|
31,819
| 6,630,733,769
|
IssuesEvent
|
2017-09-25 02:02:03
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
closed
|
Field note visual height glitch on Chrome
|
3.x 4.x defect
|
Found: 4.2.0
Ext.NET forum thread: [Bug when displaying Field 3.x and 4.x ](http://forums.ext.net/showthread.php?61738)
On chrome, the spacing between fields looks broken, while it works fine in IE and FireFox.
This simple CSS override seems to address the issue:
```html
<style type="text/css">
.x-form-item-body {
height: initial
}
</style>
```
|
1.0
|
Field note visual height glitch on Chrome - Found: 4.2.0
Ext.NET forum thread: [Bug when displaying Field 3.x and 4.x ](http://forums.ext.net/showthread.php?61738)
On chrome, the spacing between fields looks broken, while it works fine in IE and FireFox.
This simple CSS override seems to address the issue:
```html
<style type="text/css">
.x-form-item-body {
height: initial
}
</style>
```
|
defect
|
field note visual height glitch on chrome found ext net forum thread on chrome the spacing between fields looks broken while it works fine in ie and firefox this simple css override seems to address the issue html x form item body height initial
| 1
|
43,857
| 11,868,853,957
|
IssuesEvent
|
2020-03-26 09:56:03
|
contao/contao
|
https://api.github.com/repos/contao/contao
|
closed
|
Leeres Auswahlmenü für Ursprungstemplate bei Vergleich von Templates mit abweichendem Präfix
|
defect
|
Beim Vergleich von Templates ist das Auswahlmenü für das Ursprungstemplate leer, wenn das zu vergleichende Template ein abweichendes Präfix hat (zum Beispiel `x_x.html5`). Dadurch ist es nicht möglich, einen Vergleich durchzuführen.
Entweder könnte der Button zur Vergleichsansicht in diesem Fall deaktiviert werden oder es könnten wie bei Templates ohne Präfix (zum Beispiel `x.html5`) im Auswahlmenü alle Templates angezeigt werden.
|
1.0
|
Leeres Auswahlmenü für Ursprungstemplate bei Vergleich von Templates mit abweichendem Präfix - Beim Vergleich von Templates ist das Auswahlmenü für das Ursprungstemplate leer, wenn das zu vergleichende Template ein abweichendes Präfix hat (zum Beispiel `x_x.html5`). Dadurch ist es nicht möglich, einen Vergleich durchzuführen.
Entweder könnte der Button zur Vergleichsansicht in diesem Fall deaktiviert werden oder es könnten wie bei Templates ohne Präfix (zum Beispiel `x.html5`) im Auswahlmenü alle Templates angezeigt werden.
|
defect
|
leeres auswahlmenü für ursprungstemplate bei vergleich von templates mit abweichendem präfix beim vergleich von templates ist das auswahlmenü für das ursprungstemplate leer wenn das zu vergleichende template ein abweichendes präfix hat zum beispiel x x dadurch ist es nicht möglich einen vergleich durchzuführen entweder könnte der button zur vergleichsansicht in diesem fall deaktiviert werden oder es könnten wie bei templates ohne präfix zum beispiel x im auswahlmenü alle templates angezeigt werden
| 1
|
67,231
| 20,961,586,897
|
IssuesEvent
|
2022-03-27 21:44:31
|
abedmaatalla/foursquared
|
https://api.github.com/repos/abedmaatalla/foursquared
|
closed
|
Update widget assets for v1 release.
|
Priority-Medium Type-Defect auto-migrated Component-UI
|
```
1. widget_header.png needs 9-patchification, on the gradient section so
that it will stretch and contract horizontally.
2. widget_header needs the standard 3 state button imagery, per
http://d.android.com/guide/practices/ui_guidelines/widget_design.html#desig
n bullet point #4.
3. widget_footer.png needs 9-patchification, on the mockup's semi-
transparent background.
4. widget_footer needs the standard 3 state button imagery, per
http://d.android.com/guide/practices/ui_guidelines/widget_design.html#desig
n bullet point #4.
5. both assets should be no wider than 220 pixels as to fit in a 3 cell
wide widget on the G1
I think these are the only things we'll need for art to get v1 out the
door, thats why its so specific.
```
Original issue reported on code.google.com by `jlapenna` on 3 Nov 2009 at 6:40
- Blocking: #28
|
1.0
|
Update widget assets for v1 release. - ```
1. widget_header.png needs 9-patchification, on the gradient section so
that it will stretch and contract horizontally.
2. widget_header needs the standard 3 state button imagery, per
http://d.android.com/guide/practices/ui_guidelines/widget_design.html#desig
n bullet point #4.
3. widget_footer.png needs 9-patchification, on the mockup's semi-
transparent background.
4. widget_footer needs the standard 3 state button imagery, per
http://d.android.com/guide/practices/ui_guidelines/widget_design.html#desig
n bullet point #4.
5. both assets should be no wider than 220 pixels as to fit in a 3 cell
wide widget on the G1
I think these are the only things we'll need for art to get v1 out the
door, thats why its so specific.
```
Original issue reported on code.google.com by `jlapenna` on 3 Nov 2009 at 6:40
- Blocking: #28
|
defect
|
update widget assets for release widget header png needs patchification on the gradient section so that it will stretch and contract horizontally widget header needs the standard state button imagery per n bullet point widget footer png needs patchification on the mockup s semi transparent background widget footer needs the standard state button imagery per n bullet point both assets should be no wider than pixels as to fit in a cell wide widget on the i think these are the only things we ll need for art to get out the door thats why its so specific original issue reported on code google com by jlapenna on nov at blocking
| 1
|
316,598
| 9,652,191,818
|
IssuesEvent
|
2019-05-18 15:02:55
|
servicemesher/istio-official-translation
|
https://api.github.com/repos/servicemesher/istio-official-translation
|
closed
|
/blog/2018/egress-https/index.md
|
finished lang/zh priority/P0 sync/update version/1.1
|
Source File: [/blog/2018/egress-https/index.md](https://github.com/istio/istio.io/tree/release-1.1/content/blog/2018/egress-https/index.md)
Diff:
~~~diff
diff --git a/content/blog/2018/egress-https/index.md b/content/blog/2018/egress-https/index.md
index b33b6538..8ebc0b0d 100644
--- a/content/blog/2018/egress-https/index.md
+++ b/content/blog/2018/egress-https/index.md
@@ -44,10 +44,9 @@ Here is a copy of the end-to-end architecture of the application from the origin
Perform the steps in the
[Deploying the application](/docs/examples/bookinfo/#deploying-the-application),
-[Confirm the app is running](/docs/examples/bookinfo/#confirm-the-app-is-accessible-from-outside-the-cluster),
+[Confirm the app is running](/docs/examples/bookinfo/#confirm-the-app-is-accessible-from-outside-the-cluster), and
[Apply default destination rules](/docs/examples/bookinfo/#apply-default-destination-rules)
-sections, and
-[change Istio to the blocking-egress-by-default policy](/docs/tasks/traffic-management/egress/#change-to-the-blocking-by-default-policy).
+sections.
## Bookinfo with HTTPS access to a Google Books web service
@@ -212,14 +211,15 @@ Here is how both patterns are supported in the
{{< text ruby >}}
uri = URI.parse('https://www.googleapis.com/books/v1/volumes?q=isbn:' + isbn)
-http = Net::HTTP.new(uri.host, ENV['DO_NOT_ENCRYPT'] === 'true' ? 80:443)
+http = Net::HTTP.new(uri.host, uri.port)
...
unless ENV['DO_NOT_ENCRYPT'] === 'true' then
http.use_ssl = true
end
{{< /text >}}
-When the `DO_NOT_ENCRYPT` environment variable is defined, the request is performed without SSL (plain HTTP) to port 80.
+Note that the port is derived by the `URI.parse` from the URI's schema (`https://`) to be `443`, the default HTTPS port.
+When the `DO_NOT_ENCRYPT` environment variable is defined, the request is performed without SSL (plain HTTP).
You can set the `DO_NOT_ENCRYPT` environment variable to _"true"_ in the
[Kubernetes deployment spec of details v2]({{< github_file >}}/samples/bookinfo/platform/kube/bookinfo-details-v2.yaml),
@@ -250,8 +250,7 @@ In the next section you will configure TLS origination for accessing an external
$ kubectl apply -f @samples/bookinfo/networking/virtual-service-details-v2.yaml@
{{< /text >}}
-1. Create a mesh-external service entry for `www.google.apis` , a virtual service to rewrite the destination port from
- 80 to 443, and a destination rule to perform TLS origination.
+1. Create a mesh-external service entry for `www.google.apis` and a destination rule to perform TLS origination.
{{< text bash >}}
$ kubectl apply -f - <<EOF
@@ -263,31 +262,12 @@ In the next section you will configure TLS origination for accessing an external
hosts:
- www.googleapis.com
ports:
- - number: 80
- name: http
- protocol: HTTP
- number: 443
- name: https
- protocol: HTTPS
+ name: http-port-for-tls-origination
+ protocol: HTTP
resolution: DNS
---
apiVersion: networking.istio.io/v1alpha3
- kind: VirtualService
- metadata:
- name: rewrite-port-for-googleapis
- spec:
- hosts:
- - www.googleapis.com
- http:
- - match:
- - port: 80
- route:
- - destination:
- host: www.googleapis.com
- port:
- number: 443
- ---
- apiVersion: networking.istio.io/v1alpha3
kind: DestinationRule
metadata:
name: originate-tls-for-googleapis
@@ -300,19 +280,22 @@ In the next section you will configure TLS origination for accessing an external
- port:
number: 443
tls:
- mode: SIMPLE # initiates HTTPS when accessing www.googleapis.com
+ mode: SIMPLE # initiates HTTPS when accessing edition.cnn.com
EOF
{{< /text >}}
-1. Access the web page of the application and verify that the book details are displayed without errors.
+ Note that port `443` is designated by a name with the prefix `http-`, and its protocol is specified as `HTTP`. Note
+ that you are not required to use port 443 to send HTTP requests for TLS origination.
+ [This example](/docs/examples/advanced-gateways/egress-tls-origination/) shows how to perform TLS
+ origination with port rewriting.
-1. [Enable Envoy’s access logging](/docs/tasks/telemetry/logs/access-log/#enable-envoy-s-access-logging)
+1. Access the web page of the application and verify that the book details are displayed without errors.
1. Check the log of of the sidecar proxy of _details v2_ and see the HTTP request.
{{< text bash >}}
$ kubectl logs $(kubectl get pods -l app=details -l version=v2 -o jsonpath='{.items[0].metadata.name}') istio-proxy | grep googleapis
- [2018-08-09T11:32:58.171Z] "GET /books/v1/volumes?q=isbn:0486424618 HTTP/1.1" 200 - 0 1050 264 264 "-" "Ruby" "b993bae7-4288-9241-81a5-4cde93b2e3a6" "www.googleapis.com:80" "172.217.20.74:80"
+ [2018-08-09T11:32:58.171Z] "GET /books/v1/volumes?q=isbn:0486424618 HTTP/1.1" 200 - 0 1050 264 264 "-" "Ruby" "b993bae7-4288-9241-81a5-4cde93b2e3a6" "www.googleapis.com:443" "172.217.20.74:443"
EOF
{{< /text >}}
@@ -323,7 +306,6 @@ In the next section you will configure TLS origination for accessing an external
{{< text bash >}}
$ kubectl delete serviceentry googleapis
-$ kubectl delete virtualservice rewrite-port-for-googleapis
$ kubectl delete destinationrule originate-tls-for-googleapis
$ kubectl delete -f @samples/bookinfo/networking/virtual-service-details-v2.yaml@
$ kubectl delete -f @samples/bookinfo/platform/kube/bookinfo-details-v2.yaml@
~~~
|
1.0
|
/blog/2018/egress-https/index.md - Source File: [/blog/2018/egress-https/index.md](https://github.com/istio/istio.io/tree/release-1.1/content/blog/2018/egress-https/index.md)
Diff:
~~~diff
diff --git a/content/blog/2018/egress-https/index.md b/content/blog/2018/egress-https/index.md
index b33b6538..8ebc0b0d 100644
--- a/content/blog/2018/egress-https/index.md
+++ b/content/blog/2018/egress-https/index.md
@@ -44,10 +44,9 @@ Here is a copy of the end-to-end architecture of the application from the origin
Perform the steps in the
[Deploying the application](/docs/examples/bookinfo/#deploying-the-application),
-[Confirm the app is running](/docs/examples/bookinfo/#confirm-the-app-is-accessible-from-outside-the-cluster),
+[Confirm the app is running](/docs/examples/bookinfo/#confirm-the-app-is-accessible-from-outside-the-cluster), and
[Apply default destination rules](/docs/examples/bookinfo/#apply-default-destination-rules)
-sections, and
-[change Istio to the blocking-egress-by-default policy](/docs/tasks/traffic-management/egress/#change-to-the-blocking-by-default-policy).
+sections.
## Bookinfo with HTTPS access to a Google Books web service
@@ -212,14 +211,15 @@ Here is how both patterns are supported in the
{{< text ruby >}}
uri = URI.parse('https://www.googleapis.com/books/v1/volumes?q=isbn:' + isbn)
-http = Net::HTTP.new(uri.host, ENV['DO_NOT_ENCRYPT'] === 'true' ? 80:443)
+http = Net::HTTP.new(uri.host, uri.port)
...
unless ENV['DO_NOT_ENCRYPT'] === 'true' then
http.use_ssl = true
end
{{< /text >}}
-When the `DO_NOT_ENCRYPT` environment variable is defined, the request is performed without SSL (plain HTTP) to port 80.
+Note that the port is derived by the `URI.parse` from the URI's schema (`https://`) to be `443`, the default HTTPS port.
+When the `DO_NOT_ENCRYPT` environment variable is defined, the request is performed without SSL (plain HTTP).
You can set the `DO_NOT_ENCRYPT` environment variable to _"true"_ in the
[Kubernetes deployment spec of details v2]({{< github_file >}}/samples/bookinfo/platform/kube/bookinfo-details-v2.yaml),
@@ -250,8 +250,7 @@ In the next section you will configure TLS origination for accessing an external
$ kubectl apply -f @samples/bookinfo/networking/virtual-service-details-v2.yaml@
{{< /text >}}
-1. Create a mesh-external service entry for `www.google.apis` , a virtual service to rewrite the destination port from
- 80 to 443, and a destination rule to perform TLS origination.
+1. Create a mesh-external service entry for `www.google.apis` and a destination rule to perform TLS origination.
{{< text bash >}}
$ kubectl apply -f - <<EOF
@@ -263,31 +262,12 @@ In the next section you will configure TLS origination for accessing an external
hosts:
- www.googleapis.com
ports:
- - number: 80
- name: http
- protocol: HTTP
- number: 443
- name: https
- protocol: HTTPS
+ name: http-port-for-tls-origination
+ protocol: HTTP
resolution: DNS
---
apiVersion: networking.istio.io/v1alpha3
- kind: VirtualService
- metadata:
- name: rewrite-port-for-googleapis
- spec:
- hosts:
- - www.googleapis.com
- http:
- - match:
- - port: 80
- route:
- - destination:
- host: www.googleapis.com
- port:
- number: 443
- ---
- apiVersion: networking.istio.io/v1alpha3
kind: DestinationRule
metadata:
name: originate-tls-for-googleapis
@@ -300,19 +280,22 @@ In the next section you will configure TLS origination for accessing an external
- port:
number: 443
tls:
- mode: SIMPLE # initiates HTTPS when accessing www.googleapis.com
+ mode: SIMPLE # initiates HTTPS when accessing edition.cnn.com
EOF
{{< /text >}}
-1. Access the web page of the application and verify that the book details are displayed without errors.
+ Note that port `443` is designated by a name with the prefix `http-`, and its protocol is specified as `HTTP`. Note
+ that you are not required to use port 443 to send HTTP requests for TLS origination.
+ [This example](/docs/examples/advanced-gateways/egress-tls-origination/) shows how to perform TLS
+ origination with port rewriting.
-1. [Enable Envoy’s access logging](/docs/tasks/telemetry/logs/access-log/#enable-envoy-s-access-logging)
+1. Access the web page of the application and verify that the book details are displayed without errors.
1. Check the log of of the sidecar proxy of _details v2_ and see the HTTP request.
{{< text bash >}}
$ kubectl logs $(kubectl get pods -l app=details -l version=v2 -o jsonpath='{.items[0].metadata.name}') istio-proxy | grep googleapis
- [2018-08-09T11:32:58.171Z] "GET /books/v1/volumes?q=isbn:0486424618 HTTP/1.1" 200 - 0 1050 264 264 "-" "Ruby" "b993bae7-4288-9241-81a5-4cde93b2e3a6" "www.googleapis.com:80" "172.217.20.74:80"
+ [2018-08-09T11:32:58.171Z] "GET /books/v1/volumes?q=isbn:0486424618 HTTP/1.1" 200 - 0 1050 264 264 "-" "Ruby" "b993bae7-4288-9241-81a5-4cde93b2e3a6" "www.googleapis.com:443" "172.217.20.74:443"
EOF
{{< /text >}}
@@ -323,7 +306,6 @@ In the next section you will configure TLS origination for accessing an external
{{< text bash >}}
$ kubectl delete serviceentry googleapis
-$ kubectl delete virtualservice rewrite-port-for-googleapis
$ kubectl delete destinationrule originate-tls-for-googleapis
$ kubectl delete -f @samples/bookinfo/networking/virtual-service-details-v2.yaml@
$ kubectl delete -f @samples/bookinfo/platform/kube/bookinfo-details-v2.yaml@
~~~
|
non_defect
|
blog egress https index md source file diff diff diff git a content blog egress https index md b content blog egress https index md index a content blog egress https index md b content blog egress https index md here is a copy of the end to end architecture of the application from the origin perform the steps in the docs examples bookinfo deploying the application docs examples bookinfo confirm the app is accessible from outside the cluster docs examples bookinfo confirm the app is accessible from outside the cluster and docs examples bookinfo apply default destination rules sections and docs tasks traffic management egress change to the blocking by default policy sections bookinfo with https access to a google books web service here is how both patterns are supported in the uri uri parse isbn http net http new uri host env true http net http new uri host uri port unless env true then http use ssl true end when the do not encrypt environment variable is defined the request is performed without ssl plain http to port note that the port is derived by the uri parse from the uri s schema to be the default https port when the do not encrypt environment variable is defined the request is performed without ssl plain http you can set the do not encrypt environment variable to true in the samples bookinfo platform kube bookinfo details yaml in the next section you will configure tls origination for accessing an external kubectl apply f samples bookinfo networking virtual service details yaml create a mesh external service entry for a virtual service to rewrite the destination port from to and a destination rule to perform tls origination create a mesh external service entry for and a destination rule to perform tls origination kubectl apply f eof in the next section you will configure tls origination for accessing an external hosts ports number name http protocol http number name https protocol https name http port for tls origination protocol http resolution dns apiversion networking istio io kind virtualservice metadata name rewrite port for googleapis spec hosts http match port route destination host port number apiversion networking istio io kind destinationrule metadata name originate tls for googleapis in the next section you will configure tls origination for accessing an external port number tls mode simple initiates https when accessing mode simple initiates https when accessing edition cnn com eof access the web page of the application and verify that the book details are displayed without errors note that port is designated by a name with the prefix http and its protocol is specified as http note that you are not required to use port to send http requests for tls origination docs examples advanced gateways egress tls origination shows how to perform tls origination with port rewriting docs tasks telemetry logs access log enable envoy s access logging access the web page of the application and verify that the book details are displayed without errors check the log of of the sidecar proxy of details and see the http request kubectl logs kubectl get pods l app details l version o jsonpath items metadata name istio proxy grep googleapis get books volumes q isbn http ruby get books volumes q isbn http ruby eof in the next section you will configure tls origination for accessing an external kubectl delete serviceentry googleapis kubectl delete virtualservice rewrite port for googleapis kubectl delete destinationrule originate tls for googleapis kubectl delete f samples bookinfo networking virtual service details yaml kubectl delete f samples bookinfo platform kube bookinfo details yaml
| 0
|
390,506
| 11,544,452,740
|
IssuesEvent
|
2020-02-18 11:30:13
|
grafana/grafana
|
https://api.github.com/repos/grafana/grafana
|
closed
|
calendar sometimes closes itself when using new timepicker
|
area/frontend needs investigation priority/important-soon type/bug
|
**What happened**:
When we select dates in Absolute time range new calendar, it sometimes closes itself and does not validate selected start & end dates.
**What you expected to happen**:
After selection, see selected period in blue as expected, and be able to change it if needed before validate.
**How to reproduce it (as minimally and precisely as possible)**:
Just select a first date on calendar, blue circle appears around date, and then select a second date. Randomly, calendar closes itself and we have to start again.
**Anything else we need to know?**: -
**Environment**:
- Grafana version: 6.6.1
- Data source type & version: postgres
- OS Grafana is installed on: debian 9.11
- User OS & Browser: chrome 79.0.3945.88 on debian
- Grafana plugins: -
- Others: -
|
1.0
|
calendar sometimes closes itself when using new timepicker - **What happened**:
When we select dates in Absolute time range new calendar, it sometimes closes itself and does not validate selected start & end dates.
**What you expected to happen**:
After selection, see selected period in blue as expected, and be able to change it if needed before validate.
**How to reproduce it (as minimally and precisely as possible)**:
Just select a first date on calendar, blue circle appears around date, and then select a second date. Randomly, calendar closes itself and we have to start again.
**Anything else we need to know?**: -
**Environment**:
- Grafana version: 6.6.1
- Data source type & version: postgres
- OS Grafana is installed on: debian 9.11
- User OS & Browser: chrome 79.0.3945.88 on debian
- Grafana plugins: -
- Others: -
|
non_defect
|
calendar sometimes closes itself when using new timepicker what happened when we select dates in absolute time range new calendar it sometimes closes itself and does not validate selected start end dates what you expected to happen after selection see selected period in blue as expected and be able to change it if needed before validate how to reproduce it as minimally and precisely as possible just select a first date on calendar blue circle appears around date and then select a second date randomly calendar closes itself and we have to start again anything else we need to know environment grafana version data source type version postgres os grafana is installed on debian user os browser chrome on debian grafana plugins others
| 0
|
38,576
| 8,915,069,857
|
IssuesEvent
|
2019-01-19 01:38:39
|
abulka/pynsource
|
https://api.github.com/repos/abulka/pynsource
|
closed
|
Lines should be able to be deleted
|
Priority-Medium Type-Defect auto-migrated enhancement
|
```
What steps will reproduce the problem?
1.Create two classes;
2.Create a line between them;
3.Try to delete the line.
There is no option for deleting lines.
1.61 Windows 7 64-bit
```
Original issue reported on code.google.com by `cristian...@yahoo.com` on 21 Jul 2014 at 11:18
|
1.0
|
Lines should be able to be deleted - ```
What steps will reproduce the problem?
1.Create two classes;
2.Create a line between them;
3.Try to delete the line.
There is no option for deleting lines.
1.61 Windows 7 64-bit
```
Original issue reported on code.google.com by `cristian...@yahoo.com` on 21 Jul 2014 at 11:18
|
defect
|
lines should be able to be deleted what steps will reproduce the problem create two classes create a line between them try to delete the line there is no option for deleting lines windows bit original issue reported on code google com by cristian yahoo com on jul at
| 1
|
65,900
| 19,770,619,089
|
IssuesEvent
|
2022-01-17 09:39:46
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: scipy.optimize.ridder() does not obey absolute tolerance
|
defect
|
### Describe your issue.
Hi,
I would like to use `scipy.optimize.ridder()` for a function of me.
I set the absolute tolerance to `1.` and expect that the resulting function call would be in the range `[-1.:+1.]`.
When looking at the second call to `ridder()`, I can see `5.000065000084806` as a result.
Did I make a mistake when calling `ridder()` or is something wrong with the result of it?
Thanks in advance!
### Reproducing Code Example
```python
import scipy.optimize as sco
import numpy as np
# import scipy
import sys, scipy, numpy;
print(scipy.__version__, numpy.__version__, sys.version_info)
amountOfCalls = 0
def myFunc(x):
global amountOfCalls
amountOfCalls += 1
y = x ** 3 + x - 5
return y
for xtol in [0.00001, 1., ]:
amountOfCalls = 0
xValue, root1_result = sco.ridder(myFunc, -1000, 1000, xtol=xtol, full_output=True, disp=True, )
print('atol:', xtol, 'nullvalue:', myFunc(xValue), 'xValue:', xValue, 'amountOfCalls:', amountOfCalls) # , ' '.join(str(root1_result).split('\n')))
```
### Error message
```shell
1.2.2 1.17.0 sys.version_info(major=3, minor=6, micro=8, releaselevel='final', serial=0)
atol: 1e-05 nullvalue: 3.9394811365944804e-05 xValue: 1.515985217781899 amountOfCalls: 31
atol: 1.0 nullvalue: 5.000065000084806 xValue: 2.000004999994985 amountOfCalls: 13
```
### SciPy/NumPy/Python version information
1.2.2 1.17.0 sys.version_info(major=3, minor=6, micro=8, releaselevel='final', serial=0)
|
1.0
|
BUG: scipy.optimize.ridder() does not obey absolute tolerance - ### Describe your issue.
Hi,
I would like to use `scipy.optimize.ridder()` for a function of me.
I set the absolute tolerance to `1.` and expect that the resulting function call would be in the range `[-1.:+1.]`.
When looking at the second call to `ridder()`, I can see `5.000065000084806` as a result.
Did I make a mistake when calling `ridder()` or is something wrong with the result of it?
Thanks in advance!
### Reproducing Code Example
```python
import scipy.optimize as sco
import numpy as np
# import scipy
import sys, scipy, numpy;
print(scipy.__version__, numpy.__version__, sys.version_info)
amountOfCalls = 0
def myFunc(x):
global amountOfCalls
amountOfCalls += 1
y = x ** 3 + x - 5
return y
for xtol in [0.00001, 1., ]:
amountOfCalls = 0
xValue, root1_result = sco.ridder(myFunc, -1000, 1000, xtol=xtol, full_output=True, disp=True, )
print('atol:', xtol, 'nullvalue:', myFunc(xValue), 'xValue:', xValue, 'amountOfCalls:', amountOfCalls) # , ' '.join(str(root1_result).split('\n')))
```
### Error message
```shell
1.2.2 1.17.0 sys.version_info(major=3, minor=6, micro=8, releaselevel='final', serial=0)
atol: 1e-05 nullvalue: 3.9394811365944804e-05 xValue: 1.515985217781899 amountOfCalls: 31
atol: 1.0 nullvalue: 5.000065000084806 xValue: 2.000004999994985 amountOfCalls: 13
```
### SciPy/NumPy/Python version information
1.2.2 1.17.0 sys.version_info(major=3, minor=6, micro=8, releaselevel='final', serial=0)
|
defect
|
bug scipy optimize ridder does not obey absolute tolerance describe your issue hi i would like to use scipy optimize ridder for a function of me i set the absolute tolerance to and expect that the resulting function call would be in the range when looking at the second call to ridder i can see as a result did i make a mistake when calling ridder or is something wrong with the result of it thanks in advance reproducing code example python import scipy optimize as sco import numpy as np import scipy import sys scipy numpy print scipy version numpy version sys version info amountofcalls def myfunc x global amountofcalls amountofcalls y x x return y for xtol in amountofcalls xvalue result sco ridder myfunc xtol xtol full output true disp true print atol xtol nullvalue myfunc xvalue xvalue xvalue amountofcalls amountofcalls join str result split n error message shell sys version info major minor micro releaselevel final serial atol nullvalue xvalue amountofcalls atol nullvalue xvalue amountofcalls scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
119,017
| 4,759,497,370
|
IssuesEvent
|
2016-10-24 22:48:29
|
kdahlquist/GRNmap
|
https://api.github.com/repos/kdahlquist/GRNmap
|
closed
|
Data Analysis Team Tasks for Week of 10/17
|
data analysis logistics priority 0
|
The tasks for this week:
* @kdahlquist has "review requested" for
- [ ] #230 (degree distribution charts)
- [ ] #245, generating and vetting input workbooks (after requested changes are made)
* @bklein7 and @Nwilli31 are working on the following:
- [ ] #245, generating and vetting input workbooks, finishing off requested changes
- [ ] #241, update 16_tests workbooks on beta branch to be consistent with our formatting guidelines
- [ ] start some model runs (details to come)
* @khorstmann and @maggie-oneil are working on the following:
- [ ] #259 @maggie-oneil and @khorstmann will work on getting the SBET toolbox up and running, starting with betweenness centrality.
* Everyone is working on TRACE documentation #170
- [ ] @khorstmann and @maggie-oneil will take first four
- [ ] @bklein7 and @Nwilli31 will take second four.
|
1.0
|
Data Analysis Team Tasks for Week of 10/17 - The tasks for this week:
* @kdahlquist has "review requested" for
- [ ] #230 (degree distribution charts)
- [ ] #245, generating and vetting input workbooks (after requested changes are made)
* @bklein7 and @Nwilli31 are working on the following:
- [ ] #245, generating and vetting input workbooks, finishing off requested changes
- [ ] #241, update 16_tests workbooks on beta branch to be consistent with our formatting guidelines
- [ ] start some model runs (details to come)
* @khorstmann and @maggie-oneil are working on the following:
- [ ] #259 @maggie-oneil and @khorstmann will work on getting the SBET toolbox up and running, starting with betweenness centrality.
* Everyone is working on TRACE documentation #170
- [ ] @khorstmann and @maggie-oneil will take first four
- [ ] @bklein7 and @Nwilli31 will take second four.
|
non_defect
|
data analysis team tasks for week of the tasks for this week kdahlquist has review requested for degree distribution charts generating and vetting input workbooks after requested changes are made and are working on the following generating and vetting input workbooks finishing off requested changes update tests workbooks on beta branch to be consistent with our formatting guidelines start some model runs details to come khorstmann and maggie oneil are working on the following maggie oneil and khorstmann will work on getting the sbet toolbox up and running starting with betweenness centrality everyone is working on trace documentation khorstmann and maggie oneil will take first four and will take second four
| 0
|
443,569
| 12,796,238,946
|
IssuesEvent
|
2020-07-02 10:05:31
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
opened
|
[Migration-3.2.0] Application Key generation page cannot view
|
Priority/Normal Type/Bug
|
### Description:
Migrate from 3.0.0 ---> 3.2.0
Redirect to application page.
Select Old app
Redirect to jet generate page
Error: blank page
Console error:

selectedGrantTypes become null.
Following the line of code:
https://github.com/wso2/carbon-apimgt/blob/master/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/AbstractAPIManager.java#L3141
apiKey.setGrantTypes(
(String) oAuthApplicationInfo.getParameter(APIConstants.JSON_GRANT_TYPES));
|
1.0
|
[Migration-3.2.0] Application Key generation page cannot view - ### Description:
Migrate from 3.0.0 ---> 3.2.0
Redirect to application page.
Select Old app
Redirect to jet generate page
Error: blank page
Console error:

selectedGrantTypes become null.
Following the line of code:
https://github.com/wso2/carbon-apimgt/blob/master/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/AbstractAPIManager.java#L3141
apiKey.setGrantTypes(
(String) oAuthApplicationInfo.getParameter(APIConstants.JSON_GRANT_TYPES));
|
non_defect
|
application key generation page cannot view description migrate from redirect to application page select old app redirect to jet generate page error blank page console error selectedgranttypes become null following the line of code apikey setgranttypes string oauthapplicationinfo getparameter apiconstants json grant types
| 0
|
2,369
| 2,607,898,411
|
IssuesEvent
|
2015-02-26 00:12:27
|
chrsmithdemos/zen-coding
|
https://api.github.com/repos/chrsmithdemos/zen-coding
|
closed
|
Поддержка скобок
|
auto-migrated Milestone-0.6 Priority-Medium Type-Defect
|
```
Очень хотелось бы поддержку скобок.
Чтобы можно было бы писать *form>(label+input:text)`*`2*
Так же это незаменимо для всяческих меню,
где выражение *nav>ul>
(li>a)`*`10* может значительно сэкономить время
по сравнению с ручной
вставкой тега A в каждый из созданных
элементов списка.
```
-----
Original issue reported on code.google.com by `alexey.kvirc@googlemail.com` on 24 Nov 2009 at 10:44
|
1.0
|
Поддержка скобок - ```
Очень хотелось бы поддержку скобок.
Чтобы можно было бы писать *form>(label+input:text)`*`2*
Так же это незаменимо для всяческих меню,
где выражение *nav>ul>
(li>a)`*`10* может значительно сэкономить время
по сравнению с ручной
вставкой тега A в каждый из созданных
элементов списка.
```
-----
Original issue reported on code.google.com by `alexey.kvirc@googlemail.com` on 24 Nov 2009 at 10:44
|
defect
|
поддержка скобок очень хотелось бы поддержку скобок чтобы можно было бы писать form label input text так же это незаменимо для всяческих меню где выражение nav ul li a может значительно сэкономить время по сравнению с ручной вставкой тега a в каждый из созданных элементов списка original issue reported on code google com by alexey kvirc googlemail com on nov at
| 1
|
31,159
| 6,443,899,794
|
IssuesEvent
|
2017-08-12 02:18:04
|
opendatakit/opendatakit
|
https://api.github.com/repos/opendatakit/opendatakit
|
closed
|
MiscTasks scanned frequently to see if form is being deleted.
|
Aggregate Priority-Medium Type-Defect
|
Originally reported on Google Code with ID 588
```
The check for form-deletion causes the misc tasks table to be scanned a lot.
Should perhaps handle this differently? e.g., would be nice if FormInfo had a flag
indicating whether the form was marked for deletion. (v2.0?).
Performance could be impacted if you are deleting forms very frequently.
```
Reported by `mitchellsundt` on 2012-06-04 19:30:40
|
1.0
|
MiscTasks scanned frequently to see if form is being deleted. - Originally reported on Google Code with ID 588
```
The check for form-deletion causes the misc tasks table to be scanned a lot.
Should perhaps handle this differently? e.g., would be nice if FormInfo had a flag
indicating whether the form was marked for deletion. (v2.0?).
Performance could be impacted if you are deleting forms very frequently.
```
Reported by `mitchellsundt` on 2012-06-04 19:30:40
|
defect
|
misctasks scanned frequently to see if form is being deleted originally reported on google code with id the check for form deletion causes the misc tasks table to be scanned a lot should perhaps handle this differently e g would be nice if forminfo had a flag indicating whether the form was marked for deletion performance could be impacted if you are deleting forms very frequently reported by mitchellsundt on
| 1
|
16,690
| 2,933,994,135
|
IssuesEvent
|
2015-06-30 04:30:14
|
prettydiff/prettydiff
|
https://api.github.com/repos/prettydiff/prettydiff
|
closed
|
Empty li mess up indentation
|
Defect Underway
|
Empty `li` are replaced with their self closing tags, but mess up indentation.
Input
```
<ul>
<li>something</li>
<li></li>
<li>something</li>
<li></li>
<li>something</li>
</ul>
```
Output
```
<ul>
<li>something</li>
<li/>
<li>
something</li><li/>
<li>something
</li>
</ul>
```
|
1.0
|
Empty li mess up indentation - Empty `li` are replaced with their self closing tags, but mess up indentation.
Input
```
<ul>
<li>something</li>
<li></li>
<li>something</li>
<li></li>
<li>something</li>
</ul>
```
Output
```
<ul>
<li>something</li>
<li/>
<li>
something</li><li/>
<li>something
</li>
</ul>
```
|
defect
|
empty li mess up indentation empty li are replaced with their self closing tags but mess up indentation input something something something output something something something
| 1
|
54,515
| 7,891,019,381
|
IssuesEvent
|
2018-06-28 10:39:04
|
i18next/react-i18next
|
https://api.github.com/repos/i18next/react-i18next
|
closed
|
I18next with a help of new react context
|
documentation
|
Hello there!
I have created an alternative way to use i18next in react with a help of context. Here is my lib - https://github.com/dmtrKovalenko/ns-react-18next.
Maybe you will post the link for it from the docs, or it can be merged to `react-i18next` if its convenient. Thank you)
|
1.0
|
I18next with a help of new react context - Hello there!
I have created an alternative way to use i18next in react with a help of context. Here is my lib - https://github.com/dmtrKovalenko/ns-react-18next.
Maybe you will post the link for it from the docs, or it can be merged to `react-i18next` if its convenient. Thank you)
|
non_defect
|
with a help of new react context hello there i have created an alternative way to use in react with a help of context here is my lib maybe you will post the link for it from the docs or it can be merged to react if its convenient thank you
| 0
|
506,901
| 14,675,491,073
|
IssuesEvent
|
2020-12-30 17:43:52
|
aims-group/metagrid
|
https://api.github.com/repos/aims-group/metagrid
|
closed
|
Non-existent routes should redirect to the homepage or show a 404
|
Platform: React Priority: Medium Type: Enhancement
|
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
React router should be configured so that non-existent routes will show a 404 with a button to return to the homepage.
|
1.0
|
Non-existent routes should redirect to the homepage or show a 404 - **Describe the solution you'd like**
A clear and concise description of what you want to happen.
React router should be configured so that non-existent routes will show a 404 with a button to return to the homepage.
|
non_defect
|
non existent routes should redirect to the homepage or show a describe the solution you d like a clear and concise description of what you want to happen react router should be configured so that non existent routes will show a with a button to return to the homepage
| 0
|
11,656
| 2,660,024,766
|
IssuesEvent
|
2015-03-19 01:42:42
|
perfsonar/project
|
https://api.github.com/repos/perfsonar/project
|
closed
|
I2Addr functions block contacting DNS server even when node name not requested
|
Milestone-Release3.5 Priority-Medium Type-Defect
|
Original [issue 1055](https://code.google.com/p/perfsonar-ps/issues/detail?id=1055) created by arlake228 on 2015-01-26T10:15:28.000Z:
<b>What steps will reproduce the problem?</b>
1. I2Addr functions block contacting DNS server even when node name not requested
2. owamp uses I2Addr functions for test sessions, but never requests the name of the peer node. Therefore, this results in unnecessary delays if there is any delay in contacting the DNS server.
<b>What is the expected output? What do you see instead?</b>
No delay in using I2Addr APIs when not requesting peer node name.
<b>What version of the product are you using? On what operating system?</b>
1.2-2 on Debian Linux
<b>Please provide any additional information below.</b>
Commits for tests and fixes available here:
https://code.google.com/r/rshearma-i2util/source/detail?r=3068e9290d34a09491d54dbb87033cc3c3931c61&name=addr-rdns
https://code.google.com/r/rshearma-i2util/source/detail?r=c43d3b2243e1da16fb6a38cb3d7f63aa16626c63&name=addr-rdns
|
1.0
|
I2Addr functions block contacting DNS server even when node name not requested - Original [issue 1055](https://code.google.com/p/perfsonar-ps/issues/detail?id=1055) created by arlake228 on 2015-01-26T10:15:28.000Z:
<b>What steps will reproduce the problem?</b>
1. I2Addr functions block contacting DNS server even when node name not requested
2. owamp uses I2Addr functions for test sessions, but never requests the name of the peer node. Therefore, this results in unnecessary delays if there is any delay in contacting the DNS server.
<b>What is the expected output? What do you see instead?</b>
No delay in using I2Addr APIs when not requesting peer node name.
<b>What version of the product are you using? On what operating system?</b>
1.2-2 on Debian Linux
<b>Please provide any additional information below.</b>
Commits for tests and fixes available here:
https://code.google.com/r/rshearma-i2util/source/detail?r=3068e9290d34a09491d54dbb87033cc3c3931c61&name=addr-rdns
https://code.google.com/r/rshearma-i2util/source/detail?r=c43d3b2243e1da16fb6a38cb3d7f63aa16626c63&name=addr-rdns
|
defect
|
functions block contacting dns server even when node name not requested original created by on what steps will reproduce the problem functions block contacting dns server even when node name not requested owamp uses functions for test sessions but never requests the name of the peer node therefore this results in unnecessary delays if there is any delay in contacting the dns server what is the expected output what do you see instead no delay in using apis when not requesting peer node name what version of the product are you using on what operating system on debian linux please provide any additional information below commits for tests and fixes available here
| 1
|
514,582
| 14,941,164,289
|
IssuesEvent
|
2021-01-25 19:20:56
|
Plaxy-Technologies-Inc/YouPlanets-Bug-Report
|
https://api.github.com/repos/Plaxy-Technologies-Inc/YouPlanets-Bug-Report
|
opened
|
An issue on post editor
|
Priority: Medium bug
|
**Describe the bug**
Bullet point effect does not apply to a post content after posted.
**Screenshots**
<img width="1170" alt="image" src="https://user-images.githubusercontent.com/40333215/105754645-541d6b00-5f18-11eb-9e19-f6e6e333f93b.png">
<img width="992" alt="image" src="https://user-images.githubusercontent.com/40333215/105754671-58e21f00-5f18-11eb-826b-38e79c8860d9.png">
|
1.0
|
An issue on post editor - **Describe the bug**
Bullet point effect does not apply to a post content after posted.
**Screenshots**
<img width="1170" alt="image" src="https://user-images.githubusercontent.com/40333215/105754645-541d6b00-5f18-11eb-9e19-f6e6e333f93b.png">
<img width="992" alt="image" src="https://user-images.githubusercontent.com/40333215/105754671-58e21f00-5f18-11eb-826b-38e79c8860d9.png">
|
non_defect
|
an issue on post editor describe the bug bullet point effect does not apply to a post content after posted screenshots img width alt image src img width alt image src
| 0
|
246,893
| 20,924,189,743
|
IssuesEvent
|
2022-03-24 20:37:19
|
Uuvana-Studios/longvinter-windows-client
|
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
|
closed
|
Bugged container in my new tent
|
bug Not Tested
|
**Describe the bug**
I used to have another tent but it got deleted (i'm pretty sure it's not from the timer)
i recently got another tent and when i entered it for the first time some of my old stuff was there including a radiator, a workbench (out of bounds) and a container in the middle of the room that had some of my old stuff
i cannot remove this container and i'm pretty sure that when trying to remove it it dupes itself by dropping but not disapearing (the item is stuck under it so not exploiting by taking 'em)
**To Reproduce**
not sure if you can reproduce it unless your tent gets removed unnaturally
**Expected behavior**
i expected an empty tent
**Screenshots**
the container is the one in the middle. i'v taken care of the workbench (couldn't get it back tho)

**Desktop (please complete the following information):**
- OS: windows 10
- Game Version beta 1.0.1
- Steam Version steapAPI 020
|
1.0
|
Bugged container in my new tent - **Describe the bug**
I used to have another tent but it got deleted (i'm pretty sure it's not from the timer)
i recently got another tent and when i entered it for the first time some of my old stuff was there including a radiator, a workbench (out of bounds) and a container in the middle of the room that had some of my old stuff
i cannot remove this container and i'm pretty sure that when trying to remove it it dupes itself by dropping but not disapearing (the item is stuck under it so not exploiting by taking 'em)
**To Reproduce**
not sure if you can reproduce it unless your tent gets removed unnaturally
**Expected behavior**
i expected an empty tent
**Screenshots**
the container is the one in the middle. i'v taken care of the workbench (couldn't get it back tho)

**Desktop (please complete the following information):**
- OS: windows 10
- Game Version beta 1.0.1
- Steam Version steapAPI 020
|
non_defect
|
bugged container in my new tent describe the bug i used to have another tent but it got deleted i m pretty sure it s not from the timer i recently got another tent and when i entered it for the first time some of my old stuff was there including a radiator a workbench out of bounds and a container in the middle of the room that had some of my old stuff i cannot remove this container and i m pretty sure that when trying to remove it it dupes itself by dropping but not disapearing the item is stuck under it so not exploiting by taking em to reproduce not sure if you can reproduce it unless your tent gets removed unnaturally expected behavior i expected an empty tent screenshots the container is the one in the middle i v taken care of the workbench couldn t get it back tho desktop please complete the following information os windows game version beta steam version steapapi
| 0
|
172,707
| 6,515,663,072
|
IssuesEvent
|
2017-08-26 18:43:28
|
pravega/pravega
|
https://api.github.com/repos/pravega/pravega
|
opened
|
Write to aborted txn never returns
|
area/client kind/bug priority/P0 status/needs-attention version/1
|
**Problem description**
I have come across a scenario in which a writer was stuck trying to complete the write of an event, but it couldn't because the txn had timed out and aborted. Specifically, here are the two consecutive appends of the same writer thread and the interleaving messages (this is from test case `MultiReaderTxnWriterWithFailoverTest`
[test.log.zip](https://github.com/pravega/pravega/files/1253896/test.log.zip)
):
```
2017-08-26 16:45:25,799 11109 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.t.s.MultiReaderTxnWriterWithFailoverTest - Writing event: 1303 into transaction: 775ba19a-5c2d-4f67-b6f1-aacb226c955c
2017-08-26 16:45:25,800 11110 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/2#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: d143d153-542b-4e29-aed1-c3b410daf23e
2017-08-26 16:45:25,800 11110 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.common.util.ReusableFutureLatch - Running releaser now, runningThread:MultiReaderTxnWriterWithFailoverTest-1
2017-08-26 16:46:22,848 68158 [MultiReaderTxnWriterWithFailoverTest-1] TRACE i.p.c.s.i.SegmentOutputStreamImpl - Sending append request: Append(segment=testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/2#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerId=d143d153-542b-4e29-aed1-c3b410daf23e, eventNumber=1, eventCount=1, data=UnpooledHeapByteBuf(ridx: 0, widx: 82, cap: 82/82), expectedLength=null)
2017-08-26 16:46:22,849 68159 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.t.s.MultiReaderTxnWriterWithFailoverTest - Writing event: 1307 into transaction: 775ba19a-5c2d-4f67-b6f1-aacb226c955c
```
Note that there was for some reason a 1-minute gap between the two writes, it is possible that there has been a failover even in the middle.
After that, I saw continuously the following pattern of messages:
```
2017-08-26 16:46:22,859 68169 [clientInternal-3] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - Running reconnect for segment:testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,859 68169 [clientInternal-3] INFO i.p.c.s.i.SegmentOutputStreamImpl - Fetching endpoint for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,859 68169 [clientInternal-3] TRACE i.p.c.stream.impl.ControllerImpl - ENTER getEndpointForSegment@5408355552528 [testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c].
2017-08-26 16:46:22,864 68174 [clientInternal-2] INFO i.p.c.s.i.SegmentOutputStreamImpl - Establishing connection to PravegaNodeUri(endpoint=172.16.23.10, port=12345) for testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,866 68176 [epollEventLoopGroup-4-8] TRACE i.p.s.protocol.netty.CommandEncoder - Encoding message to send over the wire WireCommands.SetupAppend(type=SETUP_APPEND, requestId=2, writerId=53abe18e-dfee-4209-b6c6-bef09cb07b0b, segment=testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c)
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] TRACE i.p.s.protocol.netty.CommandDecoder - Decode a message on connection: 172.16.23.10/172.16.23.10:12345. Message was No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] DEBUG i.p.c.n.i.ClientConnectionInboundHandler - 172.16.23.10 processing reply: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] WARN i.p.c.s.i.SegmentOutputStreamImpl - Connection for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c failed due to: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
java.util.concurrent.CompletionException: java.lang.IllegalArgumentException: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
Caused by: java.lang.IllegalArgumentException: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] WARN i.p.c.s.i.SegmentOutputStreamImpl - Connection for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c failed due to: null
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
```
Showing the writer trying to access the txn segment.
**Problem location**
Client writer.
**Suggestions for an improvement**
The writer should after some retries check whether the txn has aborted and error the operation accordingly.
|
1.0
|
Write to aborted txn never returns - **Problem description**
I have come across a scenario in which a writer was stuck trying to complete the write of an event, but it couldn't because the txn had timed out and aborted. Specifically, here are the two consecutive appends of the same writer thread and the interleaving messages (this is from test case `MultiReaderTxnWriterWithFailoverTest`
[test.log.zip](https://github.com/pravega/pravega/files/1253896/test.log.zip)
):
```
2017-08-26 16:45:25,799 11109 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.t.s.MultiReaderTxnWriterWithFailoverTest - Writing event: 1303 into transaction: 775ba19a-5c2d-4f67-b6f1-aacb226c955c
2017-08-26 16:45:25,800 11110 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/2#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: d143d153-542b-4e29-aed1-c3b410daf23e
2017-08-26 16:45:25,800 11110 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.common.util.ReusableFutureLatch - Running releaser now, runningThread:MultiReaderTxnWriterWithFailoverTest-1
2017-08-26 16:46:22,848 68158 [MultiReaderTxnWriterWithFailoverTest-1] TRACE i.p.c.s.i.SegmentOutputStreamImpl - Sending append request: Append(segment=testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/2#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerId=d143d153-542b-4e29-aed1-c3b410daf23e, eventNumber=1, eventCount=1, data=UnpooledHeapByteBuf(ridx: 0, widx: 82, cap: 82/82), expectedLength=null)
2017-08-26 16:46:22,849 68159 [MultiReaderTxnWriterWithFailoverTest-1] DEBUG i.p.t.s.MultiReaderTxnWriterWithFailoverTest - Writing event: 1307 into transaction: 775ba19a-5c2d-4f67-b6f1-aacb226c955c
```
Note that there was for some reason a 1-minute gap between the two writes, it is possible that there has been a failover even in the middle.
After that, I saw continuously the following pattern of messages:
```
2017-08-26 16:46:22,859 68169 [clientInternal-3] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - Running reconnect for segment:testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,859 68169 [clientInternal-3] INFO i.p.c.s.i.SegmentOutputStreamImpl - Fetching endpoint for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,859 68169 [clientInternal-3] TRACE i.p.c.stream.impl.ControllerImpl - ENTER getEndpointForSegment@5408355552528 [testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c].
2017-08-26 16:46:22,864 68174 [clientInternal-2] INFO i.p.c.s.i.SegmentOutputStreamImpl - Establishing connection to PravegaNodeUri(endpoint=172.16.23.10, port=12345) for testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,866 68176 [epollEventLoopGroup-4-8] TRACE i.p.s.protocol.netty.CommandEncoder - Encoding message to send over the wire WireCommands.SetupAppend(type=SETUP_APPEND, requestId=2, writerId=53abe18e-dfee-4209-b6c6-bef09cb07b0b, segment=testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c)
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] TRACE i.p.s.protocol.netty.CommandDecoder - Decode a message on connection: 172.16.23.10/172.16.23.10:12345. Message was No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] DEBUG i.p.c.n.i.ClientConnectionInboundHandler - 172.16.23.10 processing reply: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,867 68177 [epollEventLoopGroup-4-8] WARN i.p.c.s.i.SegmentOutputStreamImpl - Connection for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c failed due to: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
java.util.concurrent.CompletionException: java.lang.IllegalArgumentException: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
Caused by: java.lang.IllegalArgumentException: No such segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] WARN i.p.c.s.i.SegmentOutputStreamImpl - Connection for segment testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c failed due to: null
2017-08-26 16:46:22,868 68178 [epollEventLoopGroup-4-8] DEBUG i.p.c.s.i.SegmentOutputStreamImpl - (Re)connect invoked, Segment: testMultiReaderWriterTxnScope468184854/testMultiReaderWriterTxnStream/1#transaction.775ba19a5c2d4f67b6f1aacb226c955c, writerID: 53abe18e-dfee-4209-b6c6-bef09cb07b0b
```
Showing the writer trying to access the txn segment.
**Problem location**
Client writer.
**Suggestions for an improvement**
The writer should after some retries check whether the txn has aborted and error the operation accordingly.
|
non_defect
|
write to aborted txn never returns problem description i have come across a scenario in which a writer was stuck trying to complete the write of an event but it couldn t because the txn had timed out and aborted specifically here are the two consecutive appends of the same writer thread and the interleaving messages this is from test case multireadertxnwriterwithfailovertest debug i p t s multireadertxnwriterwithfailovertest writing event into transaction debug i p c s i segmentoutputstreamimpl re connect invoked segment testmultireaderwritertxnstream transaction writerid debug i p common util reusablefuturelatch running releaser now runningthread multireadertxnwriterwithfailovertest trace i p c s i segmentoutputstreamimpl sending append request append segment testmultireaderwritertxnstream transaction writerid eventnumber eventcount data unpooledheapbytebuf ridx widx cap expectedlength null debug i p t s multireadertxnwriterwithfailovertest writing event into transaction note that there was for some reason a minute gap between the two writes it is possible that there has been a failover even in the middle after that i saw continuously the following pattern of messages debug i p c s i segmentoutputstreamimpl running reconnect for segment testmultireaderwritertxnstream transaction writerid dfee info i p c s i segmentoutputstreamimpl fetching endpoint for segment testmultireaderwritertxnstream transaction writerid dfee trace i p c stream impl controllerimpl enter getendpointforsegment info i p c s i segmentoutputstreamimpl establishing connection to praveganodeuri endpoint port for testmultireaderwritertxnstream transaction writerid dfee trace i p s protocol netty commandencoder encoding message to send over the wire wirecommands setupappend type setup append requestid writerid dfee segment testmultireaderwritertxnstream transaction trace i p s protocol netty commanddecoder decode a message on connection message was no such segment testmultireaderwritertxnstream transaction debug i p c n i clientconnectioninboundhandler processing reply no such segment testmultireaderwritertxnstream transaction warn i p c s i segmentoutputstreamimpl connection for segment testmultireaderwritertxnstream transaction failed due to no such segment testmultireaderwritertxnstream transaction java util concurrent completionexception java lang illegalargumentexception no such segment testmultireaderwritertxnstream transaction caused by java lang illegalargumentexception no such segment testmultireaderwritertxnstream transaction debug i p c s i segmentoutputstreamimpl re connect invoked segment testmultireaderwritertxnstream transaction writerid dfee warn i p c s i segmentoutputstreamimpl connection for segment testmultireaderwritertxnstream transaction failed due to null debug i p c s i segmentoutputstreamimpl re connect invoked segment testmultireaderwritertxnstream transaction writerid dfee showing the writer trying to access the txn segment problem location client writer suggestions for an improvement the writer should after some retries check whether the txn has aborted and error the operation accordingly
| 0
|
48,235
| 13,067,556,619
|
IssuesEvent
|
2020-07-31 00:50:27
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
[filterscripts] bug in gfu filter? (Trac #2031)
|
Migrated from Trac combo reconstruction defect
|
When running simulationFiltering.py, this was found.
```text
ERROR (I3Module): GammaFollowUp_BDT_Up: Exception thrown (I3Module.cxx:120 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py", line 357, in <module>
main(opts)
File "/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py", line 314, in main
tray.Execute()
File "/data/user/ssarkar/IceRec/build-trunk/lib/I3Tray.py", line 256, in Execute
super(I3Tray, self).Execute()
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 668, in <lambda>
If = lambda f: If(f) and EventHasBasicReco(f) and (not EventIsDowngoing(f)) and EventPassesUpgoingPrecuts(f)
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 616, in EventPassesUpgoingPrecuts
var = get_bdt_vars_up(frame)
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 559, in get_bdt_vars_up
BayesLlhDiff = replace_nan(clip(frame['OnlineL2_BayesianFitFitParams'].logl - frame['OnlineL2_SPE2itFitFitParams'].logl, -100.), -100.),
KeyError: 'OnlineL2_BayesianFitFitParams'
```
Migrated from https://code.icecube.wisc.edu/ticket/2031
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:53",
"description": "When running simulationFiltering.py, this was found.\n\n{{{\nERROR (I3Module): GammaFollowUp_BDT_Up: Exception thrown (I3Module.cxx:120 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py\", line 357, in <module>\n main(opts)\n File \"/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py\", line 314, in main\n tray.Execute()\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 668, in <lambda>\n If = lambda f: If(f) and EventHasBasicReco(f) and (not EventIsDowngoing(f)) and EventPassesUpgoingPrecuts(f)\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 616, in EventPassesUpgoingPrecuts\n var = get_bdt_vars_up(frame)\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 559, in get_bdt_vars_up\n BayesLlhDiff = replace_nan(clip(frame['OnlineL2_BayesianFitFitParams'].logl - frame['OnlineL2_SPE2itFitFitParams'].logl, -100.), -100.),\nKeyError: 'OnlineL2_BayesianFitFitParams'\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067233566919",
"component": "combo reconstruction",
"summary": "[filterscripts] bug in gfu filter?",
"priority": "major",
"keywords": "",
"time": "2017-05-23T21:20:14",
"milestone": "",
"owner": "thomas.kintscher",
"type": "defect"
}
```
|
1.0
|
[filterscripts] bug in gfu filter? (Trac #2031) - When running simulationFiltering.py, this was found.
```text
ERROR (I3Module): GammaFollowUp_BDT_Up: Exception thrown (I3Module.cxx:120 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py", line 357, in <module>
main(opts)
File "/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py", line 314, in main
tray.Execute()
File "/data/user/ssarkar/IceRec/build-trunk/lib/I3Tray.py", line 256, in Execute
super(I3Tray, self).Execute()
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 668, in <lambda>
If = lambda f: If(f) and EventHasBasicReco(f) and (not EventIsDowngoing(f)) and EventPassesUpgoingPrecuts(f)
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 616, in EventPassesUpgoingPrecuts
var = get_bdt_vars_up(frame)
File "/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py", line 559, in get_bdt_vars_up
BayesLlhDiff = replace_nan(clip(frame['OnlineL2_BayesianFitFitParams'].logl - frame['OnlineL2_SPE2itFitFitParams'].logl, -100.), -100.),
KeyError: 'OnlineL2_BayesianFitFitParams'
```
Migrated from https://code.icecube.wisc.edu/ticket/2031
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:53",
"description": "When running simulationFiltering.py, this was found.\n\n{{{\nERROR (I3Module): GammaFollowUp_BDT_Up: Exception thrown (I3Module.cxx:120 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py\", line 357, in <module>\n main(opts)\n File \"/data/user/ssarkar/IceRec/build-trunk/filterscripts/resources/scripts/SimulationFiltering.py\", line 314, in main\n tray.Execute()\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 668, in <lambda>\n If = lambda f: If(f) and EventHasBasicReco(f) and (not EventIsDowngoing(f)) and EventPassesUpgoingPrecuts(f)\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 616, in EventPassesUpgoingPrecuts\n var = get_bdt_vars_up(frame)\n File \"/data/user/ssarkar/IceRec/build-trunk/lib/icecube/filterscripts/gfufilter.py\", line 559, in get_bdt_vars_up\n BayesLlhDiff = replace_nan(clip(frame['OnlineL2_BayesianFitFitParams'].logl - frame['OnlineL2_SPE2itFitFitParams'].logl, -100.), -100.),\nKeyError: 'OnlineL2_BayesianFitFitParams'\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067233566919",
"component": "combo reconstruction",
"summary": "[filterscripts] bug in gfu filter?",
"priority": "major",
"keywords": "",
"time": "2017-05-23T21:20:14",
"milestone": "",
"owner": "thomas.kintscher",
"type": "defect"
}
```
|
defect
|
bug in gfu filter trac when running simulationfiltering py this was found text error gammafollowup bdt up exception thrown cxx in void do void traceback most recent call last file data user ssarkar icerec build trunk filterscripts resources scripts simulationfiltering py line in main opts file data user ssarkar icerec build trunk filterscripts resources scripts simulationfiltering py line in main tray execute file data user ssarkar icerec build trunk lib py line in execute super self execute file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in if lambda f if f and eventhasbasicreco f and not eventisdowngoing f and eventpassesupgoingprecuts f file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in eventpassesupgoingprecuts var get bdt vars up frame file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in get bdt vars up bayesllhdiff replace nan clip frame logl frame logl keyerror bayesianfitfitparams migrated from json status closed changetime description when running simulationfiltering py this was found n n nerror gammafollowup bdt up exception thrown cxx in void do void ntraceback most recent call last n file data user ssarkar icerec build trunk filterscripts resources scripts simulationfiltering py line in n main opts n file data user ssarkar icerec build trunk filterscripts resources scripts simulationfiltering py line in main n tray execute n file data user ssarkar icerec build trunk lib py line in execute n super self execute n file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in n if lambda f if f and eventhasbasicreco f and not eventisdowngoing f and eventpassesupgoingprecuts f n file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in eventpassesupgoingprecuts n var get bdt vars up frame n file data user ssarkar icerec build trunk lib icecube filterscripts gfufilter py line in get bdt vars up n bayesllhdiff replace nan clip frame logl frame logl nkeyerror bayesianfitfitparams n reporter david schultz cc resolution fixed ts component combo reconstruction summary bug in gfu filter priority major keywords time milestone owner thomas kintscher type defect
| 1
|
78,387
| 27,491,818,320
|
IssuesEvent
|
2023-03-04 18:08:58
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Expire cookie when App defaultTimezone is not UTC
|
defect
|
### Description
Cakephp 4.x
How to reproduce :
Configuration
```php
# config/app.php
'defaultTimezone' => 'Europe/Paris',
```
Tests controller
```php
<?php
declare(strict_types=1);
namespace App\Controller;
use Cake\Http\Cookie\Cookie;
class TestsController extends AppController
{
public function setCookie()
{
$cookie = new Cookie('testing', 'my value');
$this->setResponse($this->response->withCookie($cookie));
return $this->redirect('/tests/readCookie');
}
public function readCookie()
{
if ($this->request->getCookieCollection()->has('testing')) {
$cookie = $this->request->getCookieCollection()->get('testing');
debug($cookie);
} else {
debug('No cookie. Yeah!');
}
exit;
}
public function expireCookie()
{
$cookie = $this->request->getCookieCollection()->get('testing');
$this->setResponse($this->response->withExpiredCookie($cookie));
return $this->redirect('/tests/readCookie');
}
}
```
Url : `/tests/setCookie` redirect to `/tests/readCookie`. Cookie is displayed
Url : `/tests/expireCookie` redirect to `/tests/readCookie`. Cookie should not be displayed but it is.
Changing the `App.defaultTimezone` fix this.
### CakePHP Version
4.4.11
### PHP Version
_No response_
|
1.0
|
Expire cookie when App defaultTimezone is not UTC - ### Description
Cakephp 4.x
How to reproduce :
Configuration
```php
# config/app.php
'defaultTimezone' => 'Europe/Paris',
```
Tests controller
```php
<?php
declare(strict_types=1);
namespace App\Controller;
use Cake\Http\Cookie\Cookie;
class TestsController extends AppController
{
public function setCookie()
{
$cookie = new Cookie('testing', 'my value');
$this->setResponse($this->response->withCookie($cookie));
return $this->redirect('/tests/readCookie');
}
public function readCookie()
{
if ($this->request->getCookieCollection()->has('testing')) {
$cookie = $this->request->getCookieCollection()->get('testing');
debug($cookie);
} else {
debug('No cookie. Yeah!');
}
exit;
}
public function expireCookie()
{
$cookie = $this->request->getCookieCollection()->get('testing');
$this->setResponse($this->response->withExpiredCookie($cookie));
return $this->redirect('/tests/readCookie');
}
}
```
Url : `/tests/setCookie` redirect to `/tests/readCookie`. Cookie is displayed
Url : `/tests/expireCookie` redirect to `/tests/readCookie`. Cookie should not be displayed but it is.
Changing the `App.defaultTimezone` fix this.
### CakePHP Version
4.4.11
### PHP Version
_No response_
|
defect
|
expire cookie when app defaulttimezone is not utc description cakephp x how to reproduce configuration php config app php defaulttimezone europe paris tests controller php php declare strict types namespace app controller use cake http cookie cookie class testscontroller extends appcontroller public function setcookie cookie new cookie testing my value this setresponse this response withcookie cookie return this redirect tests readcookie public function readcookie if this request getcookiecollection has testing cookie this request getcookiecollection get testing debug cookie else debug no cookie yeah exit public function expirecookie cookie this request getcookiecollection get testing this setresponse this response withexpiredcookie cookie return this redirect tests readcookie url tests setcookie redirect to tests readcookie cookie is displayed url tests expirecookie redirect to tests readcookie cookie should not be displayed but it is changing the app defaulttimezone fix this cakephp version php version no response
| 1
|
69,684
| 22,609,201,265
|
IssuesEvent
|
2022-06-29 15:38:25
|
zed-industries/feedback
|
https://api.github.com/repos/zed-industries/feedback
|
opened
|
Option-left/right do not reposition the cursor correctly
|
defect triage
|
**Describe the bug**
Alt-left/right behave differently than native macOS text editing controls.
**To reproduce**
Text like this (pipe is cursor):
hello.|
Press alt-left. See that the cursor is here:
hello|.
**Expected behavior**
Cursor should be here:
|hello.
**Environment:**
```
Zed 0.42.0 – /Applications/Zed.app
macOS 12.3.1
architecture x86_64
```
|
1.0
|
Option-left/right do not reposition the cursor correctly - **Describe the bug**
Alt-left/right behave differently than native macOS text editing controls.
**To reproduce**
Text like this (pipe is cursor):
hello.|
Press alt-left. See that the cursor is here:
hello|.
**Expected behavior**
Cursor should be here:
|hello.
**Environment:**
```
Zed 0.42.0 – /Applications/Zed.app
macOS 12.3.1
architecture x86_64
```
|
defect
|
option left right do not reposition the cursor correctly describe the bug alt left right behave differently than native macos text editing controls to reproduce text like this pipe is cursor hello press alt left see that the cursor is here hello expected behavior cursor should be here hello environment zed – applications zed app macos architecture
| 1
|
604,948
| 18,721,867,125
|
IssuesEvent
|
2021-11-03 12:43:41
|
woocommerce/woocommerce-admin
|
https://api.github.com/repos/woocommerce/woocommerce-admin
|
closed
|
Fatal Error in Cron Event
|
type: bug priority: critical
|
### Describe the bug
A fatal error occurs when the `wc_admin_daily` cron event runs.
### To Reproduce
<!-- Describe the steps to reproduce the behavior.-->
1. Run the cron event, e.g. `wp cron event run wc_admin_daily`
2. See fatal error.
### Actual behavior:
```
Fatal error: Uncaught Error: Call to undefined method Automattic\WooCommerce\Admin\RemoteInboxNotifications\RemoteInboxNotificationsEngine::get_data_source_poller_instance() in /var/www/html/wp-content/plugins/woocommerce-admin/src/Events.php:102
Stack trace:
#0 /var/www/html/wp-includes/class-wp-hook.php(303): Automattic\WooCommerce\Admin\Events->do_wc_admin_daily()
#1 /var/www/html/wp-includes/class-wp-hook.php(327): WP_Hook->apply_filters('', Array)
#2 /var/www/html/wp-includes/plugin.php(518): WP_Hook->do_action(Array)
#3 phar:///usr/local/bin/wp/vendor/wp-cli/cron-command/src/Cron_Event_Command.php(343): do_action_ref_array('wc_admin_daily', Array)
#4 phar:///usr/local/bin/wp/vendor/wp-cli/cron-command/src/Cron_Event_Command.php(265): Cron_Event_Command::run_event(Object(stdClass))
#5 [internal function]: Cron_Event_Command->run(Array, Array)
#6 phar:///usr/local/bin/wp/vendor/wp-cli/wp-cli/php/WP_CLI/Dispatcher/CommandFactory.php(100): call_user_func(Array, Array, Array)
#7 [internal function]: WP_CLI\Dispatcher\Com in /var/www/html/wp-content/plugins/woocommerce-admin/src/Events.php on line 102
```
### Expected behavior
No fatal error.
### Additional context
The error might have been introduced in this PR https://github.com/woocommerce/woocommerce-admin/pull/7671
|
1.0
|
Fatal Error in Cron Event - ### Describe the bug
A fatal error occurs when the `wc_admin_daily` cron event runs.
### To Reproduce
<!-- Describe the steps to reproduce the behavior.-->
1. Run the cron event, e.g. `wp cron event run wc_admin_daily`
2. See fatal error.
### Actual behavior:
```
Fatal error: Uncaught Error: Call to undefined method Automattic\WooCommerce\Admin\RemoteInboxNotifications\RemoteInboxNotificationsEngine::get_data_source_poller_instance() in /var/www/html/wp-content/plugins/woocommerce-admin/src/Events.php:102
Stack trace:
#0 /var/www/html/wp-includes/class-wp-hook.php(303): Automattic\WooCommerce\Admin\Events->do_wc_admin_daily()
#1 /var/www/html/wp-includes/class-wp-hook.php(327): WP_Hook->apply_filters('', Array)
#2 /var/www/html/wp-includes/plugin.php(518): WP_Hook->do_action(Array)
#3 phar:///usr/local/bin/wp/vendor/wp-cli/cron-command/src/Cron_Event_Command.php(343): do_action_ref_array('wc_admin_daily', Array)
#4 phar:///usr/local/bin/wp/vendor/wp-cli/cron-command/src/Cron_Event_Command.php(265): Cron_Event_Command::run_event(Object(stdClass))
#5 [internal function]: Cron_Event_Command->run(Array, Array)
#6 phar:///usr/local/bin/wp/vendor/wp-cli/wp-cli/php/WP_CLI/Dispatcher/CommandFactory.php(100): call_user_func(Array, Array, Array)
#7 [internal function]: WP_CLI\Dispatcher\Com in /var/www/html/wp-content/plugins/woocommerce-admin/src/Events.php on line 102
```
### Expected behavior
No fatal error.
### Additional context
The error might have been introduced in this PR https://github.com/woocommerce/woocommerce-admin/pull/7671
|
non_defect
|
fatal error in cron event describe the bug a fatal error occurs when the wc admin daily cron event runs to reproduce run the cron event e g wp cron event run wc admin daily see fatal error actual behavior fatal error uncaught error call to undefined method automattic woocommerce admin remoteinboxnotifications remoteinboxnotificationsengine get data source poller instance in var www html wp content plugins woocommerce admin src events php stack trace var www html wp includes class wp hook php automattic woocommerce admin events do wc admin daily var www html wp includes class wp hook php wp hook apply filters array var www html wp includes plugin php wp hook do action array phar usr local bin wp vendor wp cli cron command src cron event command php do action ref array wc admin daily array phar usr local bin wp vendor wp cli cron command src cron event command php cron event command run event object stdclass cron event command run array array phar usr local bin wp vendor wp cli wp cli php wp cli dispatcher commandfactory php call user func array array array wp cli dispatcher com in var www html wp content plugins woocommerce admin src events php on line expected behavior no fatal error additional context the error might have been introduced in this pr
| 0
|
74,566
| 25,180,875,085
|
IssuesEvent
|
2022-11-11 13:33:32
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Cover Non-clinical services and VBA services with workbench access
|
Defect ⭐️ Facilities
|
## Description

Its current state is that any vamc or vba editor could edit any any service of any other facility. They should only be able to edit services in their section.
## Acceptance Criteria
- [ ] VAMC non-clinical services are covered by workbench access restrictions.
- [ ] VBA services are covered by workbench access
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
1.0
|
Cover Non-clinical services and VBA services with workbench access - ## Description

Its current state is that any vamc or vba editor could edit any any service of any other facility. They should only be able to edit services in their section.
## Acceptance Criteria
- [ ] VAMC non-clinical services are covered by workbench access restrictions.
- [ ] VBA services are covered by workbench access
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
defect
|
cover non clinical services and vba services with workbench access description its current state is that any vamc or vba editor could edit any any service of any other facility they should only be able to edit services in their section acceptance criteria vamc non clinical services are covered by workbench access restrictions vba services are covered by workbench access cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
| 1
|
409,367
| 11,961,328,153
|
IssuesEvent
|
2020-04-05 07:53:09
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
improve message when docker is not installed for none
|
co/none-driver co/runtime/docker good first issue kind/cleanup priority/important-soon
|
when docker is not installed for none, instead of crashing, we could tell user nicely, please install docker
```
medya@none-experiment-instance-1-medya:~$ sudo minikube start --vm-driver=none
😄 minikube v1.8.2 on Ubuntu 18.04
✨ Using the none driver based on user configuration
🤹 Running on localhost (CPUs=2, Memory=7470MB, Disk=9749MB) ...
💣 Unable to start VM. Please investigate and run 'minikube delete' if possible: creating host: create: precreate:
exec: "docker": executable file not found in $PATH
😿 minikube is exiting due to an error. If the above message is not useful, open an issue:
👉 https://github.com/kubernetes/minikube/issues/new/choose
```
|
1.0
|
improve message when docker is not installed for none - when docker is not installed for none, instead of crashing, we could tell user nicely, please install docker
```
medya@none-experiment-instance-1-medya:~$ sudo minikube start --vm-driver=none
😄 minikube v1.8.2 on Ubuntu 18.04
✨ Using the none driver based on user configuration
🤹 Running on localhost (CPUs=2, Memory=7470MB, Disk=9749MB) ...
💣 Unable to start VM. Please investigate and run 'minikube delete' if possible: creating host: create: precreate:
exec: "docker": executable file not found in $PATH
😿 minikube is exiting due to an error. If the above message is not useful, open an issue:
👉 https://github.com/kubernetes/minikube/issues/new/choose
```
|
non_defect
|
improve message when docker is not installed for none when docker is not installed for none instead of crashing we could tell user nicely please install docker medya none experiment instance medya sudo minikube start vm driver none 😄 minikube on ubuntu ✨ using the none driver based on user configuration 🤹 running on localhost cpus memory disk 💣 unable to start vm please investigate and run minikube delete if possible creating host create precreate exec docker executable file not found in path 😿 minikube is exiting due to an error if the above message is not useful open an issue 👉
| 0
|
56,038
| 14,910,530,808
|
IssuesEvent
|
2021-01-22 09:43:21
|
STEllAR-GROUP/phylanx
|
https://api.github.com/repos/STEllAR-GROUP/phylanx
|
closed
|
array primitive doesn't work correctly
|
category: primitives type: compatibility issue type: defect
|
One expects the array() primitive to convert a list to vector as it does in numpy. Unfortunately, as you see in the foo() method, it instead produces an error claiming that the list is ir::range and not a numeric type.
There is a workaround. In foo2(), the call to transpose(vstack(li))[0] does what I would expect array(li) to produce. However, this is doubtless less efficient than it ought to be.
```
from phylanx import Phylanx
import numpy as np
@Phylanx
def foo():
li = [1,2,3,4,5]
print(li)
print(array(li))
@Phylanx
def foo2():
li = [1,2,3,4,5]
print(li)
print(transpose(vstack(li))[0])
foo2()
foo()
```
Produces
```
list(1, 2, 3, 4, 5)
[1, 2, 3, 4, 5]
list(1, 2, 3, 4, 5)
Traceback (most recent call last):
File "x.py", line 17, in <module>
foo()
File "/home/jovyan/.local/lib/python3.8/site-packages/phylanx-0.0.1-py3.8-linux-x86_64.egg/phylanx/ast/transducer.py", line 191, in __call__
result = self.backend.call(*mapped_args, **mapped_kwargs)
File "/home/jovyan/.local/lib/python3.8/site-packages/phylanx-0.0.1-py3.8-linux-x86_64.egg/phylanx/ast/physl.py", line 597, in call
result = phylanx.execution_tree.eval(
RuntimeError: x.py(1, 111): hstack:: primitive_argument_type does not hold a numeric value type (type held: 'phylanx::ir::range'): HPX(bad_parameter)
```
|
1.0
|
array primitive doesn't work correctly - One expects the array() primitive to convert a list to vector as it does in numpy. Unfortunately, as you see in the foo() method, it instead produces an error claiming that the list is ir::range and not a numeric type.
There is a workaround. In foo2(), the call to transpose(vstack(li))[0] does what I would expect array(li) to produce. However, this is doubtless less efficient than it ought to be.
```
from phylanx import Phylanx
import numpy as np
@Phylanx
def foo():
li = [1,2,3,4,5]
print(li)
print(array(li))
@Phylanx
def foo2():
li = [1,2,3,4,5]
print(li)
print(transpose(vstack(li))[0])
foo2()
foo()
```
Produces
```
list(1, 2, 3, 4, 5)
[1, 2, 3, 4, 5]
list(1, 2, 3, 4, 5)
Traceback (most recent call last):
File "x.py", line 17, in <module>
foo()
File "/home/jovyan/.local/lib/python3.8/site-packages/phylanx-0.0.1-py3.8-linux-x86_64.egg/phylanx/ast/transducer.py", line 191, in __call__
result = self.backend.call(*mapped_args, **mapped_kwargs)
File "/home/jovyan/.local/lib/python3.8/site-packages/phylanx-0.0.1-py3.8-linux-x86_64.egg/phylanx/ast/physl.py", line 597, in call
result = phylanx.execution_tree.eval(
RuntimeError: x.py(1, 111): hstack:: primitive_argument_type does not hold a numeric value type (type held: 'phylanx::ir::range'): HPX(bad_parameter)
```
|
defect
|
array primitive doesn t work correctly one expects the array primitive to convert a list to vector as it does in numpy unfortunately as you see in the foo method it instead produces an error claiming that the list is ir range and not a numeric type there is a workaround in the call to transpose vstack li does what i would expect array li to produce however this is doubtless less efficient than it ought to be from phylanx import phylanx import numpy as np phylanx def foo li print li print array li phylanx def li print li print transpose vstack li foo produces list list traceback most recent call last file x py line in foo file home jovyan local lib site packages phylanx linux egg phylanx ast transducer py line in call result self backend call mapped args mapped kwargs file home jovyan local lib site packages phylanx linux egg phylanx ast physl py line in call result phylanx execution tree eval runtimeerror x py hstack primitive argument type does not hold a numeric value type type held phylanx ir range hpx bad parameter
| 1
|
80,003
| 29,831,251,220
|
IssuesEvent
|
2023-06-18 09:56:10
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Cannot rm file from folder in nfs-share when folder has write permissions for others
|
Type: Defect Status: Stale
|
## Current Minimal Reproducer
Setup consists of two VMs one exposing an empty folder over NFS and the other one mounting it under `~/nfs-share`. Then to reproduce the issue do the following on the client side:
```
~/nfs-share$ mkdir test-dir
~/nfs-share$ chmod 757 test-dir/
~/nfs-share$ cd test-dir/
~/nfs-share/test-dir$ touch foo
~/nfs-share/test-dir$ rm foo
rm: cannot remove 'foo': Operation not permitted
```
What is interesting is that if we change the permissions of `test-dir` to `755` we're able to delete the file. I tried all the cases from `750` to `757` and it seems like whenever `others` have write permissions (specifically `752`, `753`, `756`, and `757`) for the directory containing the file then we are not able to delete the file. This is obviously a bug because since we are the owner of the file we should be able to delete it.
## Further Analysis
Looking at our internal VMs at Delphix we don't hit this issue with older VMs that don't contain the following commit:
https://github.com/openzfs/zfs/commit/235a85657686b678a293e33f2e77622e6b8db1e6
The above commit seems to have introduced another regression at some point that was later fixed (see https://github.com/openzfs/zfs/commit/66e6d3f128f22262e4be564c40ddc708725b6ed3 ). Unfortunately this commit does not fix the issue filed here.
Will update the issue as I dig further but I figured I file a bug here first in case people with more context of the above commit can point out the culprit quicker than myself.
|
1.0
|
Cannot rm file from folder in nfs-share when folder has write permissions for others - ## Current Minimal Reproducer
Setup consists of two VMs one exposing an empty folder over NFS and the other one mounting it under `~/nfs-share`. Then to reproduce the issue do the following on the client side:
```
~/nfs-share$ mkdir test-dir
~/nfs-share$ chmod 757 test-dir/
~/nfs-share$ cd test-dir/
~/nfs-share/test-dir$ touch foo
~/nfs-share/test-dir$ rm foo
rm: cannot remove 'foo': Operation not permitted
```
What is interesting is that if we change the permissions of `test-dir` to `755` we're able to delete the file. I tried all the cases from `750` to `757` and it seems like whenever `others` have write permissions (specifically `752`, `753`, `756`, and `757`) for the directory containing the file then we are not able to delete the file. This is obviously a bug because since we are the owner of the file we should be able to delete it.
## Further Analysis
Looking at our internal VMs at Delphix we don't hit this issue with older VMs that don't contain the following commit:
https://github.com/openzfs/zfs/commit/235a85657686b678a293e33f2e77622e6b8db1e6
The above commit seems to have introduced another regression at some point that was later fixed (see https://github.com/openzfs/zfs/commit/66e6d3f128f22262e4be564c40ddc708725b6ed3 ). Unfortunately this commit does not fix the issue filed here.
Will update the issue as I dig further but I figured I file a bug here first in case people with more context of the above commit can point out the culprit quicker than myself.
|
defect
|
cannot rm file from folder in nfs share when folder has write permissions for others current minimal reproducer setup consists of two vms one exposing an empty folder over nfs and the other one mounting it under nfs share then to reproduce the issue do the following on the client side nfs share mkdir test dir nfs share chmod test dir nfs share cd test dir nfs share test dir touch foo nfs share test dir rm foo rm cannot remove foo operation not permitted what is interesting is that if we change the permissions of test dir to we re able to delete the file i tried all the cases from to and it seems like whenever others have write permissions specifically and for the directory containing the file then we are not able to delete the file this is obviously a bug because since we are the owner of the file we should be able to delete it further analysis looking at our internal vms at delphix we don t hit this issue with older vms that don t contain the following commit the above commit seems to have introduced another regression at some point that was later fixed see unfortunately this commit does not fix the issue filed here will update the issue as i dig further but i figured i file a bug here first in case people with more context of the above commit can point out the culprit quicker than myself
| 1
|
22,253
| 3,619,270,011
|
IssuesEvent
|
2016-02-08 15:26:06
|
pavva94/snake-os
|
https://api.github.com/repos/pavva94/snake-os
|
closed
|
Not taking IP (Static and dhcp)
|
auto-migrated Priority-Medium Type-Defect
|
```
Hello there I am using the WLX 652 in my office:
The problem that arrived with me is that, earlier I was user belkin router with
static IP, I also turned of dhcp in all my devices including wlx652. As I reset
my router and I dont know the previous configuration of my router too and now
my WLX652 with snake os is not getting the ip from dhcp and when I connect it
through rj45 cable to my pc and provided a static ip after doing the hard reset
to my wlx652 also then also its not taking ip automatically. Please tell me
what to do
Thanks
Abhishek
Whatsapp: +919458406845
email: abhigyan17@gmail.com
```
Original issue reported on code.google.com by `abhigya...@gmail.com` on 13 Apr 2015 at 7:03
|
1.0
|
Not taking IP (Static and dhcp) - ```
Hello there I am using the WLX 652 in my office:
The problem that arrived with me is that, earlier I was user belkin router with
static IP, I also turned of dhcp in all my devices including wlx652. As I reset
my router and I dont know the previous configuration of my router too and now
my WLX652 with snake os is not getting the ip from dhcp and when I connect it
through rj45 cable to my pc and provided a static ip after doing the hard reset
to my wlx652 also then also its not taking ip automatically. Please tell me
what to do
Thanks
Abhishek
Whatsapp: +919458406845
email: abhigyan17@gmail.com
```
Original issue reported on code.google.com by `abhigya...@gmail.com` on 13 Apr 2015 at 7:03
|
defect
|
not taking ip static and dhcp hello there i am using the wlx in my office the problem that arrived with me is that earlier i was user belkin router with static ip i also turned of dhcp in all my devices including as i reset my router and i dont know the previous configuration of my router too and now my with snake os is not getting the ip from dhcp and when i connect it through cable to my pc and provided a static ip after doing the hard reset to my also then also its not taking ip automatically please tell me what to do thanks abhishek whatsapp email gmail com original issue reported on code google com by abhigya gmail com on apr at
| 1
|
296,095
| 22,288,467,322
|
IssuesEvent
|
2022-06-12 02:00:52
|
NickSchenck/Book-Nook
|
https://api.github.com/repos/NickSchenck/Book-Nook
|
closed
|
Mobile Responsive Design
|
documentation
|
As a user
I want the website to be mobile responsive
so that it works well on all of my devices
|
1.0
|
Mobile Responsive Design - As a user
I want the website to be mobile responsive
so that it works well on all of my devices
|
non_defect
|
mobile responsive design as a user i want the website to be mobile responsive so that it works well on all of my devices
| 0
|
18,901
| 3,092,478,010
|
IssuesEvent
|
2015-08-26 18:00:12
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Completion should not display instance members from a static context
|
Analyzer-Completion Area-Analyzer Priority-Medium Triaged Type-Defect
|
Using dart editor with analysis server, start with the code:
class C {
f1() {}
var f2;
static f3() {}
static var f4;
static void test() {
f
}
}
Place the cursor after the lone "f" in the test() method, and bring up completions.
Expected result: only the completions f3 and f4 should be shown. f1 and f2 should not be shown, because they are inaccessible from a static context.
Observed result: all four completions f1, f2, f3, and f4 are shown.
|
1.0
|
Completion should not display instance members from a static context - Using dart editor with analysis server, start with the code:
class C {
f1() {}
var f2;
static f3() {}
static var f4;
static void test() {
f
}
}
Place the cursor after the lone "f" in the test() method, and bring up completions.
Expected result: only the completions f3 and f4 should be shown. f1 and f2 should not be shown, because they are inaccessible from a static context.
Observed result: all four completions f1, f2, f3, and f4 are shown.
|
defect
|
completion should not display instance members from a static context using dart editor with analysis server start with the code class c nbsp nbsp nbsp nbsp var nbsp nbsp static nbsp nbsp static var nbsp nbsp static void test nbsp nbsp nbsp nbsp f nbsp nbsp place the cursor after the lone quot f quot in the test method and bring up completions expected result only the completions and should be shown and should not be shown because they are inaccessible from a static context observed result all four completions and are shown
| 1
|
75,722
| 26,015,693,789
|
IssuesEvent
|
2022-12-21 08:14:34
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
FTBFS on ppc64le with ieee-long-double toolchain
|
Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | gentoo
Distribution Version | rolling
Kernel Version | 5.15.80
Architecture | ppc64le
OpenZFS Version | 2.1.7 and today's master
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
compiler/libc (both clang and gcc) are configured with 128bit long double.
(gcc built with --with-long-double-format=ieee ./configure switch)
zfs fails to build with that toolchain:
gcc error is not very helpful:
```
../../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: error: expected '(' before '__volatile__'
30 | #define __asm __asm__ __volatile__
```
clang produces much more helpful output:
```
./../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: note: expanded from macro '__asm'
#define __asm __asm__ __volatile__
^
In file included from ../../module/zfs/vdev_raidz_math_powerpc_altivec.c:203:
In file included from ../../include/sys/vdev_raidz_impl.h:29:
In file included from ../../lib/libspl/include/sys/debug.h:30:
In file included from ../../lib/libspl/include/assert.h:32:
In file included from ../../lib/libspl/include/stdio.h:27:
In file included from /usr/include/stdio.h:896:
/usr/include/bits/stdio-ldbl.h:78:1: error: meaningless 'volatile' on asm outside function
__LDBL_REDIR_DECL (vasprintf)
^
/usr/include/sys/cdefs.h:569:31: note: expanded from macro '__LDBL_REDIR_DECL'
extern __typeof (name) name __asm (__ASMNAME ("__" #name "ieee128"));
^
../../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: note: expanded from macro '__asm'
```
### Describe how to reproduce the problem
ppc64le system, pass `-mabi=ieeelongdouble` to compiler, or have it configured to implicitly do it by default.
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
removing that __volatile__ on single line allows build to complete.
|
1.0
|
FTBFS on ppc64le with ieee-long-double toolchain - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | gentoo
Distribution Version | rolling
Kernel Version | 5.15.80
Architecture | ppc64le
OpenZFS Version | 2.1.7 and today's master
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
compiler/libc (both clang and gcc) are configured with 128bit long double.
(gcc built with --with-long-double-format=ieee ./configure switch)
zfs fails to build with that toolchain:
gcc error is not very helpful:
```
../../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: error: expected '(' before '__volatile__'
30 | #define __asm __asm__ __volatile__
```
clang produces much more helpful output:
```
./../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: note: expanded from macro '__asm'
#define __asm __asm__ __volatile__
^
In file included from ../../module/zfs/vdev_raidz_math_powerpc_altivec.c:203:
In file included from ../../include/sys/vdev_raidz_impl.h:29:
In file included from ../../lib/libspl/include/sys/debug.h:30:
In file included from ../../lib/libspl/include/assert.h:32:
In file included from ../../lib/libspl/include/stdio.h:27:
In file included from /usr/include/stdio.h:896:
/usr/include/bits/stdio-ldbl.h:78:1: error: meaningless 'volatile' on asm outside function
__LDBL_REDIR_DECL (vasprintf)
^
/usr/include/sys/cdefs.h:569:31: note: expanded from macro '__LDBL_REDIR_DECL'
extern __typeof (name) name __asm (__ASMNAME ("__" #name "ieee128"));
^
../../module/zfs/vdev_raidz_math_powerpc_altivec_common.h:30:23: note: expanded from macro '__asm'
```
### Describe how to reproduce the problem
ppc64le system, pass `-mabi=ieeelongdouble` to compiler, or have it configured to implicitly do it by default.
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
removing that __volatile__ on single line allows build to complete.
|
defect
|
ftbfs on with ieee long double toolchain thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name gentoo distribution version rolling kernel version architecture openzfs version and today s master command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing compiler libc both clang and gcc are configured with long double gcc built with with long double format ieee configure switch zfs fails to build with that toolchain gcc error is not very helpful module zfs vdev raidz math powerpc altivec common h error expected before volatile define asm asm volatile clang produces much more helpful output module zfs vdev raidz math powerpc altivec common h note expanded from macro asm define asm asm volatile in file included from module zfs vdev raidz math powerpc altivec c in file included from include sys vdev raidz impl h in file included from lib libspl include sys debug h in file included from lib libspl include assert h in file included from lib libspl include stdio h in file included from usr include stdio h usr include bits stdio ldbl h error meaningless volatile on asm outside function ldbl redir decl vasprintf usr include sys cdefs h note expanded from macro ldbl redir decl extern typeof name name asm asmname name module zfs vdev raidz math powerpc altivec common h note expanded from macro asm describe how to reproduce the problem system pass mabi ieeelongdouble to compiler or have it configured to implicitly do it by default important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with removing that volatile on single line allows build to complete
| 1
|
829,302
| 31,863,339,662
|
IssuesEvent
|
2023-09-15 12:33:56
|
filamentphp/filament
|
https://api.github.com/repos/filamentphp/filament
|
closed
|
Repeater defaults not working
|
bug unconfirmed low priority
|
### Package
filament/filament
### Package Version
v3.0.50
### Laravel Version
v10.23.1
### Livewire Version
_No response_
### PHP Version
PHP 8.2.6
### Problem description
When setting a default value for the Repeater, the default value is not set.
### Expected behavior
The default value should be set.
### Steps to reproduce
```
Forms\Components\Repeater::make('nicks')
->default([
['nick' => 'Jack'],
['nick' => 'Josh'],
])
->schema([
Forms\Components\TextInput::make('nick')
->required(),
]),
```
### Reproduction repository
https://github.com/muhajirrr/filament-repeater-defaults-issue
### Relevant log output
_No response_
|
1.0
|
Repeater defaults not working - ### Package
filament/filament
### Package Version
v3.0.50
### Laravel Version
v10.23.1
### Livewire Version
_No response_
### PHP Version
PHP 8.2.6
### Problem description
When setting a default value for the Repeater, the default value is not set.
### Expected behavior
The default value should be set.
### Steps to reproduce
```
Forms\Components\Repeater::make('nicks')
->default([
['nick' => 'Jack'],
['nick' => 'Josh'],
])
->schema([
Forms\Components\TextInput::make('nick')
->required(),
]),
```
### Reproduction repository
https://github.com/muhajirrr/filament-repeater-defaults-issue
### Relevant log output
_No response_
|
non_defect
|
repeater defaults not working package filament filament package version laravel version livewire version no response php version php problem description when setting a default value for the repeater the default value is not set expected behavior the default value should be set steps to reproduce forms components repeater make nicks default schema forms components textinput make nick required reproduction repository relevant log output no response
| 0
|
46,657
| 13,055,956,044
|
IssuesEvent
|
2020-07-30 03:13:45
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676)
|
Incomplete Migration Migrated from Trac defect iceprod
|
Migrated from https://code.icecube.wisc.edu/ticket/1676
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1462830916934929",
"component": "iceprod",
"summary": "[iceprod2] handle expiration of self-signed cert for webserver",
"priority": "major",
"keywords": "",
"time": "2016-04-29T16:41:30",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
1.0
|
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676) - Migrated from https://code.icecube.wisc.edu/ticket/1676
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1462830916934929",
"component": "iceprod",
"summary": "[iceprod2] handle expiration of self-signed cert for webserver",
"priority": "major",
"keywords": "",
"time": "2016-04-29T16:41:30",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
defect
|
handle expiration of self signed cert for webserver trac migrated from json status closed changetime description the default right now is to generate a self signed cert with a duration of days we could either make this time infinitely large several years or deal with the problem of expiring certs n nnote that just making a new cert isn t good enough because already running jobs will still have the old cert reporter david schultz cc resolution wontfix ts component iceprod summary handle expiration of self signed cert for webserver priority major keywords time milestone owner david schultz type defect
| 1
|
46,810
| 13,055,980,476
|
IssuesEvent
|
2020-07-30 03:17:45
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
hard-coded path for bzip2 in SimulatonFiltering.py (Trac #1913)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1913
```json
{
"status": "closed",
"changetime": "2016-11-22T19:13:12",
"description": "Line 381 has a hard-coded path to /usr/bin/bzip2. On some systems (Ubuntu) bzip2 is located in /bin/bzip2",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1479841992979744",
"component": "combo reconstruction",
"summary": "hard-coded path for bzip2 in SimulatonFiltering.py",
"priority": "major",
"keywords": "filterscripts",
"time": "2016-11-22T19:09:43",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
|
1.0
|
hard-coded path for bzip2 in SimulatonFiltering.py (Trac #1913) - Migrated from https://code.icecube.wisc.edu/ticket/1913
```json
{
"status": "closed",
"changetime": "2016-11-22T19:13:12",
"description": "Line 381 has a hard-coded path to /usr/bin/bzip2. On some systems (Ubuntu) bzip2 is located in /bin/bzip2",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1479841992979744",
"component": "combo reconstruction",
"summary": "hard-coded path for bzip2 in SimulatonFiltering.py",
"priority": "major",
"keywords": "filterscripts",
"time": "2016-11-22T19:09:43",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
|
defect
|
hard coded path for in simulatonfiltering py trac migrated from json status closed changetime description line has a hard coded path to usr bin on some systems ubuntu is located in bin reporter juancarlos cc resolution fixed ts component combo reconstruction summary hard coded path for in simulatonfiltering py priority major keywords filterscripts time milestone owner juancarlos type defect
| 1
|
324,435
| 27,809,014,401
|
IssuesEvent
|
2023-03-18 00:01:44
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix linalg.test_inv
|
Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4320254781/jobs/7540302326" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix linalg.test_inv - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4320254781/jobs/7540302326" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
non_defect
|
fix linalg test inv tensorflow img src torch img src numpy img src jax img src
| 0
|
31,798
| 6,626,156,642
|
IssuesEvent
|
2017-09-22 18:21:02
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
MooseDocs misses a few build failures
|
C: Documentation P: normal T: defect
|
### Description of the enhancement or error report
MOOSE build sometimes misses build failures (e.g. reports success when there are failures). See this job: https://moosebuild.inl.gov/job/84609/
### Rationale for the enhancement or information for reproducing the error
Unknown, @gardnerru can you shed some light on what the problem was and how it was caused/fixed?
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
This bug allows us to check in broken docs.
Tag @aeslaughter
|
1.0
|
MooseDocs misses a few build failures - ### Description of the enhancement or error report
MOOSE build sometimes misses build failures (e.g. reports success when there are failures). See this job: https://moosebuild.inl.gov/job/84609/
### Rationale for the enhancement or information for reproducing the error
Unknown, @gardnerru can you shed some light on what the problem was and how it was caused/fixed?
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
This bug allows us to check in broken docs.
Tag @aeslaughter
|
defect
|
moosedocs misses a few build failures description of the enhancement or error report moose build sometimes misses build failures e g reports success when there are failures see this job rationale for the enhancement or information for reproducing the error unknown gardnerru can you shed some light on what the problem was and how it was caused fixed identified impact i e internal object changes limited interface changes public api change or a list of specific applications impacted this bug allows us to check in broken docs tag aeslaughter
| 1
|
76,584
| 26,497,998,851
|
IssuesEvent
|
2023-01-18 08:00:20
|
vector-im/element-ios
|
https://api.github.com/repos/vector-im/element-ios
|
closed
|
The 30s back / forward buttons are jumping back and forth erratically when tapped
|
T-Defect A-Voice-Broadcast
|
### Steps to reproduce
1. Start playing a ended voice broadcast
2. Click on +/- 30s buttons to jump inside the record
### Outcome
#### What did you expect?
a +/- 30 sec jump of the cursor on the timeline
#### What happened instead?
a strange dance of the cursor before converging to the new timeline position
https://user-images.githubusercontent.com/8969772/211551161-e3396ef2-05dd-406b-91aa-8f690d113eb9.mov
Note: the cursor is updated correctly when the jump triggers a buffering. The issue is reproduced when the audio chunk is already available locally
https://user-images.githubusercontent.com/8969772/211551922-3994e1d2-3369-4ccb-8e47-b028b0b02529.mov
### Your phone model
_No response_
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
The 30s back / forward buttons are jumping back and forth erratically when tapped - ### Steps to reproduce
1. Start playing a ended voice broadcast
2. Click on +/- 30s buttons to jump inside the record
### Outcome
#### What did you expect?
a +/- 30 sec jump of the cursor on the timeline
#### What happened instead?
a strange dance of the cursor before converging to the new timeline position
https://user-images.githubusercontent.com/8969772/211551161-e3396ef2-05dd-406b-91aa-8f690d113eb9.mov
Note: the cursor is updated correctly when the jump triggers a buffering. The issue is reproduced when the audio chunk is already available locally
https://user-images.githubusercontent.com/8969772/211551922-3994e1d2-3369-4ccb-8e47-b028b0b02529.mov
### Your phone model
_No response_
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
the back forward buttons are jumping back and forth erratically when tapped steps to reproduce start playing a ended voice broadcast click on buttons to jump inside the record outcome what did you expect a sec jump of the cursor on the timeline what happened instead a strange dance of the cursor before converging to the new timeline position note the cursor is updated correctly when the jump triggers a buffering the issue is reproduced when the audio chunk is already available locally your phone model no response operating system version no response application version no response homeserver no response will you send logs no
| 1
|
304,118
| 26,254,168,741
|
IssuesEvent
|
2023-01-05 22:22:27
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: sqlalchemy failed
|
C-test-failure O-robot O-roachtest T-sql-sessions branch-release-22.2
|
roachtest.sqlalchemy [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7330134?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7330134?buildTab=artifacts#/sqlalchemy) on release-22.2 @ [fc7a9c040029eeac0791db653f1d13a3cbda4e76](https://github.com/cockroachdb/cockroach/commits/fc7a9c040029eeac0791db653f1d13a3cbda4e76):
```
test artifacts and logs in: /artifacts/sqlalchemy/run_1
orm_helpers.go:191,orm_helpers.go:117,sqlalchemy.go:239,sqlalchemy.go:47,test_runner.go:930:
Tests run on Cockroach v22.2.0-beta.2-659-gfc7a9c0400
Tests run against sqlalchemy rel_1_4_26
371 Total Tests Run
365 tests passed
6 tests failed
195 tests skipped
0 tests ignored
0 tests passed unexpectedly
6 tests failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_temp_view (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_view_schema (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_temp_table (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_schema (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_view (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table (unexpected)
For a full summary look at the sqlalchemy artifacts
An updated blocklist (sqlAlchemyBlocklist) is available in the artifacts' sqlalchemy log
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91343 roachtest: sqlalchemy failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlalchemy.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21217
|
2.0
|
roachtest: sqlalchemy failed - roachtest.sqlalchemy [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7330134?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7330134?buildTab=artifacts#/sqlalchemy) on release-22.2 @ [fc7a9c040029eeac0791db653f1d13a3cbda4e76](https://github.com/cockroachdb/cockroach/commits/fc7a9c040029eeac0791db653f1d13a3cbda4e76):
```
test artifacts and logs in: /artifacts/sqlalchemy/run_1
orm_helpers.go:191,orm_helpers.go:117,sqlalchemy.go:239,sqlalchemy.go:47,test_runner.go:930:
Tests run on Cockroach v22.2.0-beta.2-659-gfc7a9c0400
Tests run against sqlalchemy rel_1_4_26
371 Total Tests Run
365 tests passed
6 tests failed
195 tests skipped
0 tests ignored
0 tests passed unexpectedly
6 tests failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_temp_view (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_view_schema (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_temp_table (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_schema (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table_view (unexpected)
--- FAIL: test/test_suite_sqlalchemy.py::HasTableTest_cockroachdb+psycopg2_9_5_0::test_has_table (unexpected)
For a full summary look at the sqlalchemy artifacts
An updated blocklist (sqlAlchemyBlocklist) is available in the artifacts' sqlalchemy log
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91343 roachtest: sqlalchemy failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlalchemy.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21217
|
non_defect
|
roachtest sqlalchemy failed roachtest sqlalchemy with on release test artifacts and logs in artifacts sqlalchemy run orm helpers go orm helpers go sqlalchemy go sqlalchemy go test runner go tests run on cockroach beta tests run against sqlalchemy rel total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run fail test test suite sqlalchemy py hastabletest cockroachdb test has table temp view unexpected fail test test suite sqlalchemy py hastabletest cockroachdb test has table view schema unexpected fail test test suite sqlalchemy py hastabletest cockroachdb test has table temp table unexpected fail test test suite sqlalchemy py hastabletest cockroachdb test has table schema unexpected fail test test suite sqlalchemy py hastabletest cockroachdb test has table view unexpected fail test test suite sqlalchemy py hastabletest cockroachdb test has table unexpected for a full summary look at the sqlalchemy artifacts an updated blocklist sqlalchemyblocklist is available in the artifacts sqlalchemy log parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest sqlalchemy failed cc cockroachdb sql experience jira issue crdb
| 0
|
357,563
| 25,176,405,435
|
IssuesEvent
|
2022-11-11 09:39:05
|
JJoeYi/pe
|
https://api.github.com/repos/JJoeYi/pe
|
closed
|
DG Not Complete
|
type.DocumentationBug severity.VeryLow
|

Under the User Stories portion there is the phrase `more to be added`, which suggests that the User Stories portion is incomplete, which is not useful for future developers.
<!--session: 1668152526415-18fc0604-826d-4cf9-86a6-6642a28a0991-->
<!--Version: Web v3.4.4-->
|
1.0
|
DG Not Complete - 
Under the User Stories portion there is the phrase `more to be added`, which suggests that the User Stories portion is incomplete, which is not useful for future developers.
<!--session: 1668152526415-18fc0604-826d-4cf9-86a6-6642a28a0991-->
<!--Version: Web v3.4.4-->
|
non_defect
|
dg not complete under the user stories portion there is the phrase more to be added which suggests that the user stories portion is incomplete which is not useful for future developers
| 0
|
14,666
| 4,963,841,166
|
IssuesEvent
|
2016-12-03 13:13:26
|
VATSIM-UK/core
|
https://api.github.com/repos/VATSIM-UK/core
|
closed
|
Send reminder email to Referees after 7 days.
|
Acknowledged Code Functionality Module - VT Application System
|
In gitlab by @A-Lawrence on Sep 2, 2016, 15:59
As title suggests. Queue this up when they're originally requested.
|
1.0
|
Send reminder email to Referees after 7 days. - In gitlab by @A-Lawrence on Sep 2, 2016, 15:59
As title suggests. Queue this up when they're originally requested.
|
non_defect
|
send reminder email to referees after days in gitlab by a lawrence on sep as title suggests queue this up when they re originally requested
| 0
|
38,924
| 9,049,728,054
|
IssuesEvent
|
2019-02-12 06:06:47
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
opened
|
Ext.data.Query incompatibility with GridFilters GridPanel plugin
|
4.x defect review-after-extjs-upgrade sencha
|
Found: 4.8.0 (during development)
Ext.NET forum thread: N/A
Sencha thread: [[6.7.0 classic] The new Ext.data.Query issue with GridFilters plugin](https://www.sencha.com/forum/showthread.php?471396)
While using the [GridFilters plugin](https://docs.sencha.com/extjs/6.7.0/classic/Ext.grid.filters.Filters.html) with a GridPanel, the newly introduced [Query filter](https://docs.sencha.com/extjs/6.7.0/classic/Ext.data.Query.html) triggers an error, at least, the second time the grid store's `addFilter()` method is called while passing an `Ext.data.Query` filter.
This can be reproduced in the following Sencha Fiddle: [GridPanel + GridFilters + Ext.data.Query issue](https://fiddle.sencha.com/#view/editor&fiddle/2pm4).
To easily reproduce this in an Ext.NET example, use the [GridPanel > Plugins > GridFilters_Local](https://examples4.ext.net/#/GridPanel/Plugins/GridFilters_Local/) and call this client-side code twice:
```javascript
App.Store1.addFilter(new Ext.data.Query("Company like \"Co\" or Size = \"large\""));
```
*Note:* For the `Ext.data.Query` class to exist, it is imperative that the Ext JS version is at least 6.7.0.
It is possible that any `addFilter()` call after the first with `Ext.data.Query` will throw the exception, as the actual requirement is to call `removeFilter()` when there's at least one `Ext.data.Query` filter applied.
|
1.0
|
Ext.data.Query incompatibility with GridFilters GridPanel plugin - Found: 4.8.0 (during development)
Ext.NET forum thread: N/A
Sencha thread: [[6.7.0 classic] The new Ext.data.Query issue with GridFilters plugin](https://www.sencha.com/forum/showthread.php?471396)
While using the [GridFilters plugin](https://docs.sencha.com/extjs/6.7.0/classic/Ext.grid.filters.Filters.html) with a GridPanel, the newly introduced [Query filter](https://docs.sencha.com/extjs/6.7.0/classic/Ext.data.Query.html) triggers an error, at least, the second time the grid store's `addFilter()` method is called while passing an `Ext.data.Query` filter.
This can be reproduced in the following Sencha Fiddle: [GridPanel + GridFilters + Ext.data.Query issue](https://fiddle.sencha.com/#view/editor&fiddle/2pm4).
To easily reproduce this in an Ext.NET example, use the [GridPanel > Plugins > GridFilters_Local](https://examples4.ext.net/#/GridPanel/Plugins/GridFilters_Local/) and call this client-side code twice:
```javascript
App.Store1.addFilter(new Ext.data.Query("Company like \"Co\" or Size = \"large\""));
```
*Note:* For the `Ext.data.Query` class to exist, it is imperative that the Ext JS version is at least 6.7.0.
It is possible that any `addFilter()` call after the first with `Ext.data.Query` will throw the exception, as the actual requirement is to call `removeFilter()` when there's at least one `Ext.data.Query` filter applied.
|
defect
|
ext data query incompatibility with gridfilters gridpanel plugin found during development ext net forum thread n a sencha thread the new ext data query issue with gridfilters plugin while using the with a gridpanel the newly introduced triggers an error at least the second time the grid store s addfilter method is called while passing an ext data query filter this can be reproduced in the following sencha fiddle to easily reproduce this in an ext net example use the and call this client side code twice javascript app addfilter new ext data query company like co or size large note for the ext data query class to exist it is imperative that the ext js version is at least it is possible that any addfilter call after the first with ext data query will throw the exception as the actual requirement is to call removefilter when there s at least one ext data query filter applied
| 1
|
238,235
| 7,776,759,558
|
IssuesEvent
|
2018-06-05 09:04:40
|
metasfresh/metasfresh-webui-frontend
|
https://api.github.com/repos/metasfresh/metasfresh-webui-frontend
|
closed
|
reload /i18n/messages in case the language changed
|
branch:master priority:high type:bug
|
### Is this a bug or feature request?
Bug
### What is the current behavior?
Atm, the default language is German.
After user logs in, user's language is used.
Suppose your user language is not German, let's say it's English.
After you log in the static messages (those which are fetched using [/i18n/messages](http://w101.metasfresh.com:8081/swagger-ui.html#!/i-18-n-rest-controller/getMessagesUsingGET) endpoint) are still in German.
#### Which are the steps to reproduce?
Suppose your user's language is English (check your user profile).
Do login and inspect Chrome-Network tab.


Below are some of the places where static messages are used (just to quickly check which language is currently used):



etc etc etc
### What is the expected or desired behavior?
Static messages shall be in user's language.
|
1.0
|
reload /i18n/messages in case the language changed - ### Is this a bug or feature request?
Bug
### What is the current behavior?
Atm, the default language is German.
After user logs in, user's language is used.
Suppose your user language is not German, let's say it's English.
After you log in the static messages (those which are fetched using [/i18n/messages](http://w101.metasfresh.com:8081/swagger-ui.html#!/i-18-n-rest-controller/getMessagesUsingGET) endpoint) are still in German.
#### Which are the steps to reproduce?
Suppose your user's language is English (check your user profile).
Do login and inspect Chrome-Network tab.


Below are some of the places where static messages are used (just to quickly check which language is currently used):



etc etc etc
### What is the expected or desired behavior?
Static messages shall be in user's language.
|
non_defect
|
reload messages in case the language changed is this a bug or feature request bug what is the current behavior atm the default language is german after user logs in user s language is used suppose your user language is not german let s say it s english after you log in the static messages those which are fetched using endpoint are still in german which are the steps to reproduce suppose your user s language is english check your user profile do login and inspect chrome network tab below are some of the places where static messages are used just to quickly check which language is currently used etc etc etc what is the expected or desired behavior static messages shall be in user s language
| 0
|
54,394
| 13,642,823,892
|
IssuesEvent
|
2020-09-25 16:06:11
|
idaholab/HERON
|
https://api.github.com/repos/idaholab/HERON
|
opened
|
[DEFECT] Documentation cannot be made in Windows under Git Bash
|
defect
|
--------
Defect Description
--------
**Describe the defect**
##### What did you expect to see happen?
##### What did you see instead?
$ ./make_docs.sh
CONDA
raven_libraries * C:\Users\haoyuwang\.conda\envs\raven_libraries
Building manuals ...
... building in user_manual...
bash: make_win.sh: No such file or directory
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
1.0
|
[DEFECT] Documentation cannot be made in Windows under Git Bash - --------
Defect Description
--------
**Describe the defect**
##### What did you expect to see happen?
##### What did you see instead?
$ ./make_docs.sh
CONDA
raven_libraries * C:\Users\haoyuwang\.conda\envs\raven_libraries
Building manuals ...
... building in user_manual...
bash: make_win.sh: No such file or directory
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
defect
|
documentation cannot be made in windows under git bash defect description describe the defect what did you expect to see happen what did you see instead make docs sh conda raven libraries c users haoyuwang conda envs raven libraries building manuals building in user manual bash make win sh no such file or directory do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
| 1
|
83,919
| 16,396,402,643
|
IssuesEvent
|
2021-05-18 00:40:59
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
SuperPMI asm diffs and metrics improvement
|
area-CodeGen-coreclr
|
This issue proposes adding a new JIT-EE api for the JIT to report "metrics" back to SuperPMI, and how that could be used by SuperPMI replay and asm diffs.
Currently, running SuperPMI asm diffs happens as follows:
1. Run `superpmi.py asmdiffs` to orchestrate the process.
2. It runs `superpmi.exe -a`.
a. Each of baseline and diff JIT compile a method context and report results back to the "VM" (or here, the SuperPMI host).
b. superpmi.exe runs the "near differ" to determine if there is a binary diff. If so, the method context number is added to a list of method contexts with asm diffs.
3. `superpmi.py` iterates over the list of method context numbers, invoking both the baseline and diff JIT to generate textual .dasm (and, thus, compiling the functions again).
4. `superpmi.py` calls `jit-analyze` to analyze the results. It parses the textual output and generates improvement/regression information for each metric it is concerned with.
I propose a set of new JIT-EE methods:
```
void reportMetric(const char* metricName, unsigned int metricValue);
void reportMetric(const char* metricName, unsigned __int64 metricValue);
void reportMetric(const char* metricName, double metricValue);
void reportMetric(const char* metricName, const char* metricValue);
```
If we don't want named overloading, we could add the metric type to the function name. Or, we could have something like:
```
struct JitMetricValue
{
enum class JitMetricType { TypeInt, TypeLong, TypeDouble, TypeString };
JitMetricType jitMetricType;
union
{
unsigned int intValue; // jitMetricType == JitMetricType::TypeInt
unsigned __int64 longValue; // jitMetricType == JitMetricType::TypeLong
double doubleValue; // jitMetricType == JitMetricType::TypeDouble
const char* stringValue; // jitMetricType == JitMetricType::TypeString
};
};
void reportMetric(const char* metricName, JitMetricValue metricValue);
```
At the end of compilation, the JIT calls this API for each interesting metric it has computed. This should be the same set that `jit-analyze` will parse (or a superset), namely, "CodeSize, PerfScore, PrologSize, InstrCount, AllocSize, ExtraAllocBytes, DebugClauseCount, DebugVarCount". For example:
```
reportMetric("CodeSize", emitTotalCodeSize);
```
(we would also report hot/cold/data sizes).
SuperPMI will save these in a new CompileResult table (or tables). The VM would implement this function a no-op, and in fact, the JIT could optionally skip reporting them if SuperPMI is not the host.
For asm diffs, if there is a diff, instead of reporting just the method context#, superpmi.exe could also report the metrics. This could be as a file in CSV form, one column per metric (we might need to handle the case where different compilations report different metrics, in which case maybe json/xml would be better?).
Then, jit-analyze could work directly on this set of data, instead of needing to parse out the metrics from the generated text asm files. This would make jit-analyze much faster. In fact, the set of asm files we generate could be deferred until after jit-analyze is run (not before, as a prerequisite for running jit-analyze). For example, we could choose to only generate dasm files for the set of improved/regressed routines that jit-analyze reports, and not automatically generate dasm for all the method contexts with diffs. In the worst case where every function has diffs, this could save an enormous amount of time, avoiding generating dasm files for thousands of functions which will never be examined. We would need to make sure jit-analyze reports all the diffs that we as developers prefer to look at, including the simplest possible smallest diffs, not just the most impactful diffs. There would need to be a way to have the current behavior of generating all diffs, or perhaps a way to take the list of diffs and generate dasm from them, after the normal processing. Generating the .dasm files could also have a default reasonable cut-off (maximum number of generated files), since generating the dasm is not required to gather the stats.
This depends on all the interesting metrics being available during "normal", non-disasm compilation, which should be true.
So, the new process would look like:
1. Run `superpmi.py asmdiffs` to orchestrate the process.
2. It runs `superpmi.exe -a`.
a. Each of baseline and diff JIT compile a method context and report results back to the SuperPMI host, _including metrics_.
b. superpmi.exe runs the "near differ" to determine if there is a binary diff. If so, the method context number is added to a list of method contexts with asm diffs, _including metrics_.
3. `superpmi.py` calls jit-analyze to analyze the results. It reads the generated list of metrics for the difs and generates improvement/regression information for each metric it is concerned with. It needs to generate a machine-readable summary of its results, including method context numbers, in addition to the text/markdown format (there is current provision for TSV/JSON output).
4. `superpmi.py` iterates over the list of method context numbers with diffs that jit-analyze says is interesting, invoking both the baseline and diff JIT to generate textual .dasm for developers to examine.
The work for this should concurrently consider fixing https://github.com/dotnet/runtime/issues/41653, by passing summary of metrics available to superpmi.exe on to jit-analyze. In that case, by somehow providing complete aggregate code size (not just total code size of functions with diffs) to jit-analyze.
@dotnet/jit-contrib Comments?
|
1.0
|
SuperPMI asm diffs and metrics improvement - This issue proposes adding a new JIT-EE api for the JIT to report "metrics" back to SuperPMI, and how that could be used by SuperPMI replay and asm diffs.
Currently, running SuperPMI asm diffs happens as follows:
1. Run `superpmi.py asmdiffs` to orchestrate the process.
2. It runs `superpmi.exe -a`.
a. Each of baseline and diff JIT compile a method context and report results back to the "VM" (or here, the SuperPMI host).
b. superpmi.exe runs the "near differ" to determine if there is a binary diff. If so, the method context number is added to a list of method contexts with asm diffs.
3. `superpmi.py` iterates over the list of method context numbers, invoking both the baseline and diff JIT to generate textual .dasm (and, thus, compiling the functions again).
4. `superpmi.py` calls `jit-analyze` to analyze the results. It parses the textual output and generates improvement/regression information for each metric it is concerned with.
I propose a set of new JIT-EE methods:
```
void reportMetric(const char* metricName, unsigned int metricValue);
void reportMetric(const char* metricName, unsigned __int64 metricValue);
void reportMetric(const char* metricName, double metricValue);
void reportMetric(const char* metricName, const char* metricValue);
```
If we don't want named overloading, we could add the metric type to the function name. Or, we could have something like:
```
struct JitMetricValue
{
enum class JitMetricType { TypeInt, TypeLong, TypeDouble, TypeString };
JitMetricType jitMetricType;
union
{
unsigned int intValue; // jitMetricType == JitMetricType::TypeInt
unsigned __int64 longValue; // jitMetricType == JitMetricType::TypeLong
double doubleValue; // jitMetricType == JitMetricType::TypeDouble
const char* stringValue; // jitMetricType == JitMetricType::TypeString
};
};
void reportMetric(const char* metricName, JitMetricValue metricValue);
```
At the end of compilation, the JIT calls this API for each interesting metric it has computed. This should be the same set that `jit-analyze` will parse (or a superset), namely, "CodeSize, PerfScore, PrologSize, InstrCount, AllocSize, ExtraAllocBytes, DebugClauseCount, DebugVarCount". For example:
```
reportMetric("CodeSize", emitTotalCodeSize);
```
(we would also report hot/cold/data sizes).
SuperPMI will save these in a new CompileResult table (or tables). The VM would implement this function a no-op, and in fact, the JIT could optionally skip reporting them if SuperPMI is not the host.
For asm diffs, if there is a diff, instead of reporting just the method context#, superpmi.exe could also report the metrics. This could be as a file in CSV form, one column per metric (we might need to handle the case where different compilations report different metrics, in which case maybe json/xml would be better?).
Then, jit-analyze could work directly on this set of data, instead of needing to parse out the metrics from the generated text asm files. This would make jit-analyze much faster. In fact, the set of asm files we generate could be deferred until after jit-analyze is run (not before, as a prerequisite for running jit-analyze). For example, we could choose to only generate dasm files for the set of improved/regressed routines that jit-analyze reports, and not automatically generate dasm for all the method contexts with diffs. In the worst case where every function has diffs, this could save an enormous amount of time, avoiding generating dasm files for thousands of functions which will never be examined. We would need to make sure jit-analyze reports all the diffs that we as developers prefer to look at, including the simplest possible smallest diffs, not just the most impactful diffs. There would need to be a way to have the current behavior of generating all diffs, or perhaps a way to take the list of diffs and generate dasm from them, after the normal processing. Generating the .dasm files could also have a default reasonable cut-off (maximum number of generated files), since generating the dasm is not required to gather the stats.
This depends on all the interesting metrics being available during "normal", non-disasm compilation, which should be true.
So, the new process would look like:
1. Run `superpmi.py asmdiffs` to orchestrate the process.
2. It runs `superpmi.exe -a`.
a. Each of baseline and diff JIT compile a method context and report results back to the SuperPMI host, _including metrics_.
b. superpmi.exe runs the "near differ" to determine if there is a binary diff. If so, the method context number is added to a list of method contexts with asm diffs, _including metrics_.
3. `superpmi.py` calls jit-analyze to analyze the results. It reads the generated list of metrics for the difs and generates improvement/regression information for each metric it is concerned with. It needs to generate a machine-readable summary of its results, including method context numbers, in addition to the text/markdown format (there is current provision for TSV/JSON output).
4. `superpmi.py` iterates over the list of method context numbers with diffs that jit-analyze says is interesting, invoking both the baseline and diff JIT to generate textual .dasm for developers to examine.
The work for this should concurrently consider fixing https://github.com/dotnet/runtime/issues/41653, by passing summary of metrics available to superpmi.exe on to jit-analyze. In that case, by somehow providing complete aggregate code size (not just total code size of functions with diffs) to jit-analyze.
@dotnet/jit-contrib Comments?
|
non_defect
|
superpmi asm diffs and metrics improvement this issue proposes adding a new jit ee api for the jit to report metrics back to superpmi and how that could be used by superpmi replay and asm diffs currently running superpmi asm diffs happens as follows run superpmi py asmdiffs to orchestrate the process it runs superpmi exe a a each of baseline and diff jit compile a method context and report results back to the vm or here the superpmi host b superpmi exe runs the near differ to determine if there is a binary diff if so the method context number is added to a list of method contexts with asm diffs superpmi py iterates over the list of method context numbers invoking both the baseline and diff jit to generate textual dasm and thus compiling the functions again superpmi py calls jit analyze to analyze the results it parses the textual output and generates improvement regression information for each metric it is concerned with i propose a set of new jit ee methods void reportmetric const char metricname unsigned int metricvalue void reportmetric const char metricname unsigned metricvalue void reportmetric const char metricname double metricvalue void reportmetric const char metricname const char metricvalue if we don t want named overloading we could add the metric type to the function name or we could have something like struct jitmetricvalue enum class jitmetrictype typeint typelong typedouble typestring jitmetrictype jitmetrictype union unsigned int intvalue jitmetrictype jitmetrictype typeint unsigned longvalue jitmetrictype jitmetrictype typelong double doublevalue jitmetrictype jitmetrictype typedouble const char stringvalue jitmetrictype jitmetrictype typestring void reportmetric const char metricname jitmetricvalue metricvalue at the end of compilation the jit calls this api for each interesting metric it has computed this should be the same set that jit analyze will parse or a superset namely codesize perfscore prologsize instrcount allocsize extraallocbytes debugclausecount debugvarcount for example reportmetric codesize emittotalcodesize we would also report hot cold data sizes superpmi will save these in a new compileresult table or tables the vm would implement this function a no op and in fact the jit could optionally skip reporting them if superpmi is not the host for asm diffs if there is a diff instead of reporting just the method context superpmi exe could also report the metrics this could be as a file in csv form one column per metric we might need to handle the case where different compilations report different metrics in which case maybe json xml would be better then jit analyze could work directly on this set of data instead of needing to parse out the metrics from the generated text asm files this would make jit analyze much faster in fact the set of asm files we generate could be deferred until after jit analyze is run not before as a prerequisite for running jit analyze for example we could choose to only generate dasm files for the set of improved regressed routines that jit analyze reports and not automatically generate dasm for all the method contexts with diffs in the worst case where every function has diffs this could save an enormous amount of time avoiding generating dasm files for thousands of functions which will never be examined we would need to make sure jit analyze reports all the diffs that we as developers prefer to look at including the simplest possible smallest diffs not just the most impactful diffs there would need to be a way to have the current behavior of generating all diffs or perhaps a way to take the list of diffs and generate dasm from them after the normal processing generating the dasm files could also have a default reasonable cut off maximum number of generated files since generating the dasm is not required to gather the stats this depends on all the interesting metrics being available during normal non disasm compilation which should be true so the new process would look like run superpmi py asmdiffs to orchestrate the process it runs superpmi exe a a each of baseline and diff jit compile a method context and report results back to the superpmi host including metrics b superpmi exe runs the near differ to determine if there is a binary diff if so the method context number is added to a list of method contexts with asm diffs including metrics superpmi py calls jit analyze to analyze the results it reads the generated list of metrics for the difs and generates improvement regression information for each metric it is concerned with it needs to generate a machine readable summary of its results including method context numbers in addition to the text markdown format there is current provision for tsv json output superpmi py iterates over the list of method context numbers with diffs that jit analyze says is interesting invoking both the baseline and diff jit to generate textual dasm for developers to examine the work for this should concurrently consider fixing by passing summary of metrics available to superpmi exe on to jit analyze in that case by somehow providing complete aggregate code size not just total code size of functions with diffs to jit analyze dotnet jit contrib comments
| 0
|
204,427
| 23,242,576,207
|
IssuesEvent
|
2022-08-03 16:52:56
|
turkdevops/play-with-docker
|
https://api.github.com/repos/turkdevops/play-with-docker
|
opened
|
CVE-2022-30633 (Medium) detected in github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea, github.com/miekg/dns-v1.0.0
|
security vulnerability
|
## CVE-2022-30633 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea</b>, <b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>
<details><summary><b>github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- :x: **github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea** (Vulnerable Library)
</details>
<details><summary><b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>DNS library in Go</p>
<p>
Dependency Hierarchy:
- :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/play-with-docker/commit/27377d4ea18db54381a8dc972091f3c342337ec9">27377d4ea18db54381a8dc972091f3c342337ec9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-30633 golang: encoding/xml: stack exhaustion in Unmarshal
<p>Publish Date: 2022-05-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-30633>CVE-2022-30633</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-30633">https://security-tracker.debian.org/tracker/CVE-2022-30633</a></p>
<p>Release Date: 2022-05-13</p>
<p>Fix Resolution: go1.17.12,go1.18.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-30633 (Medium) detected in github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea, github.com/miekg/dns-v1.0.0 - ## CVE-2022-30633 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea</b>, <b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>
<details><summary><b>github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- :x: **github.com/golang/net-e18ecbb051101a46fc263334b127c89bc7bff7ea** (Vulnerable Library)
</details>
<details><summary><b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>DNS library in Go</p>
<p>
Dependency Hierarchy:
- :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/play-with-docker/commit/27377d4ea18db54381a8dc972091f3c342337ec9">27377d4ea18db54381a8dc972091f3c342337ec9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-30633 golang: encoding/xml: stack exhaustion in Unmarshal
<p>Publish Date: 2022-05-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-30633>CVE-2022-30633</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-30633">https://security-tracker.debian.org/tracker/CVE-2022-30633</a></p>
<p>Release Date: 2022-05-13</p>
<p>Fix Resolution: go1.17.12,go1.18.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in github com golang net github com miekg dns cve medium severity vulnerability vulnerable libraries github com golang net github com miekg dns github com golang net go supplementary network libraries dependency hierarchy x github com golang net vulnerable library github com miekg dns dns library in go dependency hierarchy x github com miekg dns vulnerable library found in head commit a href found in base branch master vulnerability details cve golang encoding xml stack exhaustion in unmarshal publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
101,872
| 16,529,729,687
|
IssuesEvent
|
2021-05-27 03:11:32
|
alpersonalwebsite/postME
|
https://api.github.com/repos/alpersonalwebsite/postME
|
opened
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz
|
security vulnerability
|
## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/dot-prop</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- optimize-css-assets-webpack-plugin-5.0.4.tgz
- cssnano-4.1.11.tgz
- cssnano-preset-default-4.0.8.tgz
- postcss-merge-longhand-4.0.11.tgz
- stylehacks-4.0.3.tgz
- postcss-selector-parser-3.1.1.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz - ## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/dot-prop</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- optimize-css-assets-webpack-plugin-5.0.4.tgz
- cssnano-4.1.11.tgz
- cssnano-preset-default-4.0.8.tgz
- postcss-merge-longhand-4.0.11.tgz
- stylehacks-4.0.3.tgz
- postcss-selector-parser-3.1.1.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in dot prop tgz cve high severity vulnerability vulnerable library dot prop tgz get set or delete a property from a nested object using a dot path library home page a href path to dependency file postme frontend package json path to vulnerable library postme frontend node modules dot prop dependency hierarchy react scripts tgz root library optimize css assets webpack plugin tgz cssnano tgz cssnano preset default tgz postcss merge longhand tgz stylehacks tgz postcss selector parser tgz x dot prop tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability in dot prop npm package versions before and versions x before allows an attacker to add arbitrary properties to javascript language constructs such as objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution dot prop step up your open source security game with whitesource
| 0
|
168,283
| 26,626,675,213
|
IssuesEvent
|
2023-01-24 15:00:34
|
patternfly/patternfly-design
|
https://api.github.com/repos/patternfly/patternfly-design
|
closed
|
Consider adjusting the visual design of PF's left nav menu.
|
Visual Design PF website
|
Consider better visually distinguishing the PF site left nav menu from the site's background.
- This is an enhancement suggestion based on the [design system competitive analysis.](https://docs.google.com/presentation/d/161XsZ6NrUSgnCkaA1e_96oxWY72VsJAa2SVmtIEhWJg/edit#slide=id.g547716335e_0_260)
- The idea is to have more distinction, inspired by [Carbon](https://carbondesignsystem.com/components/button/usage/), [Clarity](https://clarity.design/documentation/buttons), and [Pajamas](https://design.gitlab.com/components/button/).
- In its current implementation, the contrast between the menu and the site's background is low. There is not significant separation between the menu and the rest of the site.
|
1.0
|
Consider adjusting the visual design of PF's left nav menu. - Consider better visually distinguishing the PF site left nav menu from the site's background.
- This is an enhancement suggestion based on the [design system competitive analysis.](https://docs.google.com/presentation/d/161XsZ6NrUSgnCkaA1e_96oxWY72VsJAa2SVmtIEhWJg/edit#slide=id.g547716335e_0_260)
- The idea is to have more distinction, inspired by [Carbon](https://carbondesignsystem.com/components/button/usage/), [Clarity](https://clarity.design/documentation/buttons), and [Pajamas](https://design.gitlab.com/components/button/).
- In its current implementation, the contrast between the menu and the site's background is low. There is not significant separation between the menu and the rest of the site.
|
non_defect
|
consider adjusting the visual design of pf s left nav menu consider better visually distinguishing the pf site left nav menu from the site s background this is an enhancement suggestion based on the the idea is to have more distinction inspired by and in its current implementation the contrast between the menu and the site s background is low there is not significant separation between the menu and the rest of the site
| 0
|
72,960
| 31,809,212,004
|
IssuesEvent
|
2023-09-13 15:41:35
|
hashicorp/terraform-provider-google
|
https://api.github.com/repos/hashicorp/terraform-provider-google
|
closed
|
`google_monitoring_monitored_project` exponentially slow and spawns thousands of calls to cloudresourcemanager api
|
bug forward/linked service/monitoring-services
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request.
* Please do not leave _+1_ or _me too_ comments, they generate extra noise for issue followers and do not help prioritize the request.
* If you are interested in working on this issue or have submitted a pull request, please leave a comment.
* If an issue is assigned to the `modular-magician` user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If an issue is assigned to a user, that user is claiming responsibility for the issue. If an issue is assigned to `hashibot`, a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Terraform Version
Terraform v1.3.3
on linux_amd64
### Affected Resource(s)
`google_monitoring_monitored_project`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
locals {
projects = jsondecode(file("projects.json"))
}
resource "google_monitoring_monitored_project" "project" {
for_each = toset(local.projects)
metrics_scope = "my-scoping-project"
name = each.value
}
```
`projects.json` is a simple json array with project ids
```json
[
"my-scoped-project"
]
```
### Debug Output
Snippet of output:
https://gist.github.com/danjamesmay/ce3e049831e6f6bb323964717d9f02b6
### Panic Output
n/a
### Expected Behavior
It appears the Google API supports full CRUD methods for interacting with monitored projects, via project ID or project number. The [resource code](https://github.com/hashicorp/terraform-provider-google/blob/main/google/resource_monitoring_monitored_project.go#L115) seems to be doing a lot more than just interacting with this API.
API:
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes/get
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes.projects/create
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes.projects/delete
### Actual Behavior
Thousands upon thousands of calls to `cloudresourcemanager` resulting in rate limiting and very slow time to update state when reading and writing changes to the `google_monitoring_monitored_project` resource.
### Steps to Reproduce
When a project has 150+ projects listed, the time it takes for Terraform to check the state of each monitored project becomes slower and slower as rate limiting and exponential backoff limits Terraform's ability to work properly.
1. `terraform plan`
### Important Factoids
Nothing I can think of out of the ordinary.
### References
<!---
Note Google Cloud customers who are working with a dedicated Technical Account Manager / Customer Engineer: to expedite the investigation and resolution of this issue, please refer to these instructions: https://github.com/hashicorp/terraform-provider-google/wiki/Customer-Contact#raising-gcp-internal-issues-with-the-provider-development-team
--->
b/275113183
|
2.0
|
`google_monitoring_monitored_project` exponentially slow and spawns thousands of calls to cloudresourcemanager api - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request.
* Please do not leave _+1_ or _me too_ comments, they generate extra noise for issue followers and do not help prioritize the request.
* If you are interested in working on this issue or have submitted a pull request, please leave a comment.
* If an issue is assigned to the `modular-magician` user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If an issue is assigned to a user, that user is claiming responsibility for the issue. If an issue is assigned to `hashibot`, a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Terraform Version
Terraform v1.3.3
on linux_amd64
### Affected Resource(s)
`google_monitoring_monitored_project`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
locals {
projects = jsondecode(file("projects.json"))
}
resource "google_monitoring_monitored_project" "project" {
for_each = toset(local.projects)
metrics_scope = "my-scoping-project"
name = each.value
}
```
`projects.json` is a simple json array with project ids
```json
[
"my-scoped-project"
]
```
### Debug Output
Snippet of output:
https://gist.github.com/danjamesmay/ce3e049831e6f6bb323964717d9f02b6
### Panic Output
n/a
### Expected Behavior
It appears the Google API supports full CRUD methods for interacting with monitored projects, via project ID or project number. The [resource code](https://github.com/hashicorp/terraform-provider-google/blob/main/google/resource_monitoring_monitored_project.go#L115) seems to be doing a lot more than just interacting with this API.
API:
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes/get
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes.projects/create
- https://cloud.google.com/monitoring/api/ref_v3/rest/v1/locations.global.metricsScopes.projects/delete
### Actual Behavior
Thousands upon thousands of calls to `cloudresourcemanager` resulting in rate limiting and very slow time to update state when reading and writing changes to the `google_monitoring_monitored_project` resource.
### Steps to Reproduce
When a project has 150+ projects listed, the time it takes for Terraform to check the state of each monitored project becomes slower and slower as rate limiting and exponential backoff limits Terraform's ability to work properly.
1. `terraform plan`
### Important Factoids
Nothing I can think of out of the ordinary.
### References
<!---
Note Google Cloud customers who are working with a dedicated Technical Account Manager / Customer Engineer: to expedite the investigation and resolution of this issue, please refer to these instructions: https://github.com/hashicorp/terraform-provider-google/wiki/Customer-Contact#raising-gcp-internal-issues-with-the-provider-development-team
--->
b/275113183
|
non_defect
|
google monitoring monitored project exponentially slow and spawns thousands of calls to cloudresourcemanager api community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment if an issue is assigned to the modular magician user it is either in the process of being autogenerated or is planned to be autogenerated soon if an issue is assigned to a user that user is claiming responsibility for the issue if an issue is assigned to hashibot a community member has claimed the issue already terraform version terraform on linux affected resource s google monitoring monitored project terraform configuration files tf locals projects jsondecode file projects json resource google monitoring monitored project project for each toset local projects metrics scope my scoping project name each value projects json is a simple json array with project ids json my scoped project debug output snippet of output panic output n a expected behavior it appears the google api supports full crud methods for interacting with monitored projects via project id or project number the seems to be doing a lot more than just interacting with this api api actual behavior thousands upon thousands of calls to cloudresourcemanager resulting in rate limiting and very slow time to update state when reading and writing changes to the google monitoring monitored project resource steps to reproduce when a project has projects listed the time it takes for terraform to check the state of each monitored project becomes slower and slower as rate limiting and exponential backoff limits terraform s ability to work properly terraform plan important factoids nothing i can think of out of the ordinary references note google cloud customers who are working with a dedicated technical account manager customer engineer to expedite the investigation and resolution of this issue please refer to these instructions b
| 0
|
15,116
| 8,768,143,887
|
IssuesEvent
|
2018-12-17 22:00:16
|
ucsd-progsys/liquidhaskell
|
https://api.github.com/repos/ucsd-progsys/liquidhaskell
|
closed
|
Performance with PLE on (full) equational proofs
|
pbe performance
|
... is quite terrible, as shown by the file `Lists.hs` below:
```haskell
{-@ LIQUID "--reflection" @-}
{-@ LIQUID "--automatic-instances=liquidinstanceslocal" @-}
{-
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc0
Time (0.37s) for action ["Lists.lemma_app_assoc0"]
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc1
Time (0.23s) for action ["Lists.lemma_app_assoc1"]
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc2
Time (1.60s) for action ["Lists.lemma_app_assoc2"]
-}
module Lists where
import Prelude hiding ((++))
import Language.Haskell.Liquid.ProofCombinators
{-@ infixr ++ @-}
{-@ reflect ++ @-}
(++) :: [a] -> [a] -> [a]
[] ++ ys = ys
(x:xs) ++ ys = x : (xs ++ ys)
-------------------------------------------------------------------------------
-- Full equational proof
{-@ lemma_app_assoc0 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc0 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc0 [] ys zs
= ([] ++ ys) ++ zs
=== ys ++ zs
=== [] ++ (ys ++ zs)
*** QED
lemma_app_assoc0 (x:xs) ys zs
= ((x:xs) ++ ys) ++ zs
=== (x: (xs ++ ys)) ++ zs
=== x : ((xs ++ ys) ++ zs)
? lemma_app_assoc0 xs ys zs
=== x : (xs ++ (ys ++ zs))
=== (x : xs) ++ (ys ++ zs)
*** QED
-------------------------------------------------------------------------------
-- Short PLE proof
{-@ ple lemma_app_assoc1 @-}
{-@ lemma_app_assoc1 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc1 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc1 [] ys zs = ()
lemma_app_assoc1 (x:xs) ys zs = lemma_app_assoc1 xs ys zs
-------------------------------------------------------------------------------
-- Running PLE on equational proof
{-@ ple lemma_app_assoc2 @-}
{-@ lemma_app_assoc2 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc2 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc2 [] ys zs
= ([] ++ ys) ++ zs
=== ys ++ zs
=== [] ++ (ys ++ zs)
*** QED
lemma_app_assoc2 (x:xs) ys zs
= ((x:xs) ++ ys) ++ zs
=== (x: (xs ++ ys)) ++ zs
=== x : ((xs ++ ys) ++ zs)
? lemma_app_assoc2 xs ys zs
=== x : (xs ++ (ys ++ zs))
=== (x : xs) ++ (ys ++ zs)
*** QED
```
|
True
|
Performance with PLE on (full) equational proofs - ... is quite terrible, as shown by the file `Lists.hs` below:
```haskell
{-@ LIQUID "--reflection" @-}
{-@ LIQUID "--automatic-instances=liquidinstanceslocal" @-}
{-
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc0
Time (0.37s) for action ["Lists.lemma_app_assoc0"]
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc1
Time (0.23s) for action ["Lists.lemma_app_assoc1"]
$ stack exec -- liquid Lists.hs --checks=lemma_app_assoc2
Time (1.60s) for action ["Lists.lemma_app_assoc2"]
-}
module Lists where
import Prelude hiding ((++))
import Language.Haskell.Liquid.ProofCombinators
{-@ infixr ++ @-}
{-@ reflect ++ @-}
(++) :: [a] -> [a] -> [a]
[] ++ ys = ys
(x:xs) ++ ys = x : (xs ++ ys)
-------------------------------------------------------------------------------
-- Full equational proof
{-@ lemma_app_assoc0 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc0 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc0 [] ys zs
= ([] ++ ys) ++ zs
=== ys ++ zs
=== [] ++ (ys ++ zs)
*** QED
lemma_app_assoc0 (x:xs) ys zs
= ((x:xs) ++ ys) ++ zs
=== (x: (xs ++ ys)) ++ zs
=== x : ((xs ++ ys) ++ zs)
? lemma_app_assoc0 xs ys zs
=== x : (xs ++ (ys ++ zs))
=== (x : xs) ++ (ys ++ zs)
*** QED
-------------------------------------------------------------------------------
-- Short PLE proof
{-@ ple lemma_app_assoc1 @-}
{-@ lemma_app_assoc1 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc1 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc1 [] ys zs = ()
lemma_app_assoc1 (x:xs) ys zs = lemma_app_assoc1 xs ys zs
-------------------------------------------------------------------------------
-- Running PLE on equational proof
{-@ ple lemma_app_assoc2 @-}
{-@ lemma_app_assoc2 :: xs:_ -> ys:_ -> zs:_ ->
{ (xs ++ ys) ++ zs = xs ++ (ys ++ zs) } @-}
lemma_app_assoc2 :: [a] -> [a] -> [a] -> ()
lemma_app_assoc2 [] ys zs
= ([] ++ ys) ++ zs
=== ys ++ zs
=== [] ++ (ys ++ zs)
*** QED
lemma_app_assoc2 (x:xs) ys zs
= ((x:xs) ++ ys) ++ zs
=== (x: (xs ++ ys)) ++ zs
=== x : ((xs ++ ys) ++ zs)
? lemma_app_assoc2 xs ys zs
=== x : (xs ++ (ys ++ zs))
=== (x : xs) ++ (ys ++ zs)
*** QED
```
|
non_defect
|
performance with ple on full equational proofs is quite terrible as shown by the file lists hs below haskell liquid reflection liquid automatic instances liquidinstanceslocal stack exec liquid lists hs checks lemma app time for action stack exec liquid lists hs checks lemma app time for action stack exec liquid lists hs checks lemma app time for action module lists where import prelude hiding import language haskell liquid proofcombinators infixr reflect ys ys x xs ys x xs ys full equational proof lemma app xs ys zs xs ys zs xs ys zs lemma app lemma app ys zs ys zs ys zs ys zs qed lemma app x xs ys zs x xs ys zs x xs ys zs x xs ys zs lemma app xs ys zs x xs ys zs x xs ys zs qed short ple proof ple lemma app lemma app xs ys zs xs ys zs xs ys zs lemma app lemma app ys zs lemma app x xs ys zs lemma app xs ys zs running ple on equational proof ple lemma app lemma app xs ys zs xs ys zs xs ys zs lemma app lemma app ys zs ys zs ys zs ys zs qed lemma app x xs ys zs x xs ys zs x xs ys zs x xs ys zs lemma app xs ys zs x xs ys zs x xs ys zs qed
| 0
|
88,044
| 15,800,682,509
|
IssuesEvent
|
2021-04-03 00:48:08
|
DamieFC/linucrash
|
https://api.github.com/repos/DamieFC/linucrash
|
closed
|
CVE-2020-25672 (Medium) detected in soundv5.9 - autoclosed
|
security vulnerability
|
## CVE-2020-25672 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>soundv5.9</b></p></summary>
<p>
<p>Sound sub-system tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tiwai/sound.git>https://git.kernel.org/pub/scm/linux/kernel/git/tiwai/sound.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/DamieFC/linucrash/commit/e2606e963adfc968b3bb3e61b9c0700d59872554">e2606e963adfc968b3bb3e61b9c0700d59872554</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linucrash/net/nfc/llcp_sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Memory leak vulnerability was found in llcp_sock_connect() in Linux kernel through 5.12-rc2.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-25672 (Medium) detected in soundv5.9 - autoclosed - ## CVE-2020-25672 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>soundv5.9</b></p></summary>
<p>
<p>Sound sub-system tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tiwai/sound.git>https://git.kernel.org/pub/scm/linux/kernel/git/tiwai/sound.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/DamieFC/linucrash/commit/e2606e963adfc968b3bb3e61b9c0700d59872554">e2606e963adfc968b3bb3e61b9c0700d59872554</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linucrash/net/nfc/llcp_sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Memory leak vulnerability was found in llcp_sock_connect() in Linux kernel through 5.12-rc2.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library sound sub system tree library home page a href found in head commit a href found in base branch master vulnerable source files linucrash net nfc llcp sock c vulnerability details memory leak vulnerability was found in llcp sock connect in linux kernel through publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href step up your open source security game with whitesource
| 0
|
108,286
| 23,588,898,592
|
IssuesEvent
|
2022-08-23 13:48:00
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Bug ACL / URLS from Joomla 3.7
|
No Code Attached Yet J3 Issue
|
# Bug ACL / URLS from Joomla 3.7
I recently upgraded from 3.6.5 to 3.7 (and 3.7.1, 3.7.2) a site that has existed for a long time. I do all the updates regularly. First tested on a version "DEV" before modifying a version in production "PRO".
(And I do well to do that!)
Since the update of joomla 3.7, I encountered a bug with ALC rights management and URLS (whatever rewritting or not, the bug has nothing to do with the .htaccess ... tests done), I m 'Explain :
## **Context:**
On the site in question, some of the articles (news) are accessible only for defined "subscribers":
- The "public" visitor does not see ACTUALITY in the main menu.
- The "public" visitor sees on the HOME page of the site of the module blocks (Last articles or Article of a category ") with the title / category / the beginning of the intro / read more ...
- If the "public" or "subscriber (unidentified)" visitor clicks on a title / "read more", it is directed to the login page (/ component / users /? View = login) See the whole article after.
## **Bug seen since 3.7:**
If the visitor "public" or "subscriber (unidentified)" clicks on a title / "read more", it is not directed to the login page (/ component / users /? View = login) And see the article in the center of the homepage without any password request. We are not directed to a new page.
After analysis, I have found that the URLS have changed since the update 3.7 ... see the explanatory screenshot, example (The site is en fr-FR only):
**3.7**
/96-adherent-actualite/derniere-actualite/personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html
**3.6.5**
/actualites/actualite-fonction-publique/96-personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html[captures-uk.pdf](https://issues.joomla.org/uploads/1/4970d9463951d5aa4126509c282b2962.pdf)
I made "public" ACTUALITE in the main menu on joomla 3.7. The bug has disappeared, it is the right URL, we are directed to the right page
actualites/actualite-fonction-publique/96-personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html
## **Conclusion:**
It seems that a change has occurred in the URLS management with the CLA since 3.7, or a change of configuration ??
I confirm that I have only made the update. I did not change anything else ... I also tested without rewriting the URLs. I reactivated the .htaccess ... The bug remains persistent.
I have no database-level errors in the Joomla Extension / Manage / Database administration
So I'm looking for an explanation, a help, a solution ... I dry a lot!
Thanks for the help
Didier A.
|
1.0
|
Bug ACL / URLS from Joomla 3.7 - # Bug ACL / URLS from Joomla 3.7
I recently upgraded from 3.6.5 to 3.7 (and 3.7.1, 3.7.2) a site that has existed for a long time. I do all the updates regularly. First tested on a version "DEV" before modifying a version in production "PRO".
(And I do well to do that!)
Since the update of joomla 3.7, I encountered a bug with ALC rights management and URLS (whatever rewritting or not, the bug has nothing to do with the .htaccess ... tests done), I m 'Explain :
## **Context:**
On the site in question, some of the articles (news) are accessible only for defined "subscribers":
- The "public" visitor does not see ACTUALITY in the main menu.
- The "public" visitor sees on the HOME page of the site of the module blocks (Last articles or Article of a category ") with the title / category / the beginning of the intro / read more ...
- If the "public" or "subscriber (unidentified)" visitor clicks on a title / "read more", it is directed to the login page (/ component / users /? View = login) See the whole article after.
## **Bug seen since 3.7:**
If the visitor "public" or "subscriber (unidentified)" clicks on a title / "read more", it is not directed to the login page (/ component / users /? View = login) And see the article in the center of the homepage without any password request. We are not directed to a new page.
After analysis, I have found that the URLS have changed since the update 3.7 ... see the explanatory screenshot, example (The site is en fr-FR only):
**3.7**
/96-adherent-actualite/derniere-actualite/personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html
**3.6.5**
/actualites/actualite-fonction-publique/96-personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html[captures-uk.pdf](https://issues.joomla.org/uploads/1/4970d9463951d5aa4126509c282b2962.pdf)
I made "public" ACTUALITE in the main menu on joomla 3.7. The bug has disappeared, it is the right URL, we are directed to the right page
actualites/actualite-fonction-publique/96-personnels-transferes/252-etablissement-public-national-antoine-koenigswarter.html
## **Conclusion:**
It seems that a change has occurred in the URLS management with the CLA since 3.7, or a change of configuration ??
I confirm that I have only made the update. I did not change anything else ... I also tested without rewriting the URLs. I reactivated the .htaccess ... The bug remains persistent.
I have no database-level errors in the Joomla Extension / Manage / Database administration
So I'm looking for an explanation, a help, a solution ... I dry a lot!
Thanks for the help
Didier A.
|
non_defect
|
bug acl urls from joomla bug acl urls from joomla i recently upgraded from to and a site that has existed for a long time i do all the updates regularly first tested on a version dev before modifying a version in production pro and i do well to do that since the update of joomla i encountered a bug with alc rights management and urls whatever rewritting or not the bug has nothing to do with the htaccess tests done i m explain context on the site in question some of the articles news are accessible only for defined subscribers the public visitor does not see actuality in the main menu the public visitor sees on the home page of the site of the module blocks last articles or article of a category with the title category the beginning of the intro read more if the public or subscriber unidentified visitor clicks on a title read more it is directed to the login page component users view login see the whole article after bug seen since if the visitor public or subscriber unidentified clicks on a title read more it is not directed to the login page component users view login and see the article in the center of the homepage without any password request we are not directed to a new page after analysis i have found that the urls have changed since the update see the explanatory screenshot example the site is en fr fr only adherent actualite derniere actualite personnels transferes etablissement public national antoine koenigswarter html actualites actualite fonction publique personnels transferes etablissement public national antoine koenigswarter html i made public actualite in the main menu on joomla the bug has disappeared it is the right url we are directed to the right page actualites actualite fonction publique personnels transferes etablissement public national antoine koenigswarter html conclusion it seems that a change has occurred in the urls management with the cla since or a change of configuration i confirm that i have only made the update i did not change anything else i also tested without rewriting the urls i reactivated the htaccess the bug remains persistent i have no database level errors in the joomla extension manage database administration so i m looking for an explanation a help a solution i dry a lot thanks for the help didier a
| 0
|
39,942
| 9,762,211,762
|
IssuesEvent
|
2019-06-05 10:43:16
|
google/guava
|
https://api.github.com/repos/google/guava
|
closed
|
InternetDomainName.isValid is not RFC1123 compliant
|
package=net type=defect
|
_[Original issue](https://code.google.com/p/guava-libraries/issues/detail?id=1872) created by **manderson23** on 2014-10-27 at 10:59 PM_
---
InternetDomainName.isValid returns false for the parameter "8server".
RFC1123 suggests that names should be able to start with a digit. See https://groups.google.com/forum/#!topic/guava-discuss/8Sycya7Fkok on guava-discuss for a discussion of the issue.
I also asked for clarification on Server Fault at http://serverfault.com/questions/638260/is-it-valid-for-a-hostname-to-start-with-a-digit
My final suggestion on the mailing list for solving this was as follows:
"What about if the final part starts with a digit actually checking if the complete name is an IP address e.g. calling InetAddreses.isInetAddress? The discussion section in RFC 1123 does suggest a full syntactic check to prevent a dotted decimal number getting through."
but there was no response.
|
1.0
|
InternetDomainName.isValid is not RFC1123 compliant - _[Original issue](https://code.google.com/p/guava-libraries/issues/detail?id=1872) created by **manderson23** on 2014-10-27 at 10:59 PM_
---
InternetDomainName.isValid returns false for the parameter "8server".
RFC1123 suggests that names should be able to start with a digit. See https://groups.google.com/forum/#!topic/guava-discuss/8Sycya7Fkok on guava-discuss for a discussion of the issue.
I also asked for clarification on Server Fault at http://serverfault.com/questions/638260/is-it-valid-for-a-hostname-to-start-with-a-digit
My final suggestion on the mailing list for solving this was as follows:
"What about if the final part starts with a digit actually checking if the complete name is an IP address e.g. calling InetAddreses.isInetAddress? The discussion section in RFC 1123 does suggest a full syntactic check to prevent a dotted decimal number getting through."
but there was no response.
|
defect
|
internetdomainname isvalid is not compliant created by on at pm internetdomainname isvalid returns false for the parameter suggests that names should be able to start with a digit see on guava discuss for a discussion of the issue i also asked for clarification on server fault at my final suggestion on the mailing list for solving this was as follows what about if the final part starts with a digit actually checking if the complete name is an ip address e g calling inetaddreses isinetaddress the discussion section in rfc does suggest a full syntactic check to prevent a dotted decimal number getting through but there was no response
| 1
|
38,851
| 8,971,613,933
|
IssuesEvent
|
2019-01-29 16:17:38
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
closed
|
UX - cannot access room directory in riot-web/experimental
|
bug defect mvp redesign ui/ux
|
hi guys,
I'm coming from Slack/Hipchat. It is really very difficult for me to figure out how to get the list of rooms on my server, so that i can browse and join.
We are planning to use Riot/Modular for our company as a replacement for Slack and we have a lot of rooms that people can potentially join. So a browse behavior is very very important.
|
1.0
|
UX - cannot access room directory in riot-web/experimental - hi guys,
I'm coming from Slack/Hipchat. It is really very difficult for me to figure out how to get the list of rooms on my server, so that i can browse and join.
We are planning to use Riot/Modular for our company as a replacement for Slack and we have a lot of rooms that people can potentially join. So a browse behavior is very very important.
|
defect
|
ux cannot access room directory in riot web experimental hi guys i m coming from slack hipchat it is really very difficult for me to figure out how to get the list of rooms on my server so that i can browse and join we are planning to use riot modular for our company as a replacement for slack and we have a lot of rooms that people can potentially join so a browse behavior is very very important
| 1
|
9,743
| 2,615,167,001
|
IssuesEvent
|
2015-03-01 06:47:43
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
Code (0x02) and (0x03)
|
auto-migrated Priority-Triage Type-Defect
|
```
Hi,
I get these two issue, and I have no idea why. It seems to associate with no
problem. I even tried to associate via aireplay-ng just for the sake of it, no
avail. I did not know how to export the log, but I took a screen shot, so I
will upload it and share the link here. Thanks for the help.
Just noticed that the scrnshot did not capture the command, which was:
reaver -i mon0 -b XX:XX:XX:XX:XX:XX -vv
http://postimg.org/image/3u28n3muz/
And here is scrnshot no.2 with the -d command
http://postimg.org/image/t4i1jyfoh/
```
Original issue reported on code.google.com by `spiderma...@gmail.com` on 15 Oct 2013 at 10:21
|
1.0
|
Code (0x02) and (0x03) - ```
Hi,
I get these two issue, and I have no idea why. It seems to associate with no
problem. I even tried to associate via aireplay-ng just for the sake of it, no
avail. I did not know how to export the log, but I took a screen shot, so I
will upload it and share the link here. Thanks for the help.
Just noticed that the scrnshot did not capture the command, which was:
reaver -i mon0 -b XX:XX:XX:XX:XX:XX -vv
http://postimg.org/image/3u28n3muz/
And here is scrnshot no.2 with the -d command
http://postimg.org/image/t4i1jyfoh/
```
Original issue reported on code.google.com by `spiderma...@gmail.com` on 15 Oct 2013 at 10:21
|
defect
|
code and hi i get these two issue and i have no idea why it seems to associate with no problem i even tried to associate via aireplay ng just for the sake of it no avail i did not know how to export the log but i took a screen shot so i will upload it and share the link here thanks for the help just noticed that the scrnshot did not capture the command which was reaver i b xx xx xx xx xx xx vv and here is scrnshot no with the d command original issue reported on code google com by spiderma gmail com on oct at
| 1
|
66,541
| 20,264,702,345
|
IssuesEvent
|
2022-02-15 10:54:38
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Mounting raw encrypted send/recv'd dataset fails: i/o error
|
Type: Defect
|
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Proxmox 7.1 (Debian)
Distribution Version | 7.1 (debian 11.2)
Kernel Version | 5.13.19-2-pve
Architecture | amd64
OpenZFS Version | zfs-2.1.1
### Describe the problem you're observing
I'm copying (using syncoid) a zfs dataset from a laptop (running debian sid with zfs-2.1.1) to my proxmox box.
Syncoid parameters:
```
$ /usr/sbin/syncoid --sendoptions=w zpool/home target:rpool/backups
```
this dataset is encrypted with a passphrase, using aes-256-gcm.
Unlocking the dataset on the target works:
`1 / 1 key(s) successfully loaded`
but mounting the dataset returns 'I/O error'.
```
# zfs mount rpool/backup
cannot mount 'rpool/backup': Input/output error
```
### Describe how to reproduce the problem
Uncertain
### Include any warning/errors/backtraces from the system logs
None?
|
1.0
|
Mounting raw encrypted send/recv'd dataset fails: i/o error - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Proxmox 7.1 (Debian)
Distribution Version | 7.1 (debian 11.2)
Kernel Version | 5.13.19-2-pve
Architecture | amd64
OpenZFS Version | zfs-2.1.1
### Describe the problem you're observing
I'm copying (using syncoid) a zfs dataset from a laptop (running debian sid with zfs-2.1.1) to my proxmox box.
Syncoid parameters:
```
$ /usr/sbin/syncoid --sendoptions=w zpool/home target:rpool/backups
```
this dataset is encrypted with a passphrase, using aes-256-gcm.
Unlocking the dataset on the target works:
`1 / 1 key(s) successfully loaded`
but mounting the dataset returns 'I/O error'.
```
# zfs mount rpool/backup
cannot mount 'rpool/backup': Input/output error
```
### Describe how to reproduce the problem
Uncertain
### Include any warning/errors/backtraces from the system logs
None?
|
defect
|
mounting raw encrypted send recv d dataset fails i o error system information type version name distribution name proxmox debian distribution version debian kernel version pve architecture openzfs version zfs describe the problem you re observing i m copying using syncoid a zfs dataset from a laptop running debian sid with zfs to my proxmox box syncoid parameters usr sbin syncoid sendoptions w zpool home target rpool backups this dataset is encrypted with a passphrase using aes gcm unlocking the dataset on the target works key s successfully loaded but mounting the dataset returns i o error zfs mount rpool backup cannot mount rpool backup input output error describe how to reproduce the problem uncertain include any warning errors backtraces from the system logs none
| 1
|
63,924
| 18,058,821,110
|
IssuesEvent
|
2021-09-20 11:44:05
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
something went wrong
|
T-Defect
|
### Steps to reproduce
I tried to accept an invite
### What happened?
### What did you expect?
Successful join
### What happened?
"Something went wrong" screen
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Have you submitted a rageshake?
No
|
1.0
|
something went wrong - ### Steps to reproduce
I tried to accept an invite
### What happened?
### What did you expect?
Successful join
### What happened?
"Something went wrong" screen
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Have you submitted a rageshake?
No
|
defect
|
something went wrong steps to reproduce i tried to accept an invite what happened what did you expect successful join what happened something went wrong screen operating system no response application version no response how did you install the app no response homeserver no response have you submitted a rageshake no
| 1
|
59,677
| 17,023,201,418
|
IssuesEvent
|
2021-07-03 00:50:03
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Typing in diary form makes map move around
|
Component: website Priority: major Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 1.46pm, Wednesday, 30th January 2008]**
If you start the map, then start typing in the diary form, the map will move around and zoom in and out as you type. This is slightly strange.
k/+ = zoom in
m/- = zoom out
' = go east
# = go east a lot
|
1.0
|
Typing in diary form makes map move around - **[Submitted to the original trac issue database at 1.46pm, Wednesday, 30th January 2008]**
If you start the map, then start typing in the diary form, the map will move around and zoom in and out as you type. This is slightly strange.
k/+ = zoom in
m/- = zoom out
' = go east
# = go east a lot
|
defect
|
typing in diary form makes map move around if you start the map then start typing in the diary form the map will move around and zoom in and out as you type this is slightly strange k zoom in m zoom out go east go east a lot
| 1
|
194,901
| 22,281,580,978
|
IssuesEvent
|
2022-06-11 01:07:18
|
pazhanivel07/linux-4.19.72
|
https://api.github.com/repos/pazhanivel07/linux-4.19.72
|
reopened
|
CVE-2019-17666 (High) detected in linux-yoctov5.4.51
|
security vulnerability
|
## CVE-2019-17666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
rtl_p2p_noa_ie in drivers/net/wireless/realtek/rtlwifi/ps.c in the Linux kernel through 5.3.6 lacks a certain upper-bound check, leading to a buffer overflow.
<p>Publish Date: 2019-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17666>CVE-2019-17666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-17666">https://nvd.nist.gov/vuln/detail/CVE-2019-17666</a></p>
<p>Release Date: 2019-10-24</p>
<p>Fix Resolution: linux - 5.3.9.1-1;linux-lts - 4.19.82-1;linux-zen - 5.3.9.1-1;linux-hardened - 5.3.7.b-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-17666 (High) detected in linux-yoctov5.4.51 - ## CVE-2019-17666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
rtl_p2p_noa_ie in drivers/net/wireless/realtek/rtlwifi/ps.c in the Linux kernel through 5.3.6 lacks a certain upper-bound check, leading to a buffer overflow.
<p>Publish Date: 2019-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17666>CVE-2019-17666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-17666">https://nvd.nist.gov/vuln/detail/CVE-2019-17666</a></p>
<p>Release Date: 2019-10-24</p>
<p>Fix Resolution: linux - 5.3.9.1-1;linux-lts - 4.19.82-1;linux-zen - 5.3.9.1-1;linux-hardened - 5.3.7.b-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details rtl noa ie in drivers net wireless realtek rtlwifi ps c in the linux kernel through lacks a certain upper bound check leading to a buffer overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux linux lts linux zen linux hardened b step up your open source security game with whitesource
| 0
|
58,443
| 16,540,241,456
|
IssuesEvent
|
2021-05-27 15:55:51
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Setting NFS4 ACLs on exported zfs dataset fails with EIO on linux-5.12
|
Status: Triage Needed Type: Defect
|
### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Distribution Version | latest
Linux Kernel | 5.12.4
Architecture | x86_64
ZFS Version | 2.1.0_rc5
SPL Version | 2.1.0_rc5
### Describe the problem you're observing
Setting an nfs4 acl on an nfs client fails with EIO if the server is exporting from ZFS.
This started happening after upgrading Linux from 5.10 to 5.12, and ZFS to 2.1.0_rc5 alongside it.
I downgraded Linux back to 5.10 afterwards, keeping ZFS 2.1.0_rc5, and the problem disappeared.
I also tested an ext4 based NFS export on Linux 5.12 and could not observe the issue either.
So this specifically happens with ZFS on Linux 5.12.
Looking at strace on failing application on the NFS client, it's specifically the setxattr() syscall that returns EIO.
Though on the NFS server, the file in question does not have any xattrs, since NFS probably is translating the NFS ACL to a posix one.
### Describe how to reproduce the problem
Run `nfs4_editfacl` on any file on the export on the NFS client, and make any random change and save. It will fail with EIO.
Can also be reproduced with pure python:
`python3 -c 'import os, shutil; print(repr(os.listxattr("testfile"))); shutil.copystat("testfile", "testfile.copy")'`
Should first print at least the `system.nfs4_acl` xattr, and then fail in shutil with EIO.
|
1.0
|
Setting NFS4 ACLs on exported zfs dataset fails with EIO on linux-5.12 - ### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Distribution Version | latest
Linux Kernel | 5.12.4
Architecture | x86_64
ZFS Version | 2.1.0_rc5
SPL Version | 2.1.0_rc5
### Describe the problem you're observing
Setting an nfs4 acl on an nfs client fails with EIO if the server is exporting from ZFS.
This started happening after upgrading Linux from 5.10 to 5.12, and ZFS to 2.1.0_rc5 alongside it.
I downgraded Linux back to 5.10 afterwards, keeping ZFS 2.1.0_rc5, and the problem disappeared.
I also tested an ext4 based NFS export on Linux 5.12 and could not observe the issue either.
So this specifically happens with ZFS on Linux 5.12.
Looking at strace on failing application on the NFS client, it's specifically the setxattr() syscall that returns EIO.
Though on the NFS server, the file in question does not have any xattrs, since NFS probably is translating the NFS ACL to a posix one.
### Describe how to reproduce the problem
Run `nfs4_editfacl` on any file on the export on the NFS client, and make any random change and save. It will fail with EIO.
Can also be reproduced with pure python:
`python3 -c 'import os, shutil; print(repr(os.listxattr("testfile"))); shutil.copystat("testfile", "testfile.copy")'`
Should first print at least the `system.nfs4_acl` xattr, and then fail in shutil with EIO.
|
defect
|
setting acls on exported zfs dataset fails with eio on linux system information type version name distribution name gentoo distribution version latest linux kernel architecture zfs version spl version describe the problem you re observing setting an acl on an nfs client fails with eio if the server is exporting from zfs this started happening after upgrading linux from to and zfs to alongside it i downgraded linux back to afterwards keeping zfs and the problem disappeared i also tested an based nfs export on linux and could not observe the issue either so this specifically happens with zfs on linux looking at strace on failing application on the nfs client it s specifically the setxattr syscall that returns eio though on the nfs server the file in question does not have any xattrs since nfs probably is translating the nfs acl to a posix one describe how to reproduce the problem run editfacl on any file on the export on the nfs client and make any random change and save it will fail with eio can also be reproduced with pure python c import os shutil print repr os listxattr testfile shutil copystat testfile testfile copy should first print at least the system acl xattr and then fail in shutil with eio
| 1
|
35,938
| 9,690,686,555
|
IssuesEvent
|
2019-05-24 09:15:03
|
mapbox/mapbox-gl-js
|
https://api.github.com/repos/mapbox/mapbox-gl-js
|
opened
|
Adopt object spread and polyfill Object.assign
|
refactoring :building_construction:
|
Currently we use a [custom `extend` utility function](https://github.com/mapbox/mapbox-gl-js/blob/c7f994115b86c30331e568f9c8d833d8dc35a7b5/src/util/util.js#L142-L159) because `Object.assign` isn't supported in IE11.
However, if we polyfilled `Object.assign` in IE, we could switch to object spread (`{...obj}`) syntax across the code base, which provides many benefits:
- An easier to read, more concise syntax.
- When we ditch IE11 support in future and start providing an untranspiled build, we'll get better performance for free because native spreads are faster than `Object.assign`.
- Strictly typing `extend` and all code that uses in Flow is really hard, which I discovered while working on #8281. It's much easier with spreads because it syntactically indicates the full shape of an object from the start — no type conversions/refinements need to happen to transform an object through multiple extensions.
|
1.0
|
Adopt object spread and polyfill Object.assign - Currently we use a [custom `extend` utility function](https://github.com/mapbox/mapbox-gl-js/blob/c7f994115b86c30331e568f9c8d833d8dc35a7b5/src/util/util.js#L142-L159) because `Object.assign` isn't supported in IE11.
However, if we polyfilled `Object.assign` in IE, we could switch to object spread (`{...obj}`) syntax across the code base, which provides many benefits:
- An easier to read, more concise syntax.
- When we ditch IE11 support in future and start providing an untranspiled build, we'll get better performance for free because native spreads are faster than `Object.assign`.
- Strictly typing `extend` and all code that uses in Flow is really hard, which I discovered while working on #8281. It's much easier with spreads because it syntactically indicates the full shape of an object from the start — no type conversions/refinements need to happen to transform an object through multiple extensions.
|
non_defect
|
adopt object spread and polyfill object assign currently we use a because object assign isn t supported in however if we polyfilled object assign in ie we could switch to object spread obj syntax across the code base which provides many benefits an easier to read more concise syntax when we ditch support in future and start providing an untranspiled build we ll get better performance for free because native spreads are faster than object assign strictly typing extend and all code that uses in flow is really hard which i discovered while working on it s much easier with spreads because it syntactically indicates the full shape of an object from the start — no type conversions refinements need to happen to transform an object through multiple extensions
| 0
|
52,499
| 13,224,789,220
|
IssuesEvent
|
2020-08-17 19:51:06
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
python3 __nonzero__ compatibility (Trac #2341)
|
Incomplete Migration Migrated from Trac combo core defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2341">https://code.icecube.wisc.edu/projects/icecube/ticket/2341</a>, reported by egansterand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-08-20T16:50:37",
"_ts": "1566319837741391",
"description": "Running muon level3 processing using python 3 (py3-v4) results in more events passing the cuts compared to using running it with python 2 (py2-v3.1.1).\nThe problem is the `CleanInputStreams` traysegment in `level3_Functions.py`, especially the following part (lines 79-90):\n{{{\ndef muon_stream(frame):\n if frame.Has(\"FilterMask\"):\n for filter in [\"MuonFilter_12\", \"MuonFilter_13\"]:\n if frame[\"FilterMask\"].has_key(filter) and bool(frame[\"FilterMask\"][filter]):\n return True\n else:\n return False\n if frame.Has(\"QFilterMask\"):\n for filter in [\"EHEFilter_12\", \"EHEFilter_13\"]:\n if frame[\"QFilterMask\"].has_key(filter) and bool(frame[\"QFilterMask\"][filter]):\n return True\n return False\n}}}\n\n`bool(frame[\"FilterMask\"][filter])` will always evaluate to `True` in python 3, regardless of whether the filter mask is actually true or not. This results in events not passing the muon filter to be passed on for level3 processing.\n\nAdditionally, there is one print statement in `level3_Master.py` line 219 that is not yet python 3 compatible.",
"reporter": "eganster",
"cc": "",
"resolution": "fixed",
"time": "2019-08-08T13:40:34",
"component": "combo core",
"summary": "python3 __nonzero__ compatibility",
"priority": "critical",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
python3 __nonzero__ compatibility (Trac #2341) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2341">https://code.icecube.wisc.edu/projects/icecube/ticket/2341</a>, reported by egansterand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-08-20T16:50:37",
"_ts": "1566319837741391",
"description": "Running muon level3 processing using python 3 (py3-v4) results in more events passing the cuts compared to using running it with python 2 (py2-v3.1.1).\nThe problem is the `CleanInputStreams` traysegment in `level3_Functions.py`, especially the following part (lines 79-90):\n{{{\ndef muon_stream(frame):\n if frame.Has(\"FilterMask\"):\n for filter in [\"MuonFilter_12\", \"MuonFilter_13\"]:\n if frame[\"FilterMask\"].has_key(filter) and bool(frame[\"FilterMask\"][filter]):\n return True\n else:\n return False\n if frame.Has(\"QFilterMask\"):\n for filter in [\"EHEFilter_12\", \"EHEFilter_13\"]:\n if frame[\"QFilterMask\"].has_key(filter) and bool(frame[\"QFilterMask\"][filter]):\n return True\n return False\n}}}\n\n`bool(frame[\"FilterMask\"][filter])` will always evaluate to `True` in python 3, regardless of whether the filter mask is actually true or not. This results in events not passing the muon filter to be passed on for level3 processing.\n\nAdditionally, there is one print statement in `level3_Master.py` line 219 that is not yet python 3 compatible.",
"reporter": "eganster",
"cc": "",
"resolution": "fixed",
"time": "2019-08-08T13:40:34",
"component": "combo core",
"summary": "python3 __nonzero__ compatibility",
"priority": "critical",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
defect
|
nonzero compatibility trac migrated from json status closed changetime ts description running muon processing using python results in more events passing the cuts compared to using running it with python nthe problem is the cleaninputstreams traysegment in functions py especially the following part lines n ndef muon stream frame n if frame has filtermask n for filter in n if frame has key filter and bool frame n return true n else n return false n if frame has qfiltermask n for filter in n if frame has key filter and bool frame n return true n return false n n n bool frame will always evaluate to true in python regardless of whether the filter mask is actually true or not this results in events not passing the muon filter to be passed on for processing n nadditionally there is one print statement in master py line that is not yet python compatible reporter eganster cc resolution fixed time component combo core summary nonzero compatibility priority critical keywords milestone autumnal equinox owner kjmeagher type defect
| 1
|
50,010
| 13,187,306,577
|
IssuesEvent
|
2020-08-13 02:59:51
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
editing repo permissions w/o a shell (Trac #1)
|
Migrated from Trac defect infrastructure
|
need some way for people without accounts to edit repository permissions,
for example whoever is in charge of perms for the daq repo
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/1
, reported by troy and owned by cgils_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-06-06T01:55:52",
"description": "need some way for people without accounts to edit repository permissions,\nfor example whoever is in charge of perms for the daq repo",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1181094952000000",
"component": "infrastructure",
"summary": "editing repo permissions w/o a shell",
"priority": "major",
"keywords": "",
"time": "2007-06-03T02:58:19",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
editing repo permissions w/o a shell (Trac #1) - need some way for people without accounts to edit repository permissions,
for example whoever is in charge of perms for the daq repo
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/1
, reported by troy and owned by cgils_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-06-06T01:55:52",
"description": "need some way for people without accounts to edit repository permissions,\nfor example whoever is in charge of perms for the daq repo",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1181094952000000",
"component": "infrastructure",
"summary": "editing repo permissions w/o a shell",
"priority": "major",
"keywords": "",
"time": "2007-06-03T02:58:19",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
defect
|
editing repo permissions w o a shell trac need some way for people without accounts to edit repository permissions for example whoever is in charge of perms for the daq repo migrated from reported by troy and owned by cgils json status closed changetime description need some way for people without accounts to edit repository permissions nfor example whoever is in charge of perms for the daq repo reporter troy cc resolution fixed ts component infrastructure summary editing repo permissions w o a shell priority major keywords time milestone owner cgils type defect
| 1
|
675,133
| 23,080,255,117
|
IssuesEvent
|
2022-07-26 06:24:53
|
ballerina-platform/module-ballerina-c2c
|
https://api.github.com/repos/ballerina-platform/module-ballerina-c2c
|
closed
|
Add descriptions to c2c schema
|
Type/Improvement Priority/Low
|
**Description:**
We need to add descriptions to c2c schema, this will generate comments for create cloud.toml vscode action as well.
**Describe your problem(s)**
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
1.0
|
Add descriptions to c2c schema - **Description:**
We need to add descriptions to c2c schema, this will generate comments for create cloud.toml vscode action as well.
**Describe your problem(s)**
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
non_defect
|
add descriptions to schema description we need to add descriptions to schema this will generate comments for create cloud toml vscode action as well describe your problem s describe your solution s related issues optional suggested labels optional suggested assignees optional
| 0
|
439,032
| 12,676,194,290
|
IssuesEvent
|
2020-06-19 04:19:48
|
a2000-erp-team/WEBERP
|
https://api.github.com/repos/a2000-erp-team/WEBERP
|
opened
|
SAL-POS-POS-REC-[Spelling mistake: Should be (GST INCLUSIVE); Co Reg no & GST Reg no should be mapped to Company maint fields; Stack GST Reg field under Co Reg and center align.]
|
ADRIAN High Priority
|

67. Spelling mistake: Should be (GST INCLUSIVE); Co Reg no & GST Reg no should be mapped to Company maint fields; Stack GST Reg field under Co Reg and center align.
|
1.0
|
SAL-POS-POS-REC-[Spelling mistake: Should be (GST INCLUSIVE); Co Reg no & GST Reg no should be mapped to Company maint fields; Stack GST Reg field under Co Reg and center align.] - 
67. Spelling mistake: Should be (GST INCLUSIVE); Co Reg no & GST Reg no should be mapped to Company maint fields; Stack GST Reg field under Co Reg and center align.
|
non_defect
|
sal pos pos rec spelling mistake should be gst inclusive co reg no gst reg no should be mapped to company maint fields stack gst reg field under co reg and center align
| 0
|
51,956
| 21,919,736,309
|
IssuesEvent
|
2022-05-22 11:42:55
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
closed
|
Add support for `${workspaceFolder}` in C_Cpp.files.exclude/codeAnalysis.exlude
|
Language Service Feature: Configuration more votes needed enhancement
|
It doesn't work with files.exclude because in that case all the references are in regards to the workspace folder, but for the settings we add, it might be good to add support for those variables to avoid confusion, i.e. see issues like https://github.com/microsoft/vscode-cpptools/issues/8792 and others.
|
1.0
|
Add support for `${workspaceFolder}` in C_Cpp.files.exclude/codeAnalysis.exlude - It doesn't work with files.exclude because in that case all the references are in regards to the workspace folder, but for the settings we add, it might be good to add support for those variables to avoid confusion, i.e. see issues like https://github.com/microsoft/vscode-cpptools/issues/8792 and others.
|
non_defect
|
add support for workspacefolder in c cpp files exclude codeanalysis exlude it doesn t work with files exclude because in that case all the references are in regards to the workspace folder but for the settings we add it might be good to add support for those variables to avoid confusion i e see issues like and others
| 0
|
34,557
| 7,457,202,448
|
IssuesEvent
|
2018-03-30 02:35:16
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Andmestruktuuride dokumentatsiooni leht katki
|
C: AIS P: highest R: fixed T: defect
|
**Reported by sven syld on 6 Feb 2014 13:49 UTC**
http://dev.raju.teepub/doc/data/Dira/DescriptionUnitBundle/Model/DescriptionUnit annab vea:
FatalErrorException: Error: Invalid serialization data for DateTime object in /home/dev/dev/htdocs/vendor/jms/serializer/src/JMS/Serializer/Construction/UnserializeObjectConstructor.php line 28
|
1.0
|
Andmestruktuuride dokumentatsiooni leht katki - **Reported by sven syld on 6 Feb 2014 13:49 UTC**
http://dev.raju.teepub/doc/data/Dira/DescriptionUnitBundle/Model/DescriptionUnit annab vea:
FatalErrorException: Error: Invalid serialization data for DateTime object in /home/dev/dev/htdocs/vendor/jms/serializer/src/JMS/Serializer/Construction/UnserializeObjectConstructor.php line 28
|
defect
|
andmestruktuuride dokumentatsiooni leht katki reported by sven syld on feb utc annab vea fatalerrorexception error invalid serialization data for datetime object in home dev dev htdocs vendor jms serializer src jms serializer construction unserializeobjectconstructor php line
| 1
|
30,320
| 6,105,805,513
|
IssuesEvent
|
2017-06-21 01:15:11
|
ironjan/metal-only
|
https://api.github.com/repos/ironjan/metal-only
|
opened
|
Check why minSdk was set to 18
|
defect
|
It was done in 4a404b7042f3ff2c69195899a7ee09288682a787 and seems like an accident. Noticed this on release
|
1.0
|
Check why minSdk was set to 18 - It was done in 4a404b7042f3ff2c69195899a7ee09288682a787 and seems like an accident. Noticed this on release
|
defect
|
check why minsdk was set to it was done in and seems like an accident noticed this on release
| 1
|
69,120
| 22,165,514,789
|
IssuesEvent
|
2022-06-05 06:30:31
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Image banner text should not be selected
|
T-Defect
|
### Steps to reproduce
1. Upload an image to a room
2. Select the text around the image
### Outcome
#### What did you expect?
The image file name should not be selected.
#### What happened instead?
The image file can be selected.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Image banner text should not be selected - ### Steps to reproduce
1. Upload an image to a room
2. Select the text around the image
### Outcome
#### What did you expect?
The image file name should not be selected.
#### What happened instead?
The image file can be selected.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
image banner text should not be selected steps to reproduce upload an image to a room select the text around the image outcome what did you expect the image file name should not be selected what happened instead the image file can be selected operating system no response browser information no response url for webapp localhost application version develop branch homeserver no response will you send logs no
| 1
|
21,694
| 3,547,169,704
|
IssuesEvent
|
2016-01-20 08:19:12
|
evpobr/fictionbookeditor
|
https://api.github.com/repos/evpobr/fictionbookeditor
|
closed
|
Отображение превью FBShell.dll
|
auto-migrated Priority-Medium Type-Defect
|
```
Версия 2.6.6, установлена на Windows 7 Ultimate 64x. Не
отображаются обложки книг при
предварительном просмотре в проводнике.
Попытка зарегистрировать CLSID GUID в
под-раздел
IconHandler(HKEY_CLASSES_ROOT\FictionBook.2\shellex\IconHandler), как это
описано в MSDN:
http://msdn.microsoft.com/en-us/library/cc144122%28v=vs.85%29.aspx ,
результатов не дала. Сейчас GUID почему-то
пишется в ключ {BB2E617C-0920-11d1-9A0B-00C04FC2D6C1}. Кроме
того, не происходит запись GUID, привязанного
к функционалу просмотра, в раздел
Approved(HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Shell
Extensions\Approved). Ручная правка также не дала
результатов. На 32-битной системе с теме же
книгами предварительный просмотр работал
нормально. Можно предположить, что
проблема в работе FBshell.dll под 64-битной ОС.
```
Original issue reported on code.google.com by `C4.Grey` on 3 Dec 2012 at 9:57
|
1.0
|
Отображение превью FBShell.dll - ```
Версия 2.6.6, установлена на Windows 7 Ultimate 64x. Не
отображаются обложки книг при
предварительном просмотре в проводнике.
Попытка зарегистрировать CLSID GUID в
под-раздел
IconHandler(HKEY_CLASSES_ROOT\FictionBook.2\shellex\IconHandler), как это
описано в MSDN:
http://msdn.microsoft.com/en-us/library/cc144122%28v=vs.85%29.aspx ,
результатов не дала. Сейчас GUID почему-то
пишется в ключ {BB2E617C-0920-11d1-9A0B-00C04FC2D6C1}. Кроме
того, не происходит запись GUID, привязанного
к функционалу просмотра, в раздел
Approved(HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Shell
Extensions\Approved). Ручная правка также не дала
результатов. На 32-битной системе с теме же
книгами предварительный просмотр работал
нормально. Можно предположить, что
проблема в работе FBshell.dll под 64-битной ОС.
```
Original issue reported on code.google.com by `C4.Grey` on 3 Dec 2012 at 9:57
|
defect
|
отображение превью fbshell dll версия установлена на windows ultimate не отображаются обложки книг при предварительном просмотре в проводнике попытка зарегистрировать clsid guid в под раздел iconhandler hkey classes root fictionbook shellex iconhandler как это описано в msdn результатов не дала сейчас guid почему то пишется в ключ кроме того не происходит запись guid привязанного к функционалу просмотра в раздел approved hkey local machine software microsoft windows currentversion shell extensions approved ручная правка также не дала результатов на битной системе с теме же книгами предварительный просмотр работал нормально можно предположить что проблема в работе fbshell dll под битной ос original issue reported on code google com by grey on dec at
| 1
|
50,414
| 13,187,481,120
|
IssuesEvent
|
2020-08-13 03:33:02
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
bzip writer craps out on empty files (Trac #654)
|
Migrated from Trac dataio defect
|
As reported by Dipo:
Hi,
trying to extract the SLOP stream from L2 prime files results in a few
empty _SLOP files, this results in an bzip2 error if a ".bz2"
extension is given for the output _SLOP file.
======
File "/net/user/i3filter/IC79_L2_Prime/icerec/RHEL_6.0_amd64_V04_00_01/lib/I3Tray.py",
line 126, in Finish
self.the_tray.Finish()
RuntimeError: bzip2 error
=======
It seems happy to produce empty files with ".gz" extensions.
Dipo
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/654
, reported by blaufuss and owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-08T16:26:45",
"description": "As reported by Dipo:\n\nHi,\ntrying to extract the SLOP stream from L2 prime files results in a few\nempty _SLOP files, this results in an bzip2 error if a \".bz2\"\nextension is given for the output _SLOP file.\n\n======\nFile \"/net/user/i3filter/IC79_L2_Prime/icerec/RHEL_6.0_amd64_V04_00_01/lib/I3Tray.py\",\nline 126, in Finish\n self.the_tray.Finish()\nRuntimeError: bzip2 error\n=======\n\nIt seems happy to produce empty files with \".gz\" extensions.\n\n\nDipo",
"reporter": "blaufuss",
"cc": "",
"resolution": "wontfix",
"_ts": "1412785605024440",
"component": "dataio",
"summary": "bzip writer craps out on empty files",
"priority": "normal",
"keywords": "",
"time": "2011-10-12T17:45:43",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
bzip writer craps out on empty files (Trac #654) - As reported by Dipo:
Hi,
trying to extract the SLOP stream from L2 prime files results in a few
empty _SLOP files, this results in an bzip2 error if a ".bz2"
extension is given for the output _SLOP file.
======
File "/net/user/i3filter/IC79_L2_Prime/icerec/RHEL_6.0_amd64_V04_00_01/lib/I3Tray.py",
line 126, in Finish
self.the_tray.Finish()
RuntimeError: bzip2 error
=======
It seems happy to produce empty files with ".gz" extensions.
Dipo
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/654
, reported by blaufuss and owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-08T16:26:45",
"description": "As reported by Dipo:\n\nHi,\ntrying to extract the SLOP stream from L2 prime files results in a few\nempty _SLOP files, this results in an bzip2 error if a \".bz2\"\nextension is given for the output _SLOP file.\n\n======\nFile \"/net/user/i3filter/IC79_L2_Prime/icerec/RHEL_6.0_amd64_V04_00_01/lib/I3Tray.py\",\nline 126, in Finish\n self.the_tray.Finish()\nRuntimeError: bzip2 error\n=======\n\nIt seems happy to produce empty files with \".gz\" extensions.\n\n\nDipo",
"reporter": "blaufuss",
"cc": "",
"resolution": "wontfix",
"_ts": "1412785605024440",
"component": "dataio",
"summary": "bzip writer craps out on empty files",
"priority": "normal",
"keywords": "",
"time": "2011-10-12T17:45:43",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
bzip writer craps out on empty files trac as reported by dipo hi trying to extract the slop stream from prime files results in a few empty slop files this results in an error if a extension is given for the output slop file file net user prime icerec rhel lib py line in finish self the tray finish runtimeerror error it seems happy to produce empty files with gz extensions dipo migrated from reported by blaufuss and owned by blaufuss json status closed changetime description as reported by dipo n nhi ntrying to extract the slop stream from prime files results in a few nempty slop files this results in an error if a nextension is given for the output slop file n n nfile net user prime icerec rhel lib py nline in finish n self the tray finish nruntimeerror error n n nit seems happy to produce empty files with gz extensions n n ndipo reporter blaufuss cc resolution wontfix ts component dataio summary bzip writer craps out on empty files priority normal keywords time milestone owner blaufuss type defect
| 1
|
47,254
| 5,872,817,732
|
IssuesEvent
|
2017-05-15 12:36:37
|
PSEBergclubBern/BergclubBern
|
https://api.github.com/repos/PSEBergclubBern/BergclubBern
|
closed
|
Kundentests: Unterlagen für die verschiedenen Rollen/Use-Cases
|
testing
|
Test/Feedback-Unterlagen bereitstellen.
Aktueller Stand Verfügbarkeit (gemäss Kunde):
Christina als Sekretariat und Tourenleiterin ist ab Osterdienstag verfügbar.
Gallus als Kassier am 21./22. April und ab 01.05.2017 verfügbar.
Ruedi als Tourenchef und Tourenleiter ab 24.04. verfügbar.
Franziska als Redaktion ab Osterdienstag vefügbar . 26.-29.04.2017 abwesend.
Bettina als Tourenchef und Tourenleiter BCB Jugend ab 24.04.verfügbar.
Esther als Mutationen ist verfügbar.
Meine Wenigkeit ist verfügbar.
Tourenleiter die bis jetzt nicht involviert waren:
Jürg Anliker, Tourenleiter
Elisabeth Christen Nussbaum
|
1.0
|
Kundentests: Unterlagen für die verschiedenen Rollen/Use-Cases - Test/Feedback-Unterlagen bereitstellen.
Aktueller Stand Verfügbarkeit (gemäss Kunde):
Christina als Sekretariat und Tourenleiterin ist ab Osterdienstag verfügbar.
Gallus als Kassier am 21./22. April und ab 01.05.2017 verfügbar.
Ruedi als Tourenchef und Tourenleiter ab 24.04. verfügbar.
Franziska als Redaktion ab Osterdienstag vefügbar . 26.-29.04.2017 abwesend.
Bettina als Tourenchef und Tourenleiter BCB Jugend ab 24.04.verfügbar.
Esther als Mutationen ist verfügbar.
Meine Wenigkeit ist verfügbar.
Tourenleiter die bis jetzt nicht involviert waren:
Jürg Anliker, Tourenleiter
Elisabeth Christen Nussbaum
|
non_defect
|
kundentests unterlagen für die verschiedenen rollen use cases test feedback unterlagen bereitstellen aktueller stand verfügbarkeit gemäss kunde christina als sekretariat und tourenleiterin ist ab osterdienstag verfügbar gallus als kassier am april und ab verfügbar ruedi als tourenchef und tourenleiter ab verfügbar franziska als redaktion ab osterdienstag vefügbar abwesend bettina als tourenchef und tourenleiter bcb jugend ab verfügbar esther als mutationen ist verfügbar meine wenigkeit ist verfügbar tourenleiter die bis jetzt nicht involviert waren jürg anliker tourenleiter elisabeth christen nussbaum
| 0
|
688,691
| 23,592,788,575
|
IssuesEvent
|
2022-08-23 16:31:59
|
gravityview/GravityView
|
https://api.github.com/repos/gravityview/GravityView
|
closed
|
WPML: Search Bar search fields don't get translated
|
Compat: Plugin Core: I18n Difficulty: Medium Priority: Medium Core: Search Bar
|
Should use the correct language drop-down options, instead of only the primary language options.
See https://secure.helpscout.net/conversation/285832163/7577/
┆Issue is synchronized with this [Asana task](https://app.asana.com/0/995529792029955/995651954915336)
|
1.0
|
WPML: Search Bar search fields don't get translated - Should use the correct language drop-down options, instead of only the primary language options.
See https://secure.helpscout.net/conversation/285832163/7577/
┆Issue is synchronized with this [Asana task](https://app.asana.com/0/995529792029955/995651954915336)
|
non_defect
|
wpml search bar search fields don t get translated should use the correct language drop down options instead of only the primary language options see ┆issue is synchronized with this
| 0
|
55,134
| 14,234,499,263
|
IssuesEvent
|
2020-11-18 13:40:54
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Ambiguous field identifier when parsing unions with unqualified columns and parseWithMetaLookups
|
C: Parser E: All Editions P: Medium T: Defect
|
Given this schema:
```sql
create table t1 (a int);
create table t2 (a int);
```
Parsing this valid SQL:
```sql
select a
from t1
union all
select a
from t2
```
Yields:
```
Ambiguous field identifier: [2:1] select a
[*]from t1
union all
select a
from t2
```
This may well be a 3.15.0 regression that isn't productive yet
----
Related: https://github.com/jOOQ/jOOQ/issues/10699
|
1.0
|
Ambiguous field identifier when parsing unions with unqualified columns and parseWithMetaLookups - Given this schema:
```sql
create table t1 (a int);
create table t2 (a int);
```
Parsing this valid SQL:
```sql
select a
from t1
union all
select a
from t2
```
Yields:
```
Ambiguous field identifier: [2:1] select a
[*]from t1
union all
select a
from t2
```
This may well be a 3.15.0 regression that isn't productive yet
----
Related: https://github.com/jOOQ/jOOQ/issues/10699
|
defect
|
ambiguous field identifier when parsing unions with unqualified columns and parsewithmetalookups given this schema sql create table a int create table a int parsing this valid sql sql select a from union all select a from yields ambiguous field identifier select a from union all select a from this may well be a regression that isn t productive yet related
| 1
|
163,779
| 25,869,660,482
|
IssuesEvent
|
2022-12-14 00:58:26
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
add configuration to remove rounded corners
|
ux *as-designed
|
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
## Existing Issues
- #151968 is closed and locked
<!-- Describe the feature you'd like. -->
## Summary

rounded corners are a very subjective topic. such a [drastic change](https://code.visualstudio.com/updates/v1_71#_rounded-buttons) should be provided a way to disable it.

there was an option to disable rounded selection, but there weren't one to disable rounded buttons.
|
1.0
|
add configuration to remove rounded corners - <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
## Existing Issues
- #151968 is closed and locked
<!-- Describe the feature you'd like. -->
## Summary

rounded corners are a very subjective topic. such a [drastic change](https://code.visualstudio.com/updates/v1_71#_rounded-buttons) should be provided a way to disable it.

there was an option to disable rounded selection, but there weren't one to disable rounded buttons.
|
non_defect
|
add configuration to remove rounded corners existing issues is closed and locked summary rounded corners are a very subjective topic such a should be provided a way to disable it there was an option to disable rounded selection but there weren t one to disable rounded buttons
| 0
|
3,928
| 2,610,084,195
|
IssuesEvent
|
2015-02-26 18:25:40
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳彩光祛痘痘
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳彩光祛痘痘【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:56
|
1.0
|
深圳彩光祛痘痘 - ```
深圳彩光祛痘痘【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:56
|
defect
|
深圳彩光祛痘痘 深圳彩光祛痘痘【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at
| 1
|
349,417
| 10,469,027,956
|
IssuesEvent
|
2019-09-22 17:48:29
|
MarcTowler/Discord-RPG-Bot
|
https://api.github.com/repos/MarcTowler/Discord-RPG-Bot
|
opened
|
Bring in a random NPC for the inn
|
Low Priority enhancement
|
a barmaid of sorts, drops random dialogue, potentially answer stuff like 8ball with a "she ignores the question" for anything not scripted...
|
1.0
|
Bring in a random NPC for the inn - a barmaid of sorts, drops random dialogue, potentially answer stuff like 8ball with a "she ignores the question" for anything not scripted...
|
non_defect
|
bring in a random npc for the inn a barmaid of sorts drops random dialogue potentially answer stuff like with a she ignores the question for anything not scripted
| 0
|
4,506
| 2,610,110,473
|
IssuesEvent
|
2015-02-26 18:34:10
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Wordpress slugs are reset to default when publishing an edit.
|
auto-migrated Milestone-1.0a2 Priority-Critical Type-Defect wordpress
|
```
What's the problem?
My modified permalinks get reset to the defaults when I publish from
scribefire-chrome.
Also, the blank lines in wordpress are not recognized as paragraph breaks.
What version of ScribeFire for Chrome are you running?
0.1.0.0
```
-----
Original issue reported on code.google.com by `robjhyndman` on 9 Apr 2010 at 1:57
|
1.0
|
Wordpress slugs are reset to default when publishing an edit. - ```
What's the problem?
My modified permalinks get reset to the defaults when I publish from
scribefire-chrome.
Also, the blank lines in wordpress are not recognized as paragraph breaks.
What version of ScribeFire for Chrome are you running?
0.1.0.0
```
-----
Original issue reported on code.google.com by `robjhyndman` on 9 Apr 2010 at 1:57
|
defect
|
wordpress slugs are reset to default when publishing an edit what s the problem my modified permalinks get reset to the defaults when i publish from scribefire chrome also the blank lines in wordpress are not recognized as paragraph breaks what version of scribefire for chrome are you running original issue reported on code google com by robjhyndman on apr at
| 1
|
286,947
| 21,630,966,974
|
IssuesEvent
|
2022-05-05 09:39:43
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
closed
|
[Docs] #13308 [Bug]: occasional cyclic dependency on user's app
|
Documentation User Education Pod
|
> TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | FE Coders Pod
**Parent Ticket** | #13308
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
1.0
|
[Docs] #13308 [Bug]: occasional cyclic dependency on user's app - > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | FE Coders Pod
**Parent Ticket** | #13308
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
non_defect
|
occasional cyclic dependency on user s app todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod fe coders pod parent ticket engineer release date live date first draft auto assign priority environment
| 0
|
659,807
| 21,942,018,345
|
IssuesEvent
|
2022-05-23 19:10:33
|
googleapis/gax-php
|
https://api.github.com/repos/googleapis/gax-php
|
closed
|
syntax error, unexpected token "readonly", expecting identifier
|
priority: p2 type: bug
|
Thanks for stopping by to let us know something could be better!
**PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response.
Please run down the following list and make sure you've tried the usual "quick fixes":
- Search the issues already opened: https://github.com/googleapis/gax-php/issues
- Search StackOverflow: http://stackoverflow.com/questions/tagged/google-cloud-platform+php
If you are still having issues, please be sure to include as much information as possible:
#### Environment details
- OS: OSX M1 Chip
- PHP version: 8.1
- Package name and version: colopl/laravel-spanner
Repo that is using this package - https://github.com/colopl/laravel-spanner
#### Steps to reproduce
ParseError {#2663
#message: "syntax error, unexpected token "readonly", expecting identifier"
#code: 0
#file: "./vendor/google/cloud-spanner/src/V1/TransactionOptions/ReadOnly.php"
#line: 16
trace: {
./vendor/google/cloud-spanner/src/V1/TransactionOptions/ReadOnly.php:16 { …}
./vendor/composer/ClassLoader.php:428 { …}
./vendor/google/gax/src/Serializer.php:363 { …}
./vendor/google/gax/src/Serializer.php:416 { …}
./vendor/google/gax/src/Serializer.php:368 { …}
./vendor/google/gax/src/Serializer.php:416 { …}
./vendor/google/gax/src/Serializer.php:126 { …}
./vendor/google/cloud-spanner/src/Connection/Grpc.php:1192 { …}
./vendor/google/cloud-spanner/src/Connection/Grpc.php:790 { …}
./vendor/google/cloud-spanner/src/Operation.php:234 { …}
./vendor/google/cloud-spanner/src/Result.php:190 { …}
Google\Cloud\Spanner\Result->Google\Cloud\Spanner\{closure}() {}
./vendor/google/cloud-core/src/ExponentialBackoff.php:80 { …}
./vendor/google/cloud-spanner/src/Result.php:192 { …}
Google\Cloud\Spanner\Result->rows() {}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:246 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:428 { …}
./vendor/laravel/framework/src/Illuminate/Database/Connection.php:672 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:247 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2484 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Query/Builder.php:177 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2472 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:3006 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2473 { …}
./app/Console/Commands/Test.php:56 {
App\Console\Commands\Test->handle()
› ->orderBy('id')
› ->get();
}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36 { …}
./vendor/laravel/framework/src/Illuminate/Container/Util.php:41 { …}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:93 { …}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:37 { …}
./vendor/laravel/framework/src/Illuminate/Container/Container.php:653 { …}
./vendor/laravel/framework/src/Illuminate/Console/Command.php:136 { …}
./vendor/symfony/console/Command/Command.php:307 { …}
./vendor/laravel/framework/src/Illuminate/Console/Command.php:121 { …}
./vendor/symfony/console/Application.php:989 { …}
./vendor/symfony/console/Application.php:299 { …}
./vendor/symfony/console/Application.php:171 { …}
./vendor/laravel/framework/src/Illuminate/Console/Application.php:102 { …}
./vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:129 { …}
./artisan:37 { …}
}
}
#### Code example
```php
# example
```
Regular DB query
Hello!
I am receiving the following error. Could it be because in PHP 8.1 we cannot use ReadOnly as class name? I have seen that renaming class name helped, for example - https://github.com/schmittjoh/serializer/issues/1331 .
Making sure to follow these steps will guarantee the quickest resolution possible.
Thanks!
|
1.0
|
syntax error, unexpected token "readonly", expecting identifier - Thanks for stopping by to let us know something could be better!
**PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response.
Please run down the following list and make sure you've tried the usual "quick fixes":
- Search the issues already opened: https://github.com/googleapis/gax-php/issues
- Search StackOverflow: http://stackoverflow.com/questions/tagged/google-cloud-platform+php
If you are still having issues, please be sure to include as much information as possible:
#### Environment details
- OS: OSX M1 Chip
- PHP version: 8.1
- Package name and version: colopl/laravel-spanner
Repo that is using this package - https://github.com/colopl/laravel-spanner
#### Steps to reproduce
ParseError {#2663
#message: "syntax error, unexpected token "readonly", expecting identifier"
#code: 0
#file: "./vendor/google/cloud-spanner/src/V1/TransactionOptions/ReadOnly.php"
#line: 16
trace: {
./vendor/google/cloud-spanner/src/V1/TransactionOptions/ReadOnly.php:16 { …}
./vendor/composer/ClassLoader.php:428 { …}
./vendor/google/gax/src/Serializer.php:363 { …}
./vendor/google/gax/src/Serializer.php:416 { …}
./vendor/google/gax/src/Serializer.php:368 { …}
./vendor/google/gax/src/Serializer.php:416 { …}
./vendor/google/gax/src/Serializer.php:126 { …}
./vendor/google/cloud-spanner/src/Connection/Grpc.php:1192 { …}
./vendor/google/cloud-spanner/src/Connection/Grpc.php:790 { …}
./vendor/google/cloud-spanner/src/Operation.php:234 { …}
./vendor/google/cloud-spanner/src/Result.php:190 { …}
Google\Cloud\Spanner\Result->Google\Cloud\Spanner\{closure}() {}
./vendor/google/cloud-core/src/ExponentialBackoff.php:80 { …}
./vendor/google/cloud-spanner/src/Result.php:192 { …}
Google\Cloud\Spanner\Result->rows() {}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:246 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:428 { …}
./vendor/laravel/framework/src/Illuminate/Database/Connection.php:672 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Connection.php:247 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2484 { …}
./vendor/colopl/laravel-spanner/src/Colopl/Spanner/Query/Builder.php:177 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2472 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:3006 { …}
./vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php:2473 { …}
./app/Console/Commands/Test.php:56 {
App\Console\Commands\Test->handle()
› ->orderBy('id')
› ->get();
}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36 { …}
./vendor/laravel/framework/src/Illuminate/Container/Util.php:41 { …}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:93 { …}
./vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:37 { …}
./vendor/laravel/framework/src/Illuminate/Container/Container.php:653 { …}
./vendor/laravel/framework/src/Illuminate/Console/Command.php:136 { …}
./vendor/symfony/console/Command/Command.php:307 { …}
./vendor/laravel/framework/src/Illuminate/Console/Command.php:121 { …}
./vendor/symfony/console/Application.php:989 { …}
./vendor/symfony/console/Application.php:299 { …}
./vendor/symfony/console/Application.php:171 { …}
./vendor/laravel/framework/src/Illuminate/Console/Application.php:102 { …}
./vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:129 { …}
./artisan:37 { …}
}
}
#### Code example
```php
# example
```
Regular DB query
Hello!
I am receiving the following error. Could it be because in PHP 8.1 we cannot use ReadOnly as class name? I have seen that renaming class name helped, for example - https://github.com/schmittjoh/serializer/issues/1331 .
Making sure to follow these steps will guarantee the quickest resolution possible.
Thanks!
|
non_defect
|
syntax error unexpected token readonly expecting identifier thanks for stopping by to let us know something could be better please read if you have a support contract with google please create an issue in the instead of filing on github this will ensure a timely response please run down the following list and make sure you ve tried the usual quick fixes search the issues already opened search stackoverflow if you are still having issues please be sure to include as much information as possible environment details os osx chip php version package name and version colopl laravel spanner repo that is using this package steps to reproduce parseerror message syntax error unexpected token readonly expecting identifier code file vendor google cloud spanner src transactionoptions readonly php line trace vendor google cloud spanner src transactionoptions readonly php … vendor composer classloader php … vendor google gax src serializer php … vendor google gax src serializer php … vendor google gax src serializer php … vendor google gax src serializer php … vendor google gax src serializer php … vendor google cloud spanner src connection grpc php … vendor google cloud spanner src connection grpc php … vendor google cloud spanner src operation php … vendor google cloud spanner src result php … google cloud spanner result google cloud spanner closure vendor google cloud core src exponentialbackoff php … vendor google cloud spanner src result php … google cloud spanner result rows vendor colopl laravel spanner src colopl spanner connection php … vendor colopl laravel spanner src colopl spanner connection php … vendor laravel framework src illuminate database connection php … vendor colopl laravel spanner src colopl spanner connection php … vendor laravel framework src illuminate database query builder php … vendor colopl laravel spanner src colopl spanner query builder php … vendor laravel framework src illuminate database query builder php … vendor laravel framework src illuminate database query builder php … vendor laravel framework src illuminate database query builder php … app console commands test php app console commands test handle › orderby id › get vendor laravel framework src illuminate container boundmethod php … vendor laravel framework src illuminate container util php … vendor laravel framework src illuminate container boundmethod php … vendor laravel framework src illuminate container boundmethod php … vendor laravel framework src illuminate container container php … vendor laravel framework src illuminate console command php … vendor symfony console command command php … vendor laravel framework src illuminate console command php … vendor symfony console application php … vendor symfony console application php … vendor symfony console application php … vendor laravel framework src illuminate console application php … vendor laravel framework src illuminate foundation console kernel php … artisan … code example php example regular db query hello i am receiving the following error could it be because in php we cannot use readonly as class name i have seen that renaming class name helped for example making sure to follow these steps will guarantee the quickest resolution possible thanks
| 0
|
435,837
| 12,542,107,447
|
IssuesEvent
|
2020-06-05 13:31:24
|
ComPWA/expertsystem
|
https://api.github.com/repos/ComPWA/expertsystem
|
closed
|
Create template for Architecture Decision Records
|
Priority: Medium Type: Maintenance
|
Some nice resources on ADRs:
- [Homepage of the ADR GitHub organization](https://adr.github.io/)
- [SATURN 2017 Talk: Architecture Decision Records in Action](https://www.youtube.com/watch?v=41NVge3_cYo)
- [Markdown Architectural Decision](https://adr.github.io/madr/)
Probably best to follow the structure of the [adr/madr](https://github.com/adr/madr) repo.
|
1.0
|
Create template for Architecture Decision Records - Some nice resources on ADRs:
- [Homepage of the ADR GitHub organization](https://adr.github.io/)
- [SATURN 2017 Talk: Architecture Decision Records in Action](https://www.youtube.com/watch?v=41NVge3_cYo)
- [Markdown Architectural Decision](https://adr.github.io/madr/)
Probably best to follow the structure of the [adr/madr](https://github.com/adr/madr) repo.
|
non_defect
|
create template for architecture decision records some nice resources on adrs probably best to follow the structure of the repo
| 0
|
86,228
| 15,755,443,110
|
IssuesEvent
|
2021-03-31 01:47:23
|
jesper247/learnGitBranching
|
https://api.github.com/repos/jesper247/learnGitBranching
|
opened
|
CVE-2020-8244 (Medium) detected in bl-1.2.2.tgz
|
security vulnerability
|
## CVE-2020-8244 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bl-1.2.2.tgz</b></p></summary>
<p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p>
<p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.2.tgz">https://registry.npmjs.org/bl/-/bl-1.2.2.tgz</a></p>
<p>Path to dependency file: learnGitBranching/package.json</p>
<p>Path to vulnerable library: learnGitBranching/node_modules/bl/package.json</p>
<p>
Dependency Hierarchy:
- vinyl-buffer-1.0.1.tgz (Root Library)
- :x: **bl-1.2.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls.
<p>Publish Date: 2020-08-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.2.1,3.0.1,4.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8244 (Medium) detected in bl-1.2.2.tgz - ## CVE-2020-8244 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bl-1.2.2.tgz</b></p></summary>
<p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p>
<p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.2.tgz">https://registry.npmjs.org/bl/-/bl-1.2.2.tgz</a></p>
<p>Path to dependency file: learnGitBranching/package.json</p>
<p>Path to vulnerable library: learnGitBranching/node_modules/bl/package.json</p>
<p>
Dependency Hierarchy:
- vinyl-buffer-1.0.1.tgz (Root Library)
- :x: **bl-1.2.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls.
<p>Publish Date: 2020-08-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.2.1,3.0.1,4.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in bl tgz cve medium severity vulnerability vulnerable library bl tgz buffer list collect buffers and access with a standard readable buffer interface streamable too library home page a href path to dependency file learngitbranching package json path to vulnerable library learngitbranching node modules bl package json dependency hierarchy vinyl buffer tgz root library x bl tgz vulnerable library vulnerability details a buffer over read vulnerability exists in bl and which could allow an attacker to supply user input even typed that if it ends up in consume argument and can become negative the bufferlist state can be corrupted tricking it into exposing uninitialized memory via regular slice calls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
78,015
| 27,277,446,291
|
IssuesEvent
|
2023-02-23 07:09:27
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Threads don't show preview in timeline and an empty thread panel
|
T-Defect
|
### Steps to reproduce
Currently the last two threads in https://matrix.to/#/#compound:matrix.org don't work for me.
### Outcome
#### What did you expect?
Threads to show a preview in the timeline and the messages in the thread in the threads panel.
#### What happened instead?
They render a blank preview in the timeline:

And the threads panel stays blank, too, when I click one of them:

Links to messages:
* https://matrix.to/#/!YJUhVuekhcNbIWyeId:matrix.org/$Do3nroN5f5KeIrt0QRMZIjNeV3Z7diKp1mYxCUsKNeE?via=matrix.org&via=element.io&via=t2l.io
* https://matrix.to/#/!YJUhVuekhcNbIWyeId:matrix.org/$5MMLRTRCXiyOHSoUXMo4AgJAM22SJd4ZfRKXfDgr_DI?via=matrix.org&via=element.io&via=t2l.io
### Operating system
macOS
### Application version
Element Desktop nightly
### How did you install the app?
https://packages.element.io/nightly/
### Homeserver
element.io
### Will you send logs?
Yes
|
1.0
|
Threads don't show preview in timeline and an empty thread panel - ### Steps to reproduce
Currently the last two threads in https://matrix.to/#/#compound:matrix.org don't work for me.
### Outcome
#### What did you expect?
Threads to show a preview in the timeline and the messages in the thread in the threads panel.
#### What happened instead?
They render a blank preview in the timeline:

And the threads panel stays blank, too, when I click one of them:

Links to messages:
* https://matrix.to/#/!YJUhVuekhcNbIWyeId:matrix.org/$Do3nroN5f5KeIrt0QRMZIjNeV3Z7diKp1mYxCUsKNeE?via=matrix.org&via=element.io&via=t2l.io
* https://matrix.to/#/!YJUhVuekhcNbIWyeId:matrix.org/$5MMLRTRCXiyOHSoUXMo4AgJAM22SJd4ZfRKXfDgr_DI?via=matrix.org&via=element.io&via=t2l.io
### Operating system
macOS
### Application version
Element Desktop nightly
### How did you install the app?
https://packages.element.io/nightly/
### Homeserver
element.io
### Will you send logs?
Yes
|
defect
|
threads don t show preview in timeline and an empty thread panel steps to reproduce currently the last two threads in don t work for me outcome what did you expect threads to show a preview in the timeline and the messages in the thread in the threads panel what happened instead they render a blank preview in the timeline and the threads panel stays blank too when i click one of them links to messages operating system macos application version element desktop nightly how did you install the app homeserver element io will you send logs yes
| 1
|
7,682
| 2,610,432,822
|
IssuesEvent
|
2015-02-26 20:21:48
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
lost blog post!
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
completed a posting and sent to blog - disappeared!! Using Blogger
What browser are you using?
Chrome
What version of ScribeFire are you running?
```
-----
Original issue reported on code.google.com by `paul.mcg...@ocsb.ca` on 26 Jan 2014 at 2:09
|
1.0
|
lost blog post! - ```
What's the problem?
completed a posting and sent to blog - disappeared!! Using Blogger
What browser are you using?
Chrome
What version of ScribeFire are you running?
```
-----
Original issue reported on code.google.com by `paul.mcg...@ocsb.ca` on 26 Jan 2014 at 2:09
|
defect
|
lost blog post what s the problem completed a posting and sent to blog disappeared using blogger what browser are you using chrome what version of scribefire are you running original issue reported on code google com by paul mcg ocsb ca on jan at
| 1
|
271,047
| 23,576,813,854
|
IssuesEvent
|
2022-08-23 02:12:12
|
lowRISC/opentitan
|
https://api.github.com/repos/lowRISC/opentitan
|
opened
|
[rom-e2e] rom_functests
|
Priority:P2 Type:Task SW:ROM Milestone:V2 Component:RomE2eTest
|
### Test point name
[rom_functests](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_testplan.hjson?q=rom_functests)
### Host side component
Unknown
### OpenTitanTool infrastructure implemented
Unknown
### Contact person
@alphan
### Checklist
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
### Determine which functests can be executed using ROM.
Functests test ROM components (e.g., drivers, libraries, etc.) work as intended on-chip.
However, unlike when these components are embedded in the ROM, functests are linked with
the OTTF, and run out of flash. Additionally, unlike the ROM E2E tests, functests are
booted by the test ROM.
Determine which functests can be executed using ROM to understand which tests can be
reused on the silicon.
|
1.0
|
[rom-e2e] rom_functests - ### Test point name
[rom_functests](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_testplan.hjson?q=rom_functests)
### Host side component
Unknown
### OpenTitanTool infrastructure implemented
Unknown
### Contact person
@alphan
### Checklist
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
### Determine which functests can be executed using ROM.
Functests test ROM components (e.g., drivers, libraries, etc.) work as intended on-chip.
However, unlike when these components are embedded in the ROM, functests are linked with
the OTTF, and run out of flash. Additionally, unlike the ROM E2E tests, functests are
booted by the test ROM.
Determine which functests can be executed using ROM to understand which tests can be
reused on the silicon.
|
non_defect
|
rom functests test point name host side component unknown opentitantool infrastructure implemented unknown contact person alphan checklist please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking determine which functests can be executed using rom functests test rom components e g drivers libraries etc work as intended on chip however unlike when these components are embedded in the rom functests are linked with the ottf and run out of flash additionally unlike the rom tests functests are booted by the test rom determine which functests can be executed using rom to understand which tests can be reused on the silicon
| 0
|
189,960
| 22,047,164,829
|
IssuesEvent
|
2022-05-30 04:01:38
|
pazhanivel07/linux-4.19.72
|
https://api.github.com/repos/pazhanivel07/linux-4.19.72
|
closed
|
CVE-2020-0427 (Medium) detected in linux-yoctov5.4.51 - autoclosed
|
security vulnerability
|
## CVE-2020-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pinctrl/devicetree.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pinctrl/devicetree.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In create_pinctrl of core.c, there is a possible out of bounds read due to a use after free. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android kernelAndroid ID: A-140550171
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0427>CVE-2020-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-0427">https://www.linuxkernelcves.com/cves/CVE-2020-0427</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution: v4.14.161,v4.19.92,v5.4.7,v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-0427 (Medium) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2020-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pinctrl/devicetree.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pinctrl/devicetree.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In create_pinctrl of core.c, there is a possible out of bounds read due to a use after free. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android kernelAndroid ID: A-140550171
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0427>CVE-2020-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-0427">https://www.linuxkernelcves.com/cves/CVE-2020-0427</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution: v4.14.161,v4.19.92,v5.4.7,v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux autoclosed cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers pinctrl devicetree c drivers pinctrl devicetree c vulnerability details in create pinctrl of core c there is a possible out of bounds read due to a use after free this could lead to local information disclosure with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android kernelandroid id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.