Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
8,121
| 6,417,329,906
|
IssuesEvent
|
2017-08-08 16:30:32
|
Elgg/Elgg
|
https://api.github.com/repos/Elgg/Elgg
|
closed
|
Memcache: cache more metadata
|
performance
|
Now that these are access-free, they can be more freely cached, and this can probably all be done in \Elgg\Cache\MetadataCache.
|
True
|
Memcache: cache more metadata - Now that these are access-free, they can be more freely cached, and this can probably all be done in \Elgg\Cache\MetadataCache.
|
non_test
|
memcache cache more metadata now that these are access free they can be more freely cached and this can probably all be done in elgg cache metadatacache
| 0
|
127,285
| 10,459,266,628
|
IssuesEvent
|
2019-09-20 10:29:41
|
blazzer-band/Warriors-vs-Creeps
|
https://api.github.com/repos/blazzer-band/Warriors-vs-Creeps
|
closed
|
Write tests for the core game
|
core tests
|
Write tests for the core game #3
- [ ] random seed test
- [ ] map init test
|
1.0
|
Write tests for the core game - Write tests for the core game #3
- [ ] random seed test
- [ ] map init test
|
test
|
write tests for the core game write tests for the core game random seed test map init test
| 1
|
314,616
| 27,013,848,222
|
IssuesEvent
|
2023-02-10 17:30:19
|
acikkaynak/deprem-yardim-frontend
|
https://api.github.com/repos/acikkaynak/deprem-yardim-frontend
|
closed
|
bug: Layer filter icerisinde her cluster icin ayri ayri "noktalar" checkbox'i gorunuyor
|
bug approved test-failed
|
## Hata Tanımı
Goruldugu kadariyla ekrandaki her cluster icin bir "noktalar" checkbox'i ekliyoruz. Hepsinin gruplanmasi gerekiyor.
-- bu bilgiyi ekleyiniz
** discord kullanıcı adı: @Caner Dagli#1251 **
## Hata Ortamı
rc.afetharita.com
afetharita.com
## Hatayı nasıl aldığınızı adım adım açıklayın
1. Sol taraftaki layer filtere basin
4. Hata görünüyor
## Beklenen Davranış
Sadece 1 tane "Noktalar" secenegi olmali
## Ekran Görüntüleri
<img width="501" alt="image" src="https://user-images.githubusercontent.com/7020369/218091775-ccd8f9f9-b47f-4323-8dd6-baa474e130ff.png">
## Masaüstü Bilgileri
- İşletim Sistemi: OSX
- Tarayıcı [örneğin chrome, safari]
- Sürüm [örneğin 22]
## Akıllı Telefon Bilgileri
- Cihaz: [örneğin iPhone6]
- İşletim Sistemi: [örneğin iOS8.1]
- Tarayıcı [örneğin varsayılan tarayıcı, safari]
- Sürüm [örneğin 22]
## Ek Bağlam
Sorun hakkında diğer herhangi bir bağlamı burada ekleyin.
|
1.0
|
bug: Layer filter icerisinde her cluster icin ayri ayri "noktalar" checkbox'i gorunuyor - ## Hata Tanımı
Goruldugu kadariyla ekrandaki her cluster icin bir "noktalar" checkbox'i ekliyoruz. Hepsinin gruplanmasi gerekiyor.
-- bu bilgiyi ekleyiniz
** discord kullanıcı adı: @Caner Dagli#1251 **
## Hata Ortamı
rc.afetharita.com
afetharita.com
## Hatayı nasıl aldığınızı adım adım açıklayın
1. Sol taraftaki layer filtere basin
4. Hata görünüyor
## Beklenen Davranış
Sadece 1 tane "Noktalar" secenegi olmali
## Ekran Görüntüleri
<img width="501" alt="image" src="https://user-images.githubusercontent.com/7020369/218091775-ccd8f9f9-b47f-4323-8dd6-baa474e130ff.png">
## Masaüstü Bilgileri
- İşletim Sistemi: OSX
- Tarayıcı [örneğin chrome, safari]
- Sürüm [örneğin 22]
## Akıllı Telefon Bilgileri
- Cihaz: [örneğin iPhone6]
- İşletim Sistemi: [örneğin iOS8.1]
- Tarayıcı [örneğin varsayılan tarayıcı, safari]
- Sürüm [örneğin 22]
## Ek Bağlam
Sorun hakkında diğer herhangi bir bağlamı burada ekleyin.
|
test
|
bug layer filter icerisinde her cluster icin ayri ayri noktalar checkbox i gorunuyor hata tanımı goruldugu kadariyla ekrandaki her cluster icin bir noktalar checkbox i ekliyoruz hepsinin gruplanmasi gerekiyor bu bilgiyi ekleyiniz discord kullanıcı adı caner dagli hata ortamı rc afetharita com afetharita com hatayı nasıl aldığınızı adım adım açıklayın sol taraftaki layer filtere basin hata görünüyor beklenen davranış sadece tane noktalar secenegi olmali ekran görüntüleri img width alt image src masaüstü bilgileri i̇şletim sistemi osx tarayıcı sürüm akıllı telefon bilgileri cihaz i̇şletim sistemi tarayıcı sürüm ek bağlam sorun hakkında diğer herhangi bir bağlamı burada ekleyin
| 1
|
39,201
| 5,222,846,296
|
IssuesEvent
|
2017-01-27 10:01:06
|
emfoundation/ce100-app
|
https://api.github.com/repos/emfoundation/ce100-app
|
closed
|
Member profile details management
|
please-test primary-user priority-2 user-story
|
As a primary user, I can add and edit a company logo, a mission statement and a set of tags characterising my knowledge/experience, so that I have control over how I represent my organisation on the platform and can update this over time.
Mission statement: limited number of chars?
Parent tags are only used as navigational tool and cannot be selected, nor used to select all corresponding child tags.
|
1.0
|
Member profile details management - As a primary user, I can add and edit a company logo, a mission statement and a set of tags characterising my knowledge/experience, so that I have control over how I represent my organisation on the platform and can update this over time.
Mission statement: limited number of chars?
Parent tags are only used as navigational tool and cannot be selected, nor used to select all corresponding child tags.
|
test
|
member profile details management as a primary user i can add and edit a company logo a mission statement and a set of tags characterising my knowledge experience so that i have control over how i represent my organisation on the platform and can update this over time mission statement limited number of chars parent tags are only used as navigational tool and cannot be selected nor used to select all corresponding child tags
| 1
|
361,129
| 25,327,072,625
|
IssuesEvent
|
2022-11-18 10:09:24
|
GrainLearning/grainLearning
|
https://api.github.com/repos/GrainLearning/grainLearning
|
opened
|
Installation with IDEs
|
documentation
|
We see in the TUSAIL school quite a few students use IDEs like visual studio and Pycharm.
We may want to include instructions on how to install within these IDE.
|
1.0
|
Installation with IDEs - We see in the TUSAIL school quite a few students use IDEs like visual studio and Pycharm.
We may want to include instructions on how to install within these IDE.
|
non_test
|
installation with ides we see in the tusail school quite a few students use ides like visual studio and pycharm we may want to include instructions on how to install within these ide
| 0
|
250,520
| 21,314,031,199
|
IssuesEvent
|
2022-04-16 01:50:55
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Upgrade from Chromium 100 to Chromium 101.
|
QA/Yes release-notes/include QA/Test-Plan-Specified OS/Android Chromium/upgrade major OS/Desktop
|
Upgrade from Chromium 100 to Chromium 101.
https://chromium.googlesource.com/chromium/src/+log/100.0.4896.20..101.0.4919.0/?pretty=fuller&n=10000
QA tests:
This is a major Chromium version bump, please do full passes.
**Desktop Affected areas:**
* Sync (general operation, device removing from sync chain)
* Search engines show up correctly in the settings (fresh profile and on upgrade)
* UI shows data correctly for:
- Shields
- Ad block
- Rewards page
- Rewards internals
- Tipping
- Welcome
- IPFS
- NTP
- Tor internals
- Webcompat reporter
* Shields function correctly when changing shields settings
* Settings UI: toggling these settings works as expected:
- Extensions > WebTorrent
- Extensions > Widevine
- Privacy and security > Allow privacy-preserving product analytics (P3A)
- Privacy and security > Automatically send daily usage ping to Brave
**Android Affected areas:**
* Play video in browser
* Top toolbar UI
* New tab page UI
* Partner customizations (double check that they are ignored, it applies applies only to specific vendors like Lenovo)
|
1.0
|
Upgrade from Chromium 100 to Chromium 101. - Upgrade from Chromium 100 to Chromium 101.
https://chromium.googlesource.com/chromium/src/+log/100.0.4896.20..101.0.4919.0/?pretty=fuller&n=10000
QA tests:
This is a major Chromium version bump, please do full passes.
**Desktop Affected areas:**
* Sync (general operation, device removing from sync chain)
* Search engines show up correctly in the settings (fresh profile and on upgrade)
* UI shows data correctly for:
- Shields
- Ad block
- Rewards page
- Rewards internals
- Tipping
- Welcome
- IPFS
- NTP
- Tor internals
- Webcompat reporter
* Shields function correctly when changing shields settings
* Settings UI: toggling these settings works as expected:
- Extensions > WebTorrent
- Extensions > Widevine
- Privacy and security > Allow privacy-preserving product analytics (P3A)
- Privacy and security > Automatically send daily usage ping to Brave
**Android Affected areas:**
* Play video in browser
* Top toolbar UI
* New tab page UI
* Partner customizations (double check that they are ignored, it applies applies only to specific vendors like Lenovo)
|
test
|
upgrade from chromium to chromium upgrade from chromium to chromium qa tests this is a major chromium version bump please do full passes desktop affected areas sync general operation device removing from sync chain search engines show up correctly in the settings fresh profile and on upgrade ui shows data correctly for shields ad block rewards page rewards internals tipping welcome ipfs ntp tor internals webcompat reporter shields function correctly when changing shields settings settings ui toggling these settings works as expected extensions webtorrent extensions widevine privacy and security allow privacy preserving product analytics privacy and security automatically send daily usage ping to brave android affected areas play video in browser top toolbar ui new tab page ui partner customizations double check that they are ignored it applies applies only to specific vendors like lenovo
| 1
|
50,540
| 6,101,210,710
|
IssuesEvent
|
2017-06-20 14:12:40
|
pixelhumain/co2
|
https://api.github.com/repos/pixelhumain/co2
|
closed
|
Activité territoriale - vieille data
|
to test
|
En tant qu'utilisateur, si je vais sur mon fil d'info, le colonne de droite affichant l'activité territoriale de la réunion me propose seulement des vieilles datas.
Alors que j'ai créé plusieurs événements et d'autres orga sur ce territoire.



|
1.0
|
Activité territoriale - vieille data - En tant qu'utilisateur, si je vais sur mon fil d'info, le colonne de droite affichant l'activité territoriale de la réunion me propose seulement des vieilles datas.
Alors que j'ai créé plusieurs événements et d'autres orga sur ce territoire.



|
test
|
activité territoriale vieille data en tant qu utilisateur si je vais sur mon fil d info le colonne de droite affichant l activité territoriale de la réunion me propose seulement des vieilles datas alors que j ai créé plusieurs événements et d autres orga sur ce territoire
| 1
|
158,707
| 12,423,128,304
|
IssuesEvent
|
2020-05-24 03:15:30
|
kubernetes-csi/csi-driver-smb
|
https://api.github.com/repos/kubernetes-csi/csi-driver-smb
|
closed
|
Enable golint check in travis
|
help wanted test
|
**Is your feature request related to a problem?/Why is this needed**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
**Describe the solution you'd like in detail**
<!-- A clear and concise description of what you want to happen. -->
Enable golint check in travis failed:
https://github.com/csi-driver/csi-driver-smb/pull/5
**Describe alternatives you've considered**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
|
1.0
|
Enable golint check in travis - **Is your feature request related to a problem?/Why is this needed**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
**Describe the solution you'd like in detail**
<!-- A clear and concise description of what you want to happen. -->
Enable golint check in travis failed:
https://github.com/csi-driver/csi-driver-smb/pull/5
**Describe alternatives you've considered**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
|
test
|
enable golint check in travis is your feature request related to a problem why is this needed describe the solution you d like in detail enable golint check in travis failed describe alternatives you ve considered additional context
| 1
|
269,922
| 23,476,832,026
|
IssuesEvent
|
2022-08-17 07:00:13
|
gardener/gardener
|
https://api.github.com/repos/gardener/gardener
|
closed
|
[Flaky Test] Bastion Controller Integration Test Suite
|
area/testing kind/flake
|
<!-- Please only use this template for submitting reports about flaky tests or jobs (pass or fail with no underlying change in code) in Gardener CI -->
**How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
-->
/area testing
/kind flake
**Which test(s)/suite(s) are flaking**:
- `Bastion Controller Integration Test Suite.[It] Bastion controller tests shoot exists, is not in deletion and on the same seed bastion is not yet expired should not delete Bastion`
- `Bastion Controller Integration Test Suite.[It] Bastion controller tests shoot exists, is not in deletion and on the same seed expiration timestamp should requeue and delete Bastion if its expiration timestamp is about to pass`
**CI link**:
For the first test:
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558880535189655552
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558397346193084416
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558699339210035200
For the second test:
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1556645000039632896
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1553382568210993152
Also see: https://testgrid.k8s.io/gardener-gardener#ci-gardener-integration&sort-by-flakiness=
**Reason for failure**:
In both cases, it seems like the bastion is cleaned up although it is not supposed to be deleted.
**Anything else we need to know**:
|
1.0
|
[Flaky Test] Bastion Controller Integration Test Suite - <!-- Please only use this template for submitting reports about flaky tests or jobs (pass or fail with no underlying change in code) in Gardener CI -->
**How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
-->
/area testing
/kind flake
**Which test(s)/suite(s) are flaking**:
- `Bastion Controller Integration Test Suite.[It] Bastion controller tests shoot exists, is not in deletion and on the same seed bastion is not yet expired should not delete Bastion`
- `Bastion Controller Integration Test Suite.[It] Bastion controller tests shoot exists, is not in deletion and on the same seed expiration timestamp should requeue and delete Bastion if its expiration timestamp is about to pass`
**CI link**:
For the first test:
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558880535189655552
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558397346193084416
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1558699339210035200
For the second test:
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1556645000039632896
https://prow.gardener.cloud/view/gs/gardener-prow/logs/ci-gardener-integration/1553382568210993152
Also see: https://testgrid.k8s.io/gardener-gardener#ci-gardener-integration&sort-by-flakiness=
**Reason for failure**:
In both cases, it seems like the bastion is cleaned up although it is not supposed to be deleted.
**Anything else we need to know**:
|
test
|
bastion controller integration test suite how to categorize this issue please select area kind and priority for this issue this helps the community categorizing it replace below todos or exchange the existing identifiers with those that fit best in your opinion if multiple identifiers make sense you can also state the commands multiple times e g area control plane area auto scaling area identifiers audit logging auto scaling backup certification control plane migration control plane cost delivery dev productivity disaster recovery documentation high availability logging metering monitoring networking open source ops productivity os performance quality robustness scalability security storage testing usability user management kind identifiers api change bug cleanup discussion enhancement epic impediment poc post mortem question regression task technical debt test area testing kind flake which test s suite s are flaking bastion controller integration test suite bastion controller tests shoot exists is not in deletion and on the same seed bastion is not yet expired should not delete bastion bastion controller integration test suite bastion controller tests shoot exists is not in deletion and on the same seed expiration timestamp should requeue and delete bastion if its expiration timestamp is about to pass ci link for the first test for the second test also see reason for failure in both cases it seems like the bastion is cleaned up although it is not supposed to be deleted anything else we need to know
| 1
|
1,143
| 2,698,792,789
|
IssuesEvent
|
2015-04-03 11:07:20
|
neovim/neovim
|
https://api.github.com/repos/neovim/neovim
|
closed
|
travis: test on OSX too
|
buildsystem
|
Travis is getting awesomer and awesomer. We once again see the possibility of augmenting our test capability: travis seems to support OSX now:
**NOTE:** I haven't been able to find if it's now possible to specify all that is necessary in one `.travis.yml` file. This used to be impossible in earlier betas. Information seems scarce.
- https://github.com/travis-ci/travis-ci/issues/216 (people usually link to this when they have a PR for their project, so we can look at examples there)
- http://docs.travis-ci.com/user/osx-ci-environment/
- https://github.com/citra-emu/citra/pull/7 (this seems like a good example to follow)
It seems one needs the `os` key:
```yaml
os:
- osx
- linux
```
I don't see a need of building all our current lines on OSX though, just one should be enough. Ideally there would also come a FreeBSD build with time. Travis [doesn't support FreeBSD at the moment, though](https://github.com/travis-ci/travis-ci/issues/1818).
<bountysource-plugin>
---
Want to back this issue? **[Place a bounty on it!](https://www.bountysource.com/issues/2345752-travis-test-on-osx-too?utm_campaign=plugin&utm_content=tracker%2F461131&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F461131&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
travis: test on OSX too - Travis is getting awesomer and awesomer. We once again see the possibility of augmenting our test capability: travis seems to support OSX now:
**NOTE:** I haven't been able to find if it's now possible to specify all that is necessary in one `.travis.yml` file. This used to be impossible in earlier betas. Information seems scarce.
- https://github.com/travis-ci/travis-ci/issues/216 (people usually link to this when they have a PR for their project, so we can look at examples there)
- http://docs.travis-ci.com/user/osx-ci-environment/
- https://github.com/citra-emu/citra/pull/7 (this seems like a good example to follow)
It seems one needs the `os` key:
```yaml
os:
- osx
- linux
```
I don't see a need of building all our current lines on OSX though, just one should be enough. Ideally there would also come a FreeBSD build with time. Travis [doesn't support FreeBSD at the moment, though](https://github.com/travis-ci/travis-ci/issues/1818).
<bountysource-plugin>
---
Want to back this issue? **[Place a bounty on it!](https://www.bountysource.com/issues/2345752-travis-test-on-osx-too?utm_campaign=plugin&utm_content=tracker%2F461131&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F461131&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_test
|
travis test on osx too travis is getting awesomer and awesomer we once again see the possibility of augmenting our test capability travis seems to support osx now note i haven t been able to find if it s now possible to specify all that is necessary in one travis yml file this used to be impossible in earlier betas information seems scarce people usually link to this when they have a pr for their project so we can look at examples there this seems like a good example to follow it seems one needs the os key yaml os osx linux i don t see a need of building all our current lines on osx though just one should be enough ideally there would also come a freebsd build with time travis want to back this issue we accept bounties via
| 0
|
44,349
| 13,055,259,430
|
IssuesEvent
|
2020-07-30 01:07:25
|
kenferrara/react-pwa-reference-storefront
|
https://api.github.com/repos/kenferrara/react-pwa-reference-storefront
|
opened
|
CVE-2020-8175 (Medium) detected in jpeg-js-0.3.7.tgz, jpeg-js-0.3.6.tgz
|
security vulnerability
|
## CVE-2020-8175 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jpeg-js-0.3.7.tgz</b>, <b>jpeg-js-0.3.6.tgz</b></p></summary>
<p>
<details><summary><b>jpeg-js-0.3.7.tgz</b></p></summary>
<p>A pure javascript JPEG encoder and decoder</p>
<p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.7.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.7.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/react-pwa-reference-storefront/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/react-pwa-reference-storefront/node_modules/jpeg-js/package.json</p>
<p>
Dependency Hierarchy:
- quagga-0.12.1.tgz (Root Library)
- get-pixels-3.3.2.tgz
- :x: **jpeg-js-0.3.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>jpeg-js-0.3.6.tgz</b></p></summary>
<p>A pure javascript JPEG encoder and decoder</p>
<p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.6.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.6.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/react-pwa-reference-storefront/src/components/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/react-pwa-reference-storefront/src/components/node_modules/jpeg-js/package.json</p>
<p>
Dependency Hierarchy:
- quagga-0.12.1.tgz (Root Library)
- get-pixels-3.3.2.tgz
- :x: **jpeg-js-0.3.6.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Uncontrolled resource consumption in `jpeg-js` before 0.4.0 may allow attacker to launch denial of service attacks using specially a crafted JPEG image.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8175>CVE-2020-8175</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8175">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8175</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 0.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jpeg-js","packageVersion":"0.3.7","isTransitiveDependency":true,"dependencyTree":"quagga:0.12.1;get-pixels:3.3.2;jpeg-js:0.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.4.0"},{"packageType":"javascript/Node.js","packageName":"jpeg-js","packageVersion":"0.3.6","isTransitiveDependency":true,"dependencyTree":"quagga:0.12.1;get-pixels:3.3.2;jpeg-js:0.3.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.4.0"}],"vulnerabilityIdentifier":"CVE-2020-8175","vulnerabilityDetails":"Uncontrolled resource consumption in `jpeg-js` before 0.4.0 may allow attacker to launch denial of service attacks using specially a crafted JPEG image.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8175","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-8175 (Medium) detected in jpeg-js-0.3.7.tgz, jpeg-js-0.3.6.tgz - ## CVE-2020-8175 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jpeg-js-0.3.7.tgz</b>, <b>jpeg-js-0.3.6.tgz</b></p></summary>
<p>
<details><summary><b>jpeg-js-0.3.7.tgz</b></p></summary>
<p>A pure javascript JPEG encoder and decoder</p>
<p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.7.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.7.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/react-pwa-reference-storefront/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/react-pwa-reference-storefront/node_modules/jpeg-js/package.json</p>
<p>
Dependency Hierarchy:
- quagga-0.12.1.tgz (Root Library)
- get-pixels-3.3.2.tgz
- :x: **jpeg-js-0.3.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>jpeg-js-0.3.6.tgz</b></p></summary>
<p>A pure javascript JPEG encoder and decoder</p>
<p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.6.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.6.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/react-pwa-reference-storefront/src/components/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/react-pwa-reference-storefront/src/components/node_modules/jpeg-js/package.json</p>
<p>
Dependency Hierarchy:
- quagga-0.12.1.tgz (Root Library)
- get-pixels-3.3.2.tgz
- :x: **jpeg-js-0.3.6.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Uncontrolled resource consumption in `jpeg-js` before 0.4.0 may allow attacker to launch denial of service attacks using specially a crafted JPEG image.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8175>CVE-2020-8175</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8175">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8175</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 0.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jpeg-js","packageVersion":"0.3.7","isTransitiveDependency":true,"dependencyTree":"quagga:0.12.1;get-pixels:3.3.2;jpeg-js:0.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.4.0"},{"packageType":"javascript/Node.js","packageName":"jpeg-js","packageVersion":"0.3.6","isTransitiveDependency":true,"dependencyTree":"quagga:0.12.1;get-pixels:3.3.2;jpeg-js:0.3.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.4.0"}],"vulnerabilityIdentifier":"CVE-2020-8175","vulnerabilityDetails":"Uncontrolled resource consumption in `jpeg-js` before 0.4.0 may allow attacker to launch denial of service attacks using specially a crafted JPEG image.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8175","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in jpeg js tgz jpeg js tgz cve medium severity vulnerability vulnerable libraries jpeg js tgz jpeg js tgz jpeg js tgz a pure javascript jpeg encoder and decoder library home page a href path to dependency file tmp ws scm react pwa reference storefront package json path to vulnerable library tmp ws scm react pwa reference storefront node modules jpeg js package json dependency hierarchy quagga tgz root library get pixels tgz x jpeg js tgz vulnerable library jpeg js tgz a pure javascript jpeg encoder and decoder library home page a href path to dependency file tmp ws scm react pwa reference storefront src components package json path to vulnerable library tmp ws scm react pwa reference storefront src components node modules jpeg js package json dependency hierarchy quagga tgz root library get pixels tgz x jpeg js tgz vulnerable library vulnerability details uncontrolled resource consumption in jpeg js before may allow attacker to launch denial of service attacks using specially a crafted jpeg image publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails uncontrolled resource consumption in jpeg js before may allow attacker to launch denial of service attacks using specially a crafted jpeg image vulnerabilityurl
| 0
|
50,526
| 6,100,509,111
|
IssuesEvent
|
2017-06-20 12:46:50
|
hazelcast/hazelcast-cpp-client
|
https://api.github.com/repos/hazelcast/hazelcast-cpp-client
|
closed
|
ClusterTest.testListenersWhenClusterDown failed at nightly tests cpp-centos7-nightly-32-STATIC-Release
|
test-failure
|
Test ClusterTest.testListenersWhenClusterDown failed at nightly tests cpp-centos7-nightly-32-STATIC-Release for master branch at e5548cc2785a7c551925e475f8e56f4abb1de058 (failure started after https://github.com/hazelcast/hazelcast-cpp-client/pull/246 )
The test caused segmentation fault.
The core stack trace is:
04:51:22 Using host libthread_db library "/lib64/libthread_db.so.1".
04:51:22 Core was generated by `buildSTATIC32Release/hazelcast/test/src/clientTest_STATIC_32 --gtest_output=xml'.
04:51:22 Program terminated with signal 11, Segmentation fault.
04:51:22 #0 0x086dcd5d in hazelcast::client::connection::OwnerConnectionFuture::getOrWaitForCreation() ()
04:51:22
04:51:22 Thread 3 (Thread 0xf6fa0b40 (LWP 8412)):
04:51:22 #0 0xf775d430 in __kernel_vsyscall ()
04:51:22 No symbol table info available.
04:51:22 #1 0xf72c06d1 in select () from /lib/libc.so.6
04:51:22 No symbol table info available.
04:51:22 #2 0x0877c8d6 in hazelcast::client::connection::OutSelector::listenInternal() ()
04:51:22 No symbol table info available.
04:51:22 #3 0x0877ba08 in hazelcast::client::connection::IOSelector::listen() ()
04:51:22 No symbol table info available.
04:51:22 #4 0x087421a2 in hazelcast::util::Thread::controlledThread(void*) ()
04:51:22 No symbol table info available.
04:51:22 #5 0xf7720b2c in start_thread () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #6 0xf72ca08e in clone () from /lib/libc.so.6
04:51:22 No symbol table info available.
04:51:22
04:51:22 Thread 2 (Thread 0xf6fa27c0 (LWP 4151)):
04:51:22 #0 0xf775d430 in __kernel_vsyscall ()
04:51:22 No symbol table info available.
04:51:22 #1 0xf7721d15 in pthread_join () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #2 0x087420a3 in hazelcast::util::Thread::join() ()
04:51:22 No symbol table info available.
04:51:22 #3 0x086d574f in hazelcast::client::connection::ConnectionManager::shutdown() ()
04:51:22 No symbol table info available.
04:51:22 #4 0x08726034 in hazelcast::client::spi::LifecycleService::shutdown() ()
04:51:22 No symbol table info available.
04:51:22 #5 0x08737919 in hazelcast::client::HazelcastClient::~HazelcastClient() ()
04:51:22 No symbol table info available.
04:51:22 #6 0x084ba614 in hazelcast::client::test::ClusterTest_testListenersWhenClusterDown_Test::TestBody() ()
04:51:22 No symbol table info available.
04:51:22 #7 0x087aff6c in void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) ()
04:51:22 No symbol table info available.
04:51:22 #8 0x087a4b04 in testing::Test::Run() ()
04:51:22 No symbol table info available.
04:51:22 #9 0x087a4d24 in testing::TestInfo::Run() ()
04:51:22 No symbol table info available.
04:51:22 #10 0x087a4e6f in testing::TestCase::Run() ()
04:51:22 No symbol table info available.
04:51:22 #11 0x087a7ad6 in testing::internal::UnitTestImpl::RunAllTests() [clone .part.445] ()
04:51:22 No symbol table info available.
04:51:22 #12 0x087a7e9e in testing::UnitTest::Run() ()
04:51:22 No symbol table info available.
04:51:22 #13 0x083992a8 in main ()
04:51:22 No symbol table info available.
04:51:22
04:51:22 Thread 1 (Thread 0xf679fb40 (LWP 8411)):
04:51:22 #0 0x086dcd5d in hazelcast::client::connection::OwnerConnectionFuture::getOrWaitForCreation() ()
04:51:22 No symbol table info available.
04:51:22 #1 0x086d7d74 in hazelcast::client::connection::ConnectionManager::getOrConnectResolved(hazelcast::client::Address const&) ()
04:51:22 No symbol table info available.
04:51:22 #2 0x086d82c3 in hazelcast::client::connection::ConnectionManager::getOrConnect(hazelcast::client::Address const&) ()
04:51:22 No symbol table info available.
04:51:22 #3 0x086d84f3 in hazelcast::client::connection::ConnectionManager::getOrConnect(hazelcast::client::Address const&, int) ()
04:51:22 No symbol table info available.
04:51:22 #4 0x086d89ef in hazelcast::client::connection::ConnectionManager::getRandomConnection(int, std::string const&, int) ()
04:51:22 No symbol table info available.
04:51:22 #5 0x08722a09 in hazelcast::client::spi::InvocationService::resend(boost::shared_ptr<hazelcast::client::connection::CallPromise>, std::string const&) ()
04:51:22 No symbol table info available.
04:51:22 #6 0x08728dc6 in hazelcast::client::spi::ServerListenerService::retryFailedListener(boost::shared_ptr<hazelcast::client::connection::CallPromise>) ()
04:51:22 No symbol table info available.
04:51:22 #7 0x08721105 in hazelcast::client::spi::InvocationService::cleanEventHandlers(hazelcast::client::connection::Connection&) ()
04:51:22 No symbol table info available.
04:51:22 #8 0x08723310 in hazelcast::client::spi::InvocationService::cleanResources(hazelcast::client::connection::Connection&) ()
04:51:22 No symbol table info available.
04:51:22 #9 0x086d31fe in hazelcast::client::connection::Connection::close() ()
04:51:22 No symbol table info available.
04:51:22 #10 0x08776e1d in hazelcast::client::connection::IOHandler::handleSocketException(std::string const&) ()
04:51:22 No symbol table info available.
04:51:22 #11 0x08776fbd in hazelcast::client::connection::ReadHandler::handle() ()
04:51:22 No symbol table info available.
04:51:22 #12 0x0877c69e in hazelcast::client::connection::InSelector::listenInternal() ()
04:51:22 No symbol table info available.
04:51:22 #13 0x0877ba08 in hazelcast::client::connection::IOSelector::listen() ()
04:51:22 No symbol table info available.
04:51:22 #14 0x087421a2 in hazelcast::util::Thread::controlledThread(void*) ()
04:51:22 No symbol table info available.
04:51:22 #15 0xf7720b2c in start_thread () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #16 0xf72ca08e in clone () from /lib/libc.so.6
|
1.0
|
ClusterTest.testListenersWhenClusterDown failed at nightly tests cpp-centos7-nightly-32-STATIC-Release - Test ClusterTest.testListenersWhenClusterDown failed at nightly tests cpp-centos7-nightly-32-STATIC-Release for master branch at e5548cc2785a7c551925e475f8e56f4abb1de058 (failure started after https://github.com/hazelcast/hazelcast-cpp-client/pull/246 )
The test caused segmentation fault.
The core stack trace is:
04:51:22 Using host libthread_db library "/lib64/libthread_db.so.1".
04:51:22 Core was generated by `buildSTATIC32Release/hazelcast/test/src/clientTest_STATIC_32 --gtest_output=xml'.
04:51:22 Program terminated with signal 11, Segmentation fault.
04:51:22 #0 0x086dcd5d in hazelcast::client::connection::OwnerConnectionFuture::getOrWaitForCreation() ()
04:51:22
04:51:22 Thread 3 (Thread 0xf6fa0b40 (LWP 8412)):
04:51:22 #0 0xf775d430 in __kernel_vsyscall ()
04:51:22 No symbol table info available.
04:51:22 #1 0xf72c06d1 in select () from /lib/libc.so.6
04:51:22 No symbol table info available.
04:51:22 #2 0x0877c8d6 in hazelcast::client::connection::OutSelector::listenInternal() ()
04:51:22 No symbol table info available.
04:51:22 #3 0x0877ba08 in hazelcast::client::connection::IOSelector::listen() ()
04:51:22 No symbol table info available.
04:51:22 #4 0x087421a2 in hazelcast::util::Thread::controlledThread(void*) ()
04:51:22 No symbol table info available.
04:51:22 #5 0xf7720b2c in start_thread () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #6 0xf72ca08e in clone () from /lib/libc.so.6
04:51:22 No symbol table info available.
04:51:22
04:51:22 Thread 2 (Thread 0xf6fa27c0 (LWP 4151)):
04:51:22 #0 0xf775d430 in __kernel_vsyscall ()
04:51:22 No symbol table info available.
04:51:22 #1 0xf7721d15 in pthread_join () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #2 0x087420a3 in hazelcast::util::Thread::join() ()
04:51:22 No symbol table info available.
04:51:22 #3 0x086d574f in hazelcast::client::connection::ConnectionManager::shutdown() ()
04:51:22 No symbol table info available.
04:51:22 #4 0x08726034 in hazelcast::client::spi::LifecycleService::shutdown() ()
04:51:22 No symbol table info available.
04:51:22 #5 0x08737919 in hazelcast::client::HazelcastClient::~HazelcastClient() ()
04:51:22 No symbol table info available.
04:51:22 #6 0x084ba614 in hazelcast::client::test::ClusterTest_testListenersWhenClusterDown_Test::TestBody() ()
04:51:22 No symbol table info available.
04:51:22 #7 0x087aff6c in void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) ()
04:51:22 No symbol table info available.
04:51:22 #8 0x087a4b04 in testing::Test::Run() ()
04:51:22 No symbol table info available.
04:51:22 #9 0x087a4d24 in testing::TestInfo::Run() ()
04:51:22 No symbol table info available.
04:51:22 #10 0x087a4e6f in testing::TestCase::Run() ()
04:51:22 No symbol table info available.
04:51:22 #11 0x087a7ad6 in testing::internal::UnitTestImpl::RunAllTests() [clone .part.445] ()
04:51:22 No symbol table info available.
04:51:22 #12 0x087a7e9e in testing::UnitTest::Run() ()
04:51:22 No symbol table info available.
04:51:22 #13 0x083992a8 in main ()
04:51:22 No symbol table info available.
04:51:22
04:51:22 Thread 1 (Thread 0xf679fb40 (LWP 8411)):
04:51:22 #0 0x086dcd5d in hazelcast::client::connection::OwnerConnectionFuture::getOrWaitForCreation() ()
04:51:22 No symbol table info available.
04:51:22 #1 0x086d7d74 in hazelcast::client::connection::ConnectionManager::getOrConnectResolved(hazelcast::client::Address const&) ()
04:51:22 No symbol table info available.
04:51:22 #2 0x086d82c3 in hazelcast::client::connection::ConnectionManager::getOrConnect(hazelcast::client::Address const&) ()
04:51:22 No symbol table info available.
04:51:22 #3 0x086d84f3 in hazelcast::client::connection::ConnectionManager::getOrConnect(hazelcast::client::Address const&, int) ()
04:51:22 No symbol table info available.
04:51:22 #4 0x086d89ef in hazelcast::client::connection::ConnectionManager::getRandomConnection(int, std::string const&, int) ()
04:51:22 No symbol table info available.
04:51:22 #5 0x08722a09 in hazelcast::client::spi::InvocationService::resend(boost::shared_ptr<hazelcast::client::connection::CallPromise>, std::string const&) ()
04:51:22 No symbol table info available.
04:51:22 #6 0x08728dc6 in hazelcast::client::spi::ServerListenerService::retryFailedListener(boost::shared_ptr<hazelcast::client::connection::CallPromise>) ()
04:51:22 No symbol table info available.
04:51:22 #7 0x08721105 in hazelcast::client::spi::InvocationService::cleanEventHandlers(hazelcast::client::connection::Connection&) ()
04:51:22 No symbol table info available.
04:51:22 #8 0x08723310 in hazelcast::client::spi::InvocationService::cleanResources(hazelcast::client::connection::Connection&) ()
04:51:22 No symbol table info available.
04:51:22 #9 0x086d31fe in hazelcast::client::connection::Connection::close() ()
04:51:22 No symbol table info available.
04:51:22 #10 0x08776e1d in hazelcast::client::connection::IOHandler::handleSocketException(std::string const&) ()
04:51:22 No symbol table info available.
04:51:22 #11 0x08776fbd in hazelcast::client::connection::ReadHandler::handle() ()
04:51:22 No symbol table info available.
04:51:22 #12 0x0877c69e in hazelcast::client::connection::InSelector::listenInternal() ()
04:51:22 No symbol table info available.
04:51:22 #13 0x0877ba08 in hazelcast::client::connection::IOSelector::listen() ()
04:51:22 No symbol table info available.
04:51:22 #14 0x087421a2 in hazelcast::util::Thread::controlledThread(void*) ()
04:51:22 No symbol table info available.
04:51:22 #15 0xf7720b2c in start_thread () from /lib/libpthread.so.0
04:51:22 No symbol table info available.
04:51:22 #16 0xf72ca08e in clone () from /lib/libc.so.6
|
test
|
clustertest testlistenerswhenclusterdown failed at nightly tests cpp nightly static release test clustertest testlistenerswhenclusterdown failed at nightly tests cpp nightly static release for master branch at failure started after the test caused segmentation fault the core stack trace is using host libthread db library libthread db so core was generated by hazelcast test src clienttest static gtest output xml program terminated with signal segmentation fault in hazelcast client connection ownerconnectionfuture getorwaitforcreation thread thread lwp in kernel vsyscall no symbol table info available in select from lib libc so no symbol table info available in hazelcast client connection outselector listeninternal no symbol table info available in hazelcast client connection ioselector listen no symbol table info available in hazelcast util thread controlledthread void no symbol table info available in start thread from lib libpthread so no symbol table info available in clone from lib libc so no symbol table info available thread thread lwp in kernel vsyscall no symbol table info available in pthread join from lib libpthread so no symbol table info available in hazelcast util thread join no symbol table info available in hazelcast client connection connectionmanager shutdown no symbol table info available in hazelcast client spi lifecycleservice shutdown no symbol table info available in hazelcast client hazelcastclient hazelcastclient no symbol table info available in hazelcast client test clustertest testlistenerswhenclusterdown test testbody no symbol table info available in void testing internal handleexceptionsinmethodifsupported testing test void testing test char const no symbol table info available in testing test run no symbol table info available in testing testinfo run no symbol table info available in testing testcase run no symbol table info available in testing internal unittestimpl runalltests no symbol table info available in testing unittest run no symbol table info available in main no symbol table info available thread thread lwp in hazelcast client connection ownerconnectionfuture getorwaitforcreation no symbol table info available in hazelcast client connection connectionmanager getorconnectresolved hazelcast client address const no symbol table info available in hazelcast client connection connectionmanager getorconnect hazelcast client address const no symbol table info available in hazelcast client connection connectionmanager getorconnect hazelcast client address const int no symbol table info available in hazelcast client connection connectionmanager getrandomconnection int std string const int no symbol table info available in hazelcast client spi invocationservice resend boost shared ptr std string const no symbol table info available in hazelcast client spi serverlistenerservice retryfailedlistener boost shared ptr no symbol table info available in hazelcast client spi invocationservice cleaneventhandlers hazelcast client connection connection no symbol table info available in hazelcast client spi invocationservice cleanresources hazelcast client connection connection no symbol table info available in hazelcast client connection connection close no symbol table info available in hazelcast client connection iohandler handlesocketexception std string const no symbol table info available in hazelcast client connection readhandler handle no symbol table info available in hazelcast client connection inselector listeninternal no symbol table info available in hazelcast client connection ioselector listen no symbol table info available in hazelcast util thread controlledthread void no symbol table info available in start thread from lib libpthread so no symbol table info available in clone from lib libc so
| 1
|
244,023
| 20,603,516,372
|
IssuesEvent
|
2022-03-06 16:32:58
|
pvk-2022-2/mips-emulator
|
https://api.github.com/repos/pvk-2022-2/mips-emulator
|
closed
|
Test all the instructions
|
testing
|
These instructions have no tests.
R-Type:
- [x] addu
- [x] subu
- [x] and
- [x] xor
- [x] nor
- [ ] sop32
- [ ] sop33
I-Type:
- [ ] addiu
|
1.0
|
Test all the instructions - These instructions have no tests.
R-Type:
- [x] addu
- [x] subu
- [x] and
- [x] xor
- [x] nor
- [ ] sop32
- [ ] sop33
I-Type:
- [ ] addiu
|
test
|
test all the instructions these instructions have no tests r type addu subu and xor nor i type addiu
| 1
|
24,526
| 12,312,480,754
|
IssuesEvent
|
2020-05-12 13:59:44
|
yalelibrary/YUL-DC
|
https://api.github.com/repos/yalelibrary/YUL-DC
|
opened
|
Clean up public endpoints to SpinUp cluster
|
performance team
|
**ISSUE**
Since we're not going to be moving forward with SpinUp for the time being, we should decommission the public endpoints.
**ACCEPTANCE**
Remove the configuration & and any registrations for
- [ ] The blacklight application at https://dl-test.library.yale.edu/
- [ ] The manifest service at https://iiif-test.library.yale.edu/manifests
- [ ] The image service at https://iiif-test.library.yale.edu/images
|
True
|
Clean up public endpoints to SpinUp cluster - **ISSUE**
Since we're not going to be moving forward with SpinUp for the time being, we should decommission the public endpoints.
**ACCEPTANCE**
Remove the configuration & and any registrations for
- [ ] The blacklight application at https://dl-test.library.yale.edu/
- [ ] The manifest service at https://iiif-test.library.yale.edu/manifests
- [ ] The image service at https://iiif-test.library.yale.edu/images
|
non_test
|
clean up public endpoints to spinup cluster issue since we re not going to be moving forward with spinup for the time being we should decommission the public endpoints acceptance remove the configuration and any registrations for the blacklight application at the manifest service at the image service at
| 0
|
197,230
| 14,914,000,002
|
IssuesEvent
|
2021-01-22 14:52:14
|
tModLoader/tModLoader
|
https://api.github.com/repos/tModLoader/tModLoader
|
opened
|
Server port opening error handling is faulty in some cases
|
Bug Needs Further Testing
|
### Description
A few people in discord recently mentioned that their clients would get stuck at "Found server..." when using host & play. But the server logs looked fine, with the only weird thing to note being that they didn't have "X is connecting" entries. As Necro figured out on their own, this was caused by an ASUS application taking the port, in case of which. the game should've logged the following:
```
Only one usage of each socket address (protocol/network address/port) is normally permitted (see server2.log for full trace)
Tried to run two servers on the same PC
```
This message is logged properly when the port is taken by another instance of the game's server, but potentially not in cases when driver-level applications take it. The weird thing is that this happened on a mac too, log's attached below.
### Log file
Necro's (Windows) - [Windows_server.log](https://github.com/tModLoader/tModLoader/files/5856523/Windows_server.log)
mngrusin's (Mac) - [Mac_server.log](https://github.com/tModLoader/tModLoader/files/5856526/Mac_server.log)
### Steps to reproduce
Not many ideas. On windows, installing [Armoury Crate](https://rog.asus.com/us/innovation/armoury_crate/) could work, but mac would need more testing.
### Expected behavior
The message `Tried to run two servers on the same PC` would at least be printed to the server's console & log file.
### Actual behavior
The server claims that it has started listening on its port, which is impossible.
### Additional Information
This very much could be a vanilla bug.
|
1.0
|
Server port opening error handling is faulty in some cases - ### Description
A few people in discord recently mentioned that their clients would get stuck at "Found server..." when using host & play. But the server logs looked fine, with the only weird thing to note being that they didn't have "X is connecting" entries. As Necro figured out on their own, this was caused by an ASUS application taking the port, in case of which. the game should've logged the following:
```
Only one usage of each socket address (protocol/network address/port) is normally permitted (see server2.log for full trace)
Tried to run two servers on the same PC
```
This message is logged properly when the port is taken by another instance of the game's server, but potentially not in cases when driver-level applications take it. The weird thing is that this happened on a mac too, log's attached below.
### Log file
Necro's (Windows) - [Windows_server.log](https://github.com/tModLoader/tModLoader/files/5856523/Windows_server.log)
mngrusin's (Mac) - [Mac_server.log](https://github.com/tModLoader/tModLoader/files/5856526/Mac_server.log)
### Steps to reproduce
Not many ideas. On windows, installing [Armoury Crate](https://rog.asus.com/us/innovation/armoury_crate/) could work, but mac would need more testing.
### Expected behavior
The message `Tried to run two servers on the same PC` would at least be printed to the server's console & log file.
### Actual behavior
The server claims that it has started listening on its port, which is impossible.
### Additional Information
This very much could be a vanilla bug.
|
test
|
server port opening error handling is faulty in some cases description a few people in discord recently mentioned that their clients would get stuck at found server when using host play but the server logs looked fine with the only weird thing to note being that they didn t have x is connecting entries as necro figured out on their own this was caused by an asus application taking the port in case of which the game should ve logged the following only one usage of each socket address protocol network address port is normally permitted see log for full trace tried to run two servers on the same pc this message is logged properly when the port is taken by another instance of the game s server but potentially not in cases when driver level applications take it the weird thing is that this happened on a mac too log s attached below log file necro s windows mngrusin s mac steps to reproduce not many ideas on windows installing could work but mac would need more testing expected behavior the message tried to run two servers on the same pc would at least be printed to the server s console log file actual behavior the server claims that it has started listening on its port which is impossible additional information this very much could be a vanilla bug
| 1
|
172,894
| 13,351,102,851
|
IssuesEvent
|
2020-08-30 11:52:05
|
Lumieducation/H5P-Nodejs-library
|
https://api.github.com/repos/Lumieducation/H5P-Nodejs-library
|
closed
|
Use real-world examples from H5P.org
|
[type] test
|
We used to use examples that were scraped from H5P.org to run integration tests. I've moved these tests to the examples provided by the H5P Hub as this seemed to be more reliable. It has turned out that the files downloaded from the Hub are mostly identical to the examples, but not in all cases (e.g. H5P.InteractiveVideo is quite different). So we should re-introduce the scraping and add the files as further test files.
|
1.0
|
Use real-world examples from H5P.org - We used to use examples that were scraped from H5P.org to run integration tests. I've moved these tests to the examples provided by the H5P Hub as this seemed to be more reliable. It has turned out that the files downloaded from the Hub are mostly identical to the examples, but not in all cases (e.g. H5P.InteractiveVideo is quite different). So we should re-introduce the scraping and add the files as further test files.
|
test
|
use real world examples from org we used to use examples that were scraped from org to run integration tests i ve moved these tests to the examples provided by the hub as this seemed to be more reliable it has turned out that the files downloaded from the hub are mostly identical to the examples but not in all cases e g interactivevideo is quite different so we should re introduce the scraping and add the files as further test files
| 1
|
130,617
| 10,618,153,985
|
IssuesEvent
|
2019-10-13 01:31:44
|
magento/graphql-ce
|
https://api.github.com/repos/magento/graphql-ce
|
closed
|
[Test Coverage] Extend test coverage for CustomerDownloadableGraphQl
|
Component: CustomerDownloadableGraphQl Progress: PR created good first issue test-coverage
|
### Description:
Cover with API-functional tests:
https://github.com/magento/graphql-ce/blob/d4f0f25be66f04dcba68c5d8dcb77e8b686e8f96/app/code/Magento/CustomerDownloadableGraphQl/Model/Resolver/CustomerDownloadableProducts.php#L68
https://github.com/magento/graphql-ce/blob/d4f0f25be66f04dcba68c5d8dcb77e8b686e8f96/app/code/Magento/CustomerDownloadableGraphQl/Model/Resolver/CustomerDownloadableProducts.php#L69
### TestCases:
```
```
### Examples:
[Magento\GraphQl\Customer](https://github.com/magento/graphql-ce/tree/2.3-develop/dev/tests/api-functional/testsuite/Magento/GraphQl/Customer)
[Magento\GraphQl\Quote\Customer](https://github.com/magento/graphql-ce/tree/2.3-develop/dev/tests/api-functional/testsuite/Magento/GraphQl/Quote/Customer)
### Guide:
[Graphql Functional Testing](https://devdocs.magento.com/guides/v2.3/graphql/functional-testing.html)
|
1.0
|
[Test Coverage] Extend test coverage for CustomerDownloadableGraphQl - ### Description:
Cover with API-functional tests:
https://github.com/magento/graphql-ce/blob/d4f0f25be66f04dcba68c5d8dcb77e8b686e8f96/app/code/Magento/CustomerDownloadableGraphQl/Model/Resolver/CustomerDownloadableProducts.php#L68
https://github.com/magento/graphql-ce/blob/d4f0f25be66f04dcba68c5d8dcb77e8b686e8f96/app/code/Magento/CustomerDownloadableGraphQl/Model/Resolver/CustomerDownloadableProducts.php#L69
### TestCases:
```
```
### Examples:
[Magento\GraphQl\Customer](https://github.com/magento/graphql-ce/tree/2.3-develop/dev/tests/api-functional/testsuite/Magento/GraphQl/Customer)
[Magento\GraphQl\Quote\Customer](https://github.com/magento/graphql-ce/tree/2.3-develop/dev/tests/api-functional/testsuite/Magento/GraphQl/Quote/Customer)
### Guide:
[Graphql Functional Testing](https://devdocs.magento.com/guides/v2.3/graphql/functional-testing.html)
|
test
|
extend test coverage for customerdownloadablegraphql description cover with api functional tests testcases examples guide
| 1
|
159,209
| 12,466,867,442
|
IssuesEvent
|
2020-05-28 16:07:50
|
Princeton-CDH/mep-django
|
https://api.github.com/repos/Princeton-CDH/mep-django
|
closed
|
member data export does not include postal code / arrondissement
|
awaiting testing bug
|
## Notes for testing
Check [new test member export data](https://drive.google.com/drive/u/0/folders/1wCPd4iD2kNGflsmVSJ6z6FhlJx7ETDl3) to confirm that postal codes and arrondissements are included and have the correct values. Check both CSV and JSON; check members with multiple addresses, no addresses, and addresses that are outside of Paris (should not have an arrondissement).
|
1.0
|
member data export does not include postal code / arrondissement - ## Notes for testing
Check [new test member export data](https://drive.google.com/drive/u/0/folders/1wCPd4iD2kNGflsmVSJ6z6FhlJx7ETDl3) to confirm that postal codes and arrondissements are included and have the correct values. Check both CSV and JSON; check members with multiple addresses, no addresses, and addresses that are outside of Paris (should not have an arrondissement).
|
test
|
member data export does not include postal code arrondissement notes for testing check to confirm that postal codes and arrondissements are included and have the correct values check both csv and json check members with multiple addresses no addresses and addresses that are outside of paris should not have an arrondissement
| 1
|
172,598
| 13,325,002,655
|
IssuesEvent
|
2020-08-27 09:18:05
|
phanxgames/AetherStory
|
https://api.github.com/repos/phanxgames/AetherStory
|
closed
|
Construction Obstruction Quest NPC Bug
|
Needs testing QoL Bug
|
**Describe the bug**
This bug is caused after accepting the quest and continuing to talk to each worker as they are fading away their name is still visible,
**To Reproduce**
Steps to reproduce the behavior:
1. Get the quest "Construction Obstruction"
2. Proceed to find each worker and turn in the quest
3. As the NPC starts to walk away keep your cursor over the NPC and follow its pathing and as they are vanishing you can still see their name
4. See error
**Expected behavior**
As per turning quest into each npc they would walk away and vanish with their name vanishing with the npc
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Context:**
- Envy
- Windows 10
- Patch: 0.3.85
- Build: 351
**Additional context**
N/A
|
1.0
|
Construction Obstruction Quest NPC Bug - **Describe the bug**
This bug is caused after accepting the quest and continuing to talk to each worker as they are fading away their name is still visible,
**To Reproduce**
Steps to reproduce the behavior:
1. Get the quest "Construction Obstruction"
2. Proceed to find each worker and turn in the quest
3. As the NPC starts to walk away keep your cursor over the NPC and follow its pathing and as they are vanishing you can still see their name
4. See error
**Expected behavior**
As per turning quest into each npc they would walk away and vanish with their name vanishing with the npc
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Context:**
- Envy
- Windows 10
- Patch: 0.3.85
- Build: 351
**Additional context**
N/A
|
test
|
construction obstruction quest npc bug describe the bug this bug is caused after accepting the quest and continuing to talk to each worker as they are fading away their name is still visible to reproduce steps to reproduce the behavior get the quest construction obstruction proceed to find each worker and turn in the quest as the npc starts to walk away keep your cursor over the npc and follow its pathing and as they are vanishing you can still see their name see error expected behavior as per turning quest into each npc they would walk away and vanish with their name vanishing with the npc screenshots if applicable add screenshots to help explain your problem context envy windows patch build additional context n a
| 1
|
64,024
| 6,890,949,284
|
IssuesEvent
|
2017-11-22 15:35:02
|
healthlocker/healthlocker
|
https://api.github.com/repos/healthlocker/healthlocker
|
closed
|
Restrict Repeat access to new problem tracker page
|
bug please-test priority-2 T4h
|
2 test users experiencing an Internal server error when viewing the tracking overview.
@reddog looked at the appsignal monitoring and it seems to be where the database thinks someone has more than one problem tracker (which shouldn't be possible!).
I have managed to replicate it today with the following steps:
- Login in
- Click on problem tracker
- Set up new problem tracker
- enter tracking data
- Tracking overview works
But if you then
- either click the back button to get back to the 'set up new problem tracker' page or paste in the address 'healthlocker.uk/symptom/new'
- label another problem tracker, edit the name or leave don't change anything
- click save
- enter tracking data
- click on tracking overview now gives an internal server error.
Can we set it up so:
- [ ] adding a unique index for the user_id in the symptoms table, which will allow only one problem type per person.
- [ ] Returning to this page and trying to save a second problem tracker would generate an error flash message saying
"You can only set up your problem tracker once. Track your problem now"
NB. 'Track your problem now' links to https://www.healthlocker.uk/symptom-tracker/new
|
1.0
|
Restrict Repeat access to new problem tracker page - 2 test users experiencing an Internal server error when viewing the tracking overview.
@reddog looked at the appsignal monitoring and it seems to be where the database thinks someone has more than one problem tracker (which shouldn't be possible!).
I have managed to replicate it today with the following steps:
- Login in
- Click on problem tracker
- Set up new problem tracker
- enter tracking data
- Tracking overview works
But if you then
- either click the back button to get back to the 'set up new problem tracker' page or paste in the address 'healthlocker.uk/symptom/new'
- label another problem tracker, edit the name or leave don't change anything
- click save
- enter tracking data
- click on tracking overview now gives an internal server error.
Can we set it up so:
- [ ] adding a unique index for the user_id in the symptoms table, which will allow only one problem type per person.
- [ ] Returning to this page and trying to save a second problem tracker would generate an error flash message saying
"You can only set up your problem tracker once. Track your problem now"
NB. 'Track your problem now' links to https://www.healthlocker.uk/symptom-tracker/new
|
test
|
restrict repeat access to new problem tracker page test users experiencing an internal server error when viewing the tracking overview reddog looked at the appsignal monitoring and it seems to be where the database thinks someone has more than one problem tracker which shouldn t be possible i have managed to replicate it today with the following steps login in click on problem tracker set up new problem tracker enter tracking data tracking overview works but if you then either click the back button to get back to the set up new problem tracker page or paste in the address healthlocker uk symptom new label another problem tracker edit the name or leave don t change anything click save enter tracking data click on tracking overview now gives an internal server error can we set it up so adding a unique index for the user id in the symptoms table which will allow only one problem type per person returning to this page and trying to save a second problem tracker would generate an error flash message saying you can only set up your problem tracker once track your problem now nb track your problem now links to
| 1
|
14,506
| 25,015,093,643
|
IssuesEvent
|
2022-11-03 18:06:08
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
opened
|
Support gradle-node-plugin
|
type:feature status:requirements priority-5-triage
|
### What would you like Renovate to be able to do?
Be able to parse the node, npm, yarn, and pnpm versions from the node block in the `build.gradle`.
the plugin is specified as
```
plugins {
id 'com.github.node-gradle.node' version '3.2.1'
}
```
And lib versions like this https://github.com/node-gradle/gradle-node-plugin/blob/master/docs/usage.md#configuring-the-plugin
```
node {
version = '16.14.0'
download = true
}
```
https://github.com/node-gradle/gradle-node-plugin
### If you have any ideas on how this should be implemented, please tell us here.
There is an open issue with that plugin about being able to use files like `.nvmrc` (and I would argue `.node-version` to be non-lib-specific) to define the version which would make this a moot request, however, that may or may not happen.
Without looking at the renovate codebase, If I were trying to solve this myself I would simply use regex to look for the node block and versions within once I'd verified the existence of the plugin. Granted, that means tight coupling to said plugin which would then have to be watched and would be a pain if the interface ever changed.
### Is this a feature you are interested in implementing yourself?
Maybe
|
1.0
|
Support gradle-node-plugin - ### What would you like Renovate to be able to do?
Be able to parse the node, npm, yarn, and pnpm versions from the node block in the `build.gradle`.
the plugin is specified as
```
plugins {
id 'com.github.node-gradle.node' version '3.2.1'
}
```
And lib versions like this https://github.com/node-gradle/gradle-node-plugin/blob/master/docs/usage.md#configuring-the-plugin
```
node {
version = '16.14.0'
download = true
}
```
https://github.com/node-gradle/gradle-node-plugin
### If you have any ideas on how this should be implemented, please tell us here.
There is an open issue with that plugin about being able to use files like `.nvmrc` (and I would argue `.node-version` to be non-lib-specific) to define the version which would make this a moot request, however, that may or may not happen.
Without looking at the renovate codebase, If I were trying to solve this myself I would simply use regex to look for the node block and versions within once I'd verified the existence of the plugin. Granted, that means tight coupling to said plugin which would then have to be watched and would be a pain if the interface ever changed.
### Is this a feature you are interested in implementing yourself?
Maybe
|
non_test
|
support gradle node plugin what would you like renovate to be able to do be able to parse the node npm yarn and pnpm versions from the node block in the build gradle the plugin is specified as plugins id com github node gradle node version and lib versions like this node version download true if you have any ideas on how this should be implemented please tell us here there is an open issue with that plugin about being able to use files like nvmrc and i would argue node version to be non lib specific to define the version which would make this a moot request however that may or may not happen without looking at the renovate codebase if i were trying to solve this myself i would simply use regex to look for the node block and versions within once i d verified the existence of the plugin granted that means tight coupling to said plugin which would then have to be watched and would be a pain if the interface ever changed is this a feature you are interested in implementing yourself maybe
| 0
|
441,747
| 30,797,708,712
|
IssuesEvent
|
2023-07-31 21:21:44
|
mattermost/mattermost-handbook
|
https://api.github.com/repos/mattermost/mattermost-handbook
|
closed
|
Request for Documentation: FAQ: How to stop receiving notifications for all conversations in handbook repository?
|
Needs Documentation
|
Mattermost user `jason.blais` from https://community-release.mattermost.com has requested the following be documented:
```
#### How to stop receiving notifications for all conversations in handbook repository?
I've heard a few of you are getting inundated with notifications from handbook - to resolve it, go to https://github.com/mattermost/mattermost-handbook, click the "Watch" icon in top right, then choose your notification preferences.
Generally I'd recommend choosing the `Not Watching` option where you get notified only if you're participating (e.g. submit a pull request), or are @mentioned.
```
See the original post [here](https://community-release.mattermost.com/_redirect/pl/3e93f5hxb3nt3m8ewczhydhjyr).
_This issue was generated from [Mattermost](https://mattermost.com) using the [Doc Up](https://github.com/jwilander/mattermost-plugin-docup) plugin._
|
1.0
|
Request for Documentation: FAQ: How to stop receiving notifications for all conversations in handbook repository? - Mattermost user `jason.blais` from https://community-release.mattermost.com has requested the following be documented:
```
#### How to stop receiving notifications for all conversations in handbook repository?
I've heard a few of you are getting inundated with notifications from handbook - to resolve it, go to https://github.com/mattermost/mattermost-handbook, click the "Watch" icon in top right, then choose your notification preferences.
Generally I'd recommend choosing the `Not Watching` option where you get notified only if you're participating (e.g. submit a pull request), or are @mentioned.
```
See the original post [here](https://community-release.mattermost.com/_redirect/pl/3e93f5hxb3nt3m8ewczhydhjyr).
_This issue was generated from [Mattermost](https://mattermost.com) using the [Doc Up](https://github.com/jwilander/mattermost-plugin-docup) plugin._
|
non_test
|
request for documentation faq how to stop receiving notifications for all conversations in handbook repository mattermost user jason blais from has requested the following be documented how to stop receiving notifications for all conversations in handbook repository i ve heard a few of you are getting inundated with notifications from handbook to resolve it go to click the watch icon in top right then choose your notification preferences generally i d recommend choosing the not watching option where you get notified only if you re participating e g submit a pull request or are mentioned see the original post this issue was generated from using the plugin
| 0
|
325,061
| 27,845,629,376
|
IssuesEvent
|
2023-03-20 15:19:47
|
QubesOS/updates-status
|
https://api.github.com/repos/QubesOS/updates-status
|
closed
|
video-companion v1.0.1-1 (r4.1)
|
r4.1-buster-cur-test r4.1-dom0-cur-test r4.1-bullseye-cur-test r4.1-centos-stream8-cur-test r4.1-bookworm-cur-test r4.1-fc36-cur-test r4.1-fc37-cur-test r4.1-fc38-cur-test
|
Update of video-companion to v1.0.1-1 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-video-companion/commit/7b568f15cffe8ec0fde5aab37f5afe3d62ce6885
[Changes since previous version](https://github.com/QubesOS/qubes-video-companion/compare/v2.0.0...v1.0.1-1):
QubesOS/qubes-video-companion@7b568f1 version 1.0.1-1
QubesOS/qubes-video-companion@313ce9b pylint: disable consider-using-f-string
QubesOS/qubes-video-companion@038b4f8 tests: webcam test is supposed to work now
QubesOS/qubes-video-companion@ce725b8 webcam: workaround gstreamer bug for video/x-raw source
QubesOS/qubes-video-companion@741a487 Drop --buffer-size=0, it doesn't do anything
QubesOS/qubes-video-companion@dd8e6be tests: log more details on failure
QubesOS/qubes-video-companion@9bb6481 sender/webcam: support raw source stream too
QubesOS/qubes-video-companion@9dafa4e Remove 'colorimetry' pipeline constraint
QubesOS/qubes-video-companion@b3706f8 ci: limit pylint to actual package
QubesOS/qubes-video-companion@cb91413 tests: add skeleton for webcam test too
QubesOS/qubes-video-companion@14b44ce tests: add basic integration test for screen share
QubesOS/qubes-video-companion@f710822 Add v4l-utils dependency
QubesOS/qubes-video-companion@82c1038 Make pylint happy again
QubesOS/qubes-video-companion@683e401 Reformat code according to python-black
QubesOS/qubes-video-companion@240c140 Make pylint happy
QubesOS/qubes-video-companion@7800118 Add .pylintrc from core-admin and fix deps
QubesOS/qubes-video-companion@58cf260 Make shellcheck happy
QubesOS/qubes-video-companion@a790ea9 Qubes Builder integration
QubesOS/qubes-video-companion@51f9c35 Rework .gitlab-ci.yml
QubesOS/qubes-video-companion@c7215ba Allow @default in the qrexec policies
QubesOS/qubes-video-companion@e67ad19 Use @default as the destination if none is passed
QubesOS/qubes-video-companion@4939934 Fix debian/rules
QubesOS/qubes-video-companion@418a7c1 Convert icon.png
QubesOS/qubes-video-companion@695043a Make install a bit more robust
QubesOS/qubes-video-companion@3c7d3cd Check for /etc/qubes-release in dom0
QubesOS/qubes-video-companion@7af0651 Avoid overbroad except
QubesOS/qubes-video-companion@302352f Minor Python improvements
QubesOS/qubes-video-companion@0b8b901 Add basic comments to the qrexec policies
QubesOS/qubes-video-companion@fb6db6f Don't hardcode dom0 as a destination
QubesOS/qubes-video-companion@ca39b24 Avoid tripping -o pipefail
QubesOS/qubes-video-companion@43343bc Improve documentation
QubesOS/qubes-video-companion@ae28a82 Merge branch 'packaging-improvements' from DemiMarie
QubesOS/qubes-video-companion@72cd1c6 Small fix ups
QubesOS/qubes-video-companion@d7804d3 Improved packaging script
QubesOS/qubes-video-companion@ed259e9 Split the RPM package into separate packages
QubesOS/qubes-video-companion@ce98c5d Mark various targets as .PHONY
QubesOS/qubes-video-companion@e81f522 Avoid installing v4l2loopback scripts by default
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current repo` (available 7 days from now)
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
|
8.0
|
video-companion v1.0.1-1 (r4.1) - Update of video-companion to v1.0.1-1 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-video-companion/commit/7b568f15cffe8ec0fde5aab37f5afe3d62ce6885
[Changes since previous version](https://github.com/QubesOS/qubes-video-companion/compare/v2.0.0...v1.0.1-1):
QubesOS/qubes-video-companion@7b568f1 version 1.0.1-1
QubesOS/qubes-video-companion@313ce9b pylint: disable consider-using-f-string
QubesOS/qubes-video-companion@038b4f8 tests: webcam test is supposed to work now
QubesOS/qubes-video-companion@ce725b8 webcam: workaround gstreamer bug for video/x-raw source
QubesOS/qubes-video-companion@741a487 Drop --buffer-size=0, it doesn't do anything
QubesOS/qubes-video-companion@dd8e6be tests: log more details on failure
QubesOS/qubes-video-companion@9bb6481 sender/webcam: support raw source stream too
QubesOS/qubes-video-companion@9dafa4e Remove 'colorimetry' pipeline constraint
QubesOS/qubes-video-companion@b3706f8 ci: limit pylint to actual package
QubesOS/qubes-video-companion@cb91413 tests: add skeleton for webcam test too
QubesOS/qubes-video-companion@14b44ce tests: add basic integration test for screen share
QubesOS/qubes-video-companion@f710822 Add v4l-utils dependency
QubesOS/qubes-video-companion@82c1038 Make pylint happy again
QubesOS/qubes-video-companion@683e401 Reformat code according to python-black
QubesOS/qubes-video-companion@240c140 Make pylint happy
QubesOS/qubes-video-companion@7800118 Add .pylintrc from core-admin and fix deps
QubesOS/qubes-video-companion@58cf260 Make shellcheck happy
QubesOS/qubes-video-companion@a790ea9 Qubes Builder integration
QubesOS/qubes-video-companion@51f9c35 Rework .gitlab-ci.yml
QubesOS/qubes-video-companion@c7215ba Allow @default in the qrexec policies
QubesOS/qubes-video-companion@e67ad19 Use @default as the destination if none is passed
QubesOS/qubes-video-companion@4939934 Fix debian/rules
QubesOS/qubes-video-companion@418a7c1 Convert icon.png
QubesOS/qubes-video-companion@695043a Make install a bit more robust
QubesOS/qubes-video-companion@3c7d3cd Check for /etc/qubes-release in dom0
QubesOS/qubes-video-companion@7af0651 Avoid overbroad except
QubesOS/qubes-video-companion@302352f Minor Python improvements
QubesOS/qubes-video-companion@0b8b901 Add basic comments to the qrexec policies
QubesOS/qubes-video-companion@fb6db6f Don't hardcode dom0 as a destination
QubesOS/qubes-video-companion@ca39b24 Avoid tripping -o pipefail
QubesOS/qubes-video-companion@43343bc Improve documentation
QubesOS/qubes-video-companion@ae28a82 Merge branch 'packaging-improvements' from DemiMarie
QubesOS/qubes-video-companion@72cd1c6 Small fix ups
QubesOS/qubes-video-companion@d7804d3 Improved packaging script
QubesOS/qubes-video-companion@ed259e9 Split the RPM package into separate packages
QubesOS/qubes-video-companion@ce98c5d Mark various targets as .PHONY
QubesOS/qubes-video-companion@e81f522 Avoid installing v4l2loopback scripts by default
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current repo` (available 7 days from now)
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
|
test
|
video companion update of video companion to for qubes see comments below for details built from qubesos qubes video companion version qubesos qubes video companion pylint disable consider using f string qubesos qubes video companion tests webcam test is supposed to work now qubesos qubes video companion webcam workaround gstreamer bug for video x raw source qubesos qubes video companion drop buffer size it doesn t do anything qubesos qubes video companion tests log more details on failure qubesos qubes video companion sender webcam support raw source stream too qubesos qubes video companion remove colorimetry pipeline constraint qubesos qubes video companion ci limit pylint to actual package qubesos qubes video companion tests add skeleton for webcam test too qubesos qubes video companion tests add basic integration test for screen share qubesos qubes video companion add utils dependency qubesos qubes video companion make pylint happy again qubesos qubes video companion reformat code according to python black qubesos qubes video companion make pylint happy qubesos qubes video companion add pylintrc from core admin and fix deps qubesos qubes video companion make shellcheck happy qubesos qubes video companion qubes builder integration qubesos qubes video companion rework gitlab ci yml qubesos qubes video companion allow default in the qrexec policies qubesos qubes video companion use default as the destination if none is passed qubesos qubes video companion fix debian rules qubesos qubes video companion convert icon png qubesos qubes video companion make install a bit more robust qubesos qubes video companion check for etc qubes release in qubesos qubes video companion avoid overbroad except qubesos qubes video companion minor python improvements qubesos qubes video companion add basic comments to the qrexec policies qubesos qubes video companion don t hardcode as a destination qubesos qubes video companion avoid tripping o pipefail qubesos qubes video companion improve documentation qubesos qubes video companion merge branch packaging improvements from demimarie qubesos qubes video companion small fix ups qubesos qubes video companion improved packaging script qubesos qubes video companion split the rpm package into separate packages qubesos qubes video companion mark various targets as phony qubesos qubes video companion avoid installing scripts by default referenced issues if you re release manager you can issue gpg inline signed command upload video companion current repo available days from now upload video companion current dists repo you can choose subset of distributions like vm vm available days from now upload video companion security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
| 1
|
325,108
| 27,848,223,091
|
IssuesEvent
|
2023-03-20 16:50:04
|
gear-tech/gear
|
https://api.github.com/repos/gear-tech/gear
|
closed
|
WASM-GEN: generate handle_reply method
|
C1-feature D4-test
|
### Problem to Solve
Improve simulation of request/reply interaction.
### Possible Solution
analogous to generation of `handle`/`init` methods
### Notes
_No response_
|
1.0
|
WASM-GEN: generate handle_reply method - ### Problem to Solve
Improve simulation of request/reply interaction.
### Possible Solution
analogous to generation of `handle`/`init` methods
### Notes
_No response_
|
test
|
wasm gen generate handle reply method problem to solve improve simulation of request reply interaction possible solution analogous to generation of handle init methods notes no response
| 1
|
183,871
| 31,778,484,633
|
IssuesEvent
|
2023-09-12 15:46:16
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
AnalyzeDataFlow reports captured primary ctor parameters in lambdas that do not capture any
|
Bug Area-Compilers Resolution-By Design
|
`semanticModel.AnalyzeDataFlow(node).Captured` with `node` being the expression body of the lambda (i.e. `2` in the code below) reports `x` as captured.
`x` is however only captured by `F`, not by the lambda.
```C#
using System;
class C(int x)
{
void F()
{
var f = new Func<int, int>(a => 2);
var y = x;
}
}
```
The documentation of `DataFlowAnalysis.Captured` also needs some updating:
```
/// <summary>
/// The set of the local variables that have been referenced in anonymous
/// functions within a region and therefore must be moved to a field of a frame class.
/// </summary>
public abstract ImmutableArray<ISymbol> Captured { get; }
```
|
1.0
|
AnalyzeDataFlow reports captured primary ctor parameters in lambdas that do not capture any - `semanticModel.AnalyzeDataFlow(node).Captured` with `node` being the expression body of the lambda (i.e. `2` in the code below) reports `x` as captured.
`x` is however only captured by `F`, not by the lambda.
```C#
using System;
class C(int x)
{
void F()
{
var f = new Func<int, int>(a => 2);
var y = x;
}
}
```
The documentation of `DataFlowAnalysis.Captured` also needs some updating:
```
/// <summary>
/// The set of the local variables that have been referenced in anonymous
/// functions within a region and therefore must be moved to a field of a frame class.
/// </summary>
public abstract ImmutableArray<ISymbol> Captured { get; }
```
|
non_test
|
analyzedataflow reports captured primary ctor parameters in lambdas that do not capture any semanticmodel analyzedataflow node captured with node being the expression body of the lambda i e in the code below reports x as captured x is however only captured by f not by the lambda c using system class c int x void f var f new func a var y x the documentation of dataflowanalysis captured also needs some updating the set of the local variables that have been referenced in anonymous functions within a region and therefore must be moved to a field of a frame class public abstract immutablearray captured get
| 0
|
42,283
| 5,431,374,522
|
IssuesEvent
|
2017-03-04 00:32:31
|
gnieboer/GNURadio_Windows_Build_Scripts
|
https://api.github.com/repos/gnieboer/GNURadio_Windows_Build_Scripts
|
closed
|
UDP Sources crash
|
bug in test upstream
|
See also https://github.com/pothosware/gnuradio/issues/23
udp_source_impl.cc is setting the SO_LINGER option on the UDP socket, which at least on Windows, causes a WSAENOPROTOOPT exception, because linger doesn't really mean anything for a UDP socket.
Bug is an upstream bug in GNURadio, this is for tracking purposes only.
|
1.0
|
UDP Sources crash - See also https://github.com/pothosware/gnuradio/issues/23
udp_source_impl.cc is setting the SO_LINGER option on the UDP socket, which at least on Windows, causes a WSAENOPROTOOPT exception, because linger doesn't really mean anything for a UDP socket.
Bug is an upstream bug in GNURadio, this is for tracking purposes only.
|
test
|
udp sources crash see also udp source impl cc is setting the so linger option on the udp socket which at least on windows causes a wsaenoprotoopt exception because linger doesn t really mean anything for a udp socket bug is an upstream bug in gnuradio this is for tracking purposes only
| 1
|
59,687
| 6,660,687,803
|
IssuesEvent
|
2017-10-02 02:46:14
|
borgbackup/borg
|
https://api.github.com/repos/borgbackup/borg
|
closed
|
centos: EISDIR
|
testsuite vagrant
|
a metric ton of "IsADirectoryError" due to fakeroot usage there.
centos7 works quite ok without fakeroot.
|
1.0
|
centos: EISDIR - a metric ton of "IsADirectoryError" due to fakeroot usage there.
centos7 works quite ok without fakeroot.
|
test
|
centos eisdir a metric ton of isadirectoryerror due to fakeroot usage there works quite ok without fakeroot
| 1
|
85,887
| 24,709,418,268
|
IssuesEvent
|
2022-10-19 22:25:37
|
dotnet/arcade
|
https://api.github.com/repos/dotnet/arcade
|
closed
|
Build failed: arcade-services-internal-ci/main #20221019.2
|
Build Failed
|
Build [#20221019.2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2024976) failed
## :x: : internal / arcade-services-internal-ci failed
### Summary
**Finished** - Wed, 19 Oct 2022 19:12:49 GMT
**Duration** - 199 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Validate deployment
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Test Run Failed.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Vstest failed with error. Check logs for failures. There might be failed tests.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Error: The process 'D:\a\_work\_tasks\VSTest_ef087383-ee5e-42c7-9a53-ab56c98420f9\2.210.0\Modules\DTAExecutionHost.exe' failed with exit code 1
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Vstest failed with error. Check logs for failures. There might be failed tests.
#### Validate Build Assets
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/155) - (NETCORE_ENGINEERING_TELEMETRY=Sdl) Last command failed with exit code 1.
#### Deploy
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/232) - Error: Failed to update deployment history. Error: Conflict (CODE: 409)
#### Build
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/52) - Could not parse Jtokens from file D:\a\_work\1\s\src\Maestro\maestro-angular\node_modules\resolve\test\resolver\malformed_package_json\package.json.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/64) - Could not parse Jtokens from file D:\a\_work\1\s\src\Maestro\maestro-angular\node_modules\resolve\test\resolver\malformed_package_json\package.json.
### Changes
- [0626a5de](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/ec29e881-6ad7-4427-832d-ce639ccba518/commit/0626a5deb28e9cd81b08a13132424a491d101a4f) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20221017.1 (#2064)
- [f599f273](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/ec29e881-6ad7-4427-832d-ce639ccba518/commit/f599f273949253b9826abb99f1bb12fabdb5bb00) - Přemek Vysoký - Checkout individual repo clones at target revision (#2063)
|
1.0
|
Build failed: arcade-services-internal-ci/main #20221019.2 - Build [#20221019.2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2024976) failed
## :x: : internal / arcade-services-internal-ci failed
### Summary
**Finished** - Wed, 19 Oct 2022 19:12:49 GMT
**Duration** - 199 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Validate deployment
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Test Run Failed.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Vstest failed with error. Check logs for failures. There might be failed tests.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Error: The process 'D:\a\_work\_tasks\VSTest_ef087383-ee5e-42c7-9a53-ab56c98420f9\2.210.0\Modules\DTAExecutionHost.exe' failed with exit code 1
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/298) - Vstest failed with error. Check logs for failures. There might be failed tests.
#### Validate Build Assets
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/155) - (NETCORE_ENGINEERING_TELEMETRY=Sdl) Last command failed with exit code 1.
#### Deploy
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/232) - Error: Failed to update deployment history. Error: Conflict (CODE: 409)
#### Build
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/52) - Could not parse Jtokens from file D:\a\_work\1\s\src\Maestro\maestro-angular\node_modules\resolve\test\resolver\malformed_package_json\package.json.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2024976/logs/64) - Could not parse Jtokens from file D:\a\_work\1\s\src\Maestro\maestro-angular\node_modules\resolve\test\resolver\malformed_package_json\package.json.
### Changes
- [0626a5de](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/ec29e881-6ad7-4427-832d-ce639ccba518/commit/0626a5deb28e9cd81b08a13132424a491d101a4f) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20221017.1 (#2064)
- [f599f273](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/ec29e881-6ad7-4427-832d-ce639ccba518/commit/f599f273949253b9826abb99f1bb12fabdb5bb00) - Přemek Vysoký - Checkout individual repo clones at target revision (#2063)
|
non_test
|
build failed arcade services internal ci main build failed x internal arcade services internal ci failed summary finished wed oct gmt duration minutes requested for dotnet bot reason batchedci details validate deployment x test run failed warning vstest failed with error check logs for failures there might be failed tests x error the process d a work tasks vstest modules dtaexecutionhost exe failed with exit code x vstest failed with error check logs for failures there might be failed tests validate build assets x netcore engineering telemetry sdl last command failed with exit code deploy warning error failed to update deployment history error conflict code build warning could not parse jtokens from file d a work s src maestro maestro angular node modules resolve test resolver malformed package json package json warning could not parse jtokens from file d a work s src maestro maestro angular node modules resolve test resolver malformed package json package json changes dotnet maestro update dependencies from build přemek vysoký checkout individual repo clones at target revision
| 0
|
50,863
| 6,130,087,837
|
IssuesEvent
|
2017-06-24 01:41:11
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Test {e2e.go}
|
kind/flake needs-sig priority/failing-test
|
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-kops-aws-slow/1305/
Failed: Test {e2e.go}
```
error during ./hack/ginkgo-e2e.sh --ginkgo.focus=\[Slow\] --ginkgo.skip=\[Serial\]|\[Disruptive\]|\[Flaky\]|\[Feature:.+\]|\[HPA\]|Dashboard|Services.*functioning.*NodePort: exit status 1
```
Previous issues for this test: #33361 #38663 #39788 #39877 #40371 #40469 #40478 #40483 #40668 #41048 #43025
|
1.0
|
Test {e2e.go} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-kops-aws-slow/1305/
Failed: Test {e2e.go}
```
error during ./hack/ginkgo-e2e.sh --ginkgo.focus=\[Slow\] --ginkgo.skip=\[Serial\]|\[Disruptive\]|\[Flaky\]|\[Feature:.+\]|\[HPA\]|Dashboard|Services.*functioning.*NodePort: exit status 1
```
Previous issues for this test: #33361 #38663 #39788 #39877 #40371 #40469 #40478 #40483 #40668 #41048 #43025
|
test
|
test go failed test go error during hack ginkgo sh ginkgo focus ginkgo skip dashboard services functioning nodeport exit status previous issues for this test
| 1
|
171,142
| 13,220,701,559
|
IssuesEvent
|
2020-08-17 12:53:40
|
ICIJ/datashare
|
https://api.github.com/repos/ICIJ/datashare
|
closed
|
Batch search page: fields are not aligned
|
bug front need testing
|
**Describe the bug**
In some French and Spanish only (not in English nor Japanese), the fields that describes the batch search in the right column are not right-aligned. It prevents from reading the data in Spanish.
**To Reproduce**
1. Go to 'https://datashare-demo.icij.org/#/batch-search/luxleaks/24cca19c-2f5e-418b-a66e-bf9aed4c9ad2?page=1&sort=doc_nb&order=asc
2. Translate in French or Spanish
3. See error:
<img width="1076" alt="Screenshot 2020-08-14 at 10 33 53" src="https://user-images.githubusercontent.com/17233829/90230409-ccd0b500-de19-11ea-99d9-de54db798c83.png">
<img width="1105" alt="Screenshot 2020-08-14 at 10 33 46" src="https://user-images.githubusercontent.com/17233829/90230413-ce9a7880-de19-11ea-9cb2-8d615b08e301.png">
<img width="1124" alt="Screenshot 2020-08-14 at 10 33 33" src="https://user-images.githubusercontent.com/17233829/90230418-cf330f00-de19-11ea-936b-d05fdcae735f.png">
<img width="978" alt="Screenshot 2020-08-14 at 10 33 14" src="https://user-images.githubusercontent.com/17233829/90230419-cfcba580-de19-11ea-9f9d-f7594669bdfd.png">
**Expected behavior**
Right-align the "do phrase matches" and "proximity searches" fields.
**Desktop (please complete the following information):**
- OS: MacBook Pro (13-inch, 2019, Four Thunderbolt 3 ports)
- Browser: Chrome
- Version: 7.5.2
|
1.0
|
Batch search page: fields are not aligned - **Describe the bug**
In some French and Spanish only (not in English nor Japanese), the fields that describes the batch search in the right column are not right-aligned. It prevents from reading the data in Spanish.
**To Reproduce**
1. Go to 'https://datashare-demo.icij.org/#/batch-search/luxleaks/24cca19c-2f5e-418b-a66e-bf9aed4c9ad2?page=1&sort=doc_nb&order=asc
2. Translate in French or Spanish
3. See error:
<img width="1076" alt="Screenshot 2020-08-14 at 10 33 53" src="https://user-images.githubusercontent.com/17233829/90230409-ccd0b500-de19-11ea-99d9-de54db798c83.png">
<img width="1105" alt="Screenshot 2020-08-14 at 10 33 46" src="https://user-images.githubusercontent.com/17233829/90230413-ce9a7880-de19-11ea-9cb2-8d615b08e301.png">
<img width="1124" alt="Screenshot 2020-08-14 at 10 33 33" src="https://user-images.githubusercontent.com/17233829/90230418-cf330f00-de19-11ea-936b-d05fdcae735f.png">
<img width="978" alt="Screenshot 2020-08-14 at 10 33 14" src="https://user-images.githubusercontent.com/17233829/90230419-cfcba580-de19-11ea-9f9d-f7594669bdfd.png">
**Expected behavior**
Right-align the "do phrase matches" and "proximity searches" fields.
**Desktop (please complete the following information):**
- OS: MacBook Pro (13-inch, 2019, Four Thunderbolt 3 ports)
- Browser: Chrome
- Version: 7.5.2
|
test
|
batch search page fields are not aligned describe the bug in some french and spanish only not in english nor japanese the fields that describes the batch search in the right column are not right aligned it prevents from reading the data in spanish to reproduce go to translate in french or spanish see error img width alt screenshot at src img width alt screenshot at src img width alt screenshot at src img width alt screenshot at src expected behavior right align the do phrase matches and proximity searches fields desktop please complete the following information os macbook pro inch four thunderbolt ports browser chrome version
| 1
|
84,131
| 15,720,846,360
|
IssuesEvent
|
2021-03-29 01:23:50
|
rsoreq/cwa-server
|
https://api.github.com/repos/rsoreq/cwa-server
|
opened
|
CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.35.jar
|
security vulnerability
|
## CVE-2021-24122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.35.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: cwa-server/services/submission/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.35/tomcat-embed-core-9.0.35.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.3.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.3.0.RELEASE.jar
- :x: **tomcat-embed-core-9.0.35.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.
<p>Publish Date: 2021-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p>
<p>Release Date: 2021-01-14</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"9.0.35","packageFilePaths":["/services/submission/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.3.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.3.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:9.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-24122","vulnerabilityDetails":"When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.35.jar - ## CVE-2021-24122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.35.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: cwa-server/services/submission/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.35/tomcat-embed-core-9.0.35.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.3.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.3.0.RELEASE.jar
- :x: **tomcat-embed-core-9.0.35.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.
<p>Publish Date: 2021-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p>
<p>Release Date: 2021-01-14</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"9.0.35","packageFilePaths":["/services/submission/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.3.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.3.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:9.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-24122","vulnerabilityDetails":"When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in tomcat embed core jar cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file cwa server services submission pom xml path to vulnerable library root repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when serving resources from a network location using the ntfs file system apache tomcat versions to to to and to were susceptible to jsp source code disclosure in some configurations the root cause was the unexpected behaviour of the jre api file getcanonicalpath which in turn was caused by the inconsistent behaviour of the windows api findfirstfilew in some circumstances publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina basebranches vulnerabilityidentifier cve vulnerabilitydetails when serving resources from a network location using the ntfs file system apache tomcat versions to to to and to were susceptible to jsp source code disclosure in some configurations the root cause was the unexpected behaviour of the jre api file getcanonicalpath which in turn was caused by the inconsistent behaviour of the windows api findfirstfilew in some circumstances vulnerabilityurl
| 0
|
15,653
| 3,479,431,781
|
IssuesEvent
|
2015-12-28 20:15:20
|
spacetelescope/stginga
|
https://api.github.com/repos/spacetelescope/stginga
|
closed
|
Reimplement Qt specific codes to use Ginga's general widgets
|
BackgroundSub DQInspect enhancement low-priority testing
|
~~Whenever it is available under Anaconda, we have to make sure it works under both Qt4 and Qt5.~~
Move away from Qt-specific codes by using Ginga's toolkit-agnostic widgets. Then, we can stop requiring `stginga` to only work with Qt toolkit.
|
1.0
|
Reimplement Qt specific codes to use Ginga's general widgets - ~~Whenever it is available under Anaconda, we have to make sure it works under both Qt4 and Qt5.~~
Move away from Qt-specific codes by using Ginga's toolkit-agnostic widgets. Then, we can stop requiring `stginga` to only work with Qt toolkit.
|
test
|
reimplement qt specific codes to use ginga s general widgets whenever it is available under anaconda we have to make sure it works under both and move away from qt specific codes by using ginga s toolkit agnostic widgets then we can stop requiring stginga to only work with qt toolkit
| 1
|
269,242
| 23,431,873,721
|
IssuesEvent
|
2022-08-15 04:07:18
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
opened
|
DISABLED test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32 (__main__.TestEagerFusionOpInfoCPU)
|
module: flaky-tests skipped module: unknown
|
Platforms: mac, macos
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32&suite=TestEagerFusionOpInfoCPU&file=/Users/runner/work/pytorch/pytorch/functorch/test/test_pythonkey.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7829747412).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
|
1.0
|
DISABLED test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32 (__main__.TestEagerFusionOpInfoCPU) - Platforms: mac, macos
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32&suite=TestEagerFusionOpInfoCPU&file=/Users/runner/work/pytorch/pytorch/functorch/test/test_pythonkey.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7829747412).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_aot_autograd_exhaustive_max_reduction_no_dim_cpu_float32`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
|
test
|
disabled test aot autograd exhaustive max reduction no dim cpu main testeagerfusionopinfocpu platforms mac macos this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test aot autograd exhaustive max reduction no dim cpu there should be several instances run as flaky tests are rerun in ci from which you can study the logs
| 1
|
225,965
| 7,496,770,524
|
IssuesEvent
|
2018-04-08 13:06:27
|
CS2103JAN2018-W09-B3/main
|
https://api.github.com/repos/CS2103JAN2018-W09-B3/main
|
closed
|
Implement Sync
|
priority.high type.enhancement type.epic
|
Overwrite data in Google Contacts and Calendar
- [ ] #136 Sync Contacts
- [ ] #137 Sync Calendar
|
1.0
|
Implement Sync - Overwrite data in Google Contacts and Calendar
- [ ] #136 Sync Contacts
- [ ] #137 Sync Calendar
|
non_test
|
implement sync overwrite data in google contacts and calendar sync contacts sync calendar
| 0
|
737,155
| 25,503,780,146
|
IssuesEvent
|
2022-11-28 07:40:28
|
Digital-Will-Inc/wortal-sdk-unity
|
https://api.github.com/repos/Digital-Will-Inc/wortal-sdk-unity
|
closed
|
Test backwards compatibility
|
Priority: Medium Type: Chore
|
### Description
<p>Backwards compatibility should be tested to determine which engine versions the plugin does and does not work with. Workarounds can be implemented using compiler directives for specific versions.</p>
The final versions of 2019 and 2020 should be tested only.
- Type: Chore
- Priority: Normal
|
1.0
|
Test backwards compatibility - ### Description
<p>Backwards compatibility should be tested to determine which engine versions the plugin does and does not work with. Workarounds can be implemented using compiler directives for specific versions.</p>
The final versions of 2019 and 2020 should be tested only.
- Type: Chore
- Priority: Normal
|
non_test
|
test backwards compatibility description backwards compatibility should be tested to determine which engine versions the plugin does and does not work with workarounds can be implemented using compiler directives for specific versions the final versions of and should be tested only type chore priority normal
| 0
|
166,460
| 6,305,047,014
|
IssuesEvent
|
2017-07-21 17:23:35
|
minio/minio
|
https://api.github.com/repos/minio/minio
|
closed
|
Minio server with TLS
|
priority: medium triage
|
* Version used (`minio version`):
Version: 2017-06-13T19:01:01Z
Release-Tag: RELEASE.2017-06-13T19-01-01Z
Commit-ID: 353f2d3a6e30f595a721abcd9cf1b108b041ee21
* Server type and version: Docker (minio/minio:latest)
* Operating System and version (`uname -a`):
* Link to your project: https://prefiles.goodship.eu:9000/minio/login
Hello,
I have read the documentation (https://docs.minio.io/docs/how-to-secure-access-to-minio-server-with-tls) for generate key/certificate avec Openssl. I have installer certificat on /.minio/certs
So, when I want used https, I have this error : SSL_ERROR_INTERNAL_ERROR_ALERT
I have installed the Docker Minio in Rancher environment.
Thanks for your help.
Pierre.
|
1.0
|
Minio server with TLS - * Version used (`minio version`):
Version: 2017-06-13T19:01:01Z
Release-Tag: RELEASE.2017-06-13T19-01-01Z
Commit-ID: 353f2d3a6e30f595a721abcd9cf1b108b041ee21
* Server type and version: Docker (minio/minio:latest)
* Operating System and version (`uname -a`):
* Link to your project: https://prefiles.goodship.eu:9000/minio/login
Hello,
I have read the documentation (https://docs.minio.io/docs/how-to-secure-access-to-minio-server-with-tls) for generate key/certificate avec Openssl. I have installer certificat on /.minio/certs
So, when I want used https, I have this error : SSL_ERROR_INTERNAL_ERROR_ALERT
I have installed the Docker Minio in Rancher environment.
Thanks for your help.
Pierre.
|
non_test
|
minio server with tls version used minio version version release tag release commit id server type and version docker minio minio latest operating system and version uname a link to your project hello i have read the documentation for generate key certificate avec openssl i have installer certificat on minio certs so when i want used https i have this error ssl error internal error alert i have installed the docker minio in rancher environment thanks for your help pierre
| 0
|
20,687
| 10,547,532,246
|
IssuesEvent
|
2019-10-03 01:33:15
|
TIBCOSoftware/bw-sample-for-zendesk
|
https://api.github.com/repos/TIBCOSoftware/bw-sample-for-zendesk
|
opened
|
CVE-2019-16942 (Medium) detected in jackson-databind-2.1.4.jar
|
security vulnerability
|
## CVE-2019-16942 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://wiki.fasterxml.com/JacksonHome">http://wiki.fasterxml.com/JacksonHome</a></p>
<p>Path to vulnerable library: /bw-sample-for-zendesk/tcizendesk.module/lib/jackson-databind-2.1.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.1.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942>CVE-2019-16942</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: 2.10.0.pr1</p>
</p>
</details>
<p></p>
|
True
|
CVE-2019-16942 (Medium) detected in jackson-databind-2.1.4.jar - ## CVE-2019-16942 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://wiki.fasterxml.com/JacksonHome">http://wiki.fasterxml.com/JacksonHome</a></p>
<p>Path to vulnerable library: /bw-sample-for-zendesk/tcizendesk.module/lib/jackson-databind-2.1.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.1.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942>CVE-2019-16942</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: 2.10.0.pr1</p>
</p>
</details>
<p></p>
|
non_test
|
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library bw sample for zendesk tcizendesk module lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution
| 0
|
207,544
| 15,821,727,316
|
IssuesEvent
|
2021-04-05 20:59:53
|
uoForms/App-CANBeWell
|
https://api.github.com/repos/uoForms/App-CANBeWell
|
closed
|
changes in button-Terms of Use page
|
App CanBeWell 2.0 To be tested enhancement
|
this task is changed and we need a button under the "terms of use" which says I accept the
term of use and we need to remove accept and decline button.
-Nasim-
@navpreetkaur051
|
1.0
|
changes in button-Terms of Use page - this task is changed and we need a button under the "terms of use" which says I accept the
term of use and we need to remove accept and decline button.
-Nasim-
@navpreetkaur051
|
test
|
changes in button terms of use page this task is changed and we need a button under the terms of use which says i accept the term of use and we need to remove accept and decline button nasim
| 1
|
22,225
| 3,942,031,450
|
IssuesEvent
|
2016-04-27 10:14:29
|
menatwork/syncCto
|
https://api.github.com/repos/menatwork/syncCto
|
closed
|
client status not updating in ie / error in ping.js
|
Completed Testing
|
The client legend (this round indicator) does not update (shows 'grey' all the time) under ie 11.
The javascript console outputs an error that points to here:
https://github.com/menatwork/syncCto/blob/master/system/modules/syncCto/assets/js/ping.js#L8
``SCRIPT438: Das Objekt unterstützt die Eigenschaft oder Methode "getSiblings" nicht.``
(syncCto 2.6.2)
|
1.0
|
client status not updating in ie / error in ping.js - The client legend (this round indicator) does not update (shows 'grey' all the time) under ie 11.
The javascript console outputs an error that points to here:
https://github.com/menatwork/syncCto/blob/master/system/modules/syncCto/assets/js/ping.js#L8
``SCRIPT438: Das Objekt unterstützt die Eigenschaft oder Methode "getSiblings" nicht.``
(syncCto 2.6.2)
|
test
|
client status not updating in ie error in ping js the client legend this round indicator does not update shows grey all the time under ie the javascript console outputs an error that points to here das objekt unterstützt die eigenschaft oder methode getsiblings nicht synccto
| 1
|
86,347
| 16,976,494,931
|
IssuesEvent
|
2021-06-30 00:15:47
|
uf-mil/mil
|
https://api.github.com/repos/uf-mil/mil
|
opened
|
Make New Hallway Posters for Sub 8 and 9
|
documentation electrical good first issue mechanical non-code software subjugator8 subjugator9
|
Outside MAEC126 there are some presentation posters that were made long ago. If we could replicate the style and design some new ones, that would be a great starting place for potential lab members and visiting dignitaries.
Relatively low priority, but a good starting place for new members to get familiar with the subs specs and finding out information on it, rendering images from Solidworks, and designing presentation documents.
|
1.0
|
Make New Hallway Posters for Sub 8 and 9 - Outside MAEC126 there are some presentation posters that were made long ago. If we could replicate the style and design some new ones, that would be a great starting place for potential lab members and visiting dignitaries.
Relatively low priority, but a good starting place for new members to get familiar with the subs specs and finding out information on it, rendering images from Solidworks, and designing presentation documents.
|
non_test
|
make new hallway posters for sub and outside there are some presentation posters that were made long ago if we could replicate the style and design some new ones that would be a great starting place for potential lab members and visiting dignitaries relatively low priority but a good starting place for new members to get familiar with the subs specs and finding out information on it rendering images from solidworks and designing presentation documents
| 0
|
170,891
| 13,207,810,488
|
IssuesEvent
|
2020-08-15 00:44:32
|
Tadukoo/TadukooUtil
|
https://api.github.com/repos/Tadukoo/TadukooUtil
|
opened
|
[TESTING] Tadukoo Annotation Processor, Tadukoo Lang, Tadukoo Util
|
Testing
|
**What change would you like to see?**
Proper JUnit testing for the following modules:
- Tadukoo Annotation Processor
- Tadukoo Lang
- Tadukoo Util
**How does this change help?**
Provides testing coverage for the above listed modules, that isn't currently covered. This also makes sure we're ready for Alpha v.0.1
**Additional context**
Here's a TODO list for all the classes in the modules to keep track:
Tadukoo Annotation Processor:
- [ ] base package
- [ ] AbstractAnnotationProcessor
- [ ] AnnotationProcessor
- [ ] AnnotationProcessorProcessor
- [ ] AnnotationUtil
Tadukoo Lang
- [ ] base package
- [ ] AutoCloseableUtil
- [ ] BooleanUtil
- [ ] ExceptionUtil
- [ ] FileUtil
- [ ] ListUtil
- [ ] LoggerUtil
- [ ] StringUtil
- [ ] tuple package
- [ ] Pair
- [ ] Triple
Tadukoo Util
- [ ] annotation package
- [ ] ShouldBeFinal
- [ ] annotation.process package
- [ ] ShouldBeFinalProcessor
- [ ] event package
- [ ] Event
- [ ] EventHandler
- [ ] EventListener
- [ ] functional package
- [ ] NoException
- [ ] functional.consumer package
- [ ] ThrowingConsumer
- [ ] ThrowingConsumer2
- [ ] ThrowingConsumer3
- [ ] ThrowingConsumer4
- [ ] ThrowingConsumer5
- [ ] functional.function package
- [ ] ThrowingFunction
- [ ] ThrowingFunction2
- [ ] ThrowingFunction3
- [ ] ThrowingFunction4
- [ ] ThrowingFunction5
- [ ] functional.predicate package
- [ ] ThrowingPredicate
- [ ] ThrowingPredicate2
- [ ] ThrowingPredicate3
- [ ] ThrowingPredicate4
- [ ] ThrowingPredicate5
- [ ] functional.supplier package
- [ ] ThrowingSupplier
- [ ] map package
- [ ] HashManyToManyMap
- [ ] HashMultiMap
- [ ] ManyToManyMap
- [ ] MultiMap
- [ ] TreeMultiMap
|
1.0
|
[TESTING] Tadukoo Annotation Processor, Tadukoo Lang, Tadukoo Util - **What change would you like to see?**
Proper JUnit testing for the following modules:
- Tadukoo Annotation Processor
- Tadukoo Lang
- Tadukoo Util
**How does this change help?**
Provides testing coverage for the above listed modules, that isn't currently covered. This also makes sure we're ready for Alpha v.0.1
**Additional context**
Here's a TODO list for all the classes in the modules to keep track:
Tadukoo Annotation Processor:
- [ ] base package
- [ ] AbstractAnnotationProcessor
- [ ] AnnotationProcessor
- [ ] AnnotationProcessorProcessor
- [ ] AnnotationUtil
Tadukoo Lang
- [ ] base package
- [ ] AutoCloseableUtil
- [ ] BooleanUtil
- [ ] ExceptionUtil
- [ ] FileUtil
- [ ] ListUtil
- [ ] LoggerUtil
- [ ] StringUtil
- [ ] tuple package
- [ ] Pair
- [ ] Triple
Tadukoo Util
- [ ] annotation package
- [ ] ShouldBeFinal
- [ ] annotation.process package
- [ ] ShouldBeFinalProcessor
- [ ] event package
- [ ] Event
- [ ] EventHandler
- [ ] EventListener
- [ ] functional package
- [ ] NoException
- [ ] functional.consumer package
- [ ] ThrowingConsumer
- [ ] ThrowingConsumer2
- [ ] ThrowingConsumer3
- [ ] ThrowingConsumer4
- [ ] ThrowingConsumer5
- [ ] functional.function package
- [ ] ThrowingFunction
- [ ] ThrowingFunction2
- [ ] ThrowingFunction3
- [ ] ThrowingFunction4
- [ ] ThrowingFunction5
- [ ] functional.predicate package
- [ ] ThrowingPredicate
- [ ] ThrowingPredicate2
- [ ] ThrowingPredicate3
- [ ] ThrowingPredicate4
- [ ] ThrowingPredicate5
- [ ] functional.supplier package
- [ ] ThrowingSupplier
- [ ] map package
- [ ] HashManyToManyMap
- [ ] HashMultiMap
- [ ] ManyToManyMap
- [ ] MultiMap
- [ ] TreeMultiMap
|
test
|
tadukoo annotation processor tadukoo lang tadukoo util what change would you like to see proper junit testing for the following modules tadukoo annotation processor tadukoo lang tadukoo util how does this change help provides testing coverage for the above listed modules that isn t currently covered this also makes sure we re ready for alpha v additional context here s a todo list for all the classes in the modules to keep track tadukoo annotation processor base package abstractannotationprocessor annotationprocessor annotationprocessorprocessor annotationutil tadukoo lang base package autocloseableutil booleanutil exceptionutil fileutil listutil loggerutil stringutil tuple package pair triple tadukoo util annotation package shouldbefinal annotation process package shouldbefinalprocessor event package event eventhandler eventlistener functional package noexception functional consumer package throwingconsumer functional function package throwingfunction functional predicate package throwingpredicate functional supplier package throwingsupplier map package hashmanytomanymap hashmultimap manytomanymap multimap treemultimap
| 1
|
299,670
| 25,916,843,949
|
IssuesEvent
|
2022-12-15 18:05:29
|
conda/conda
|
https://api.github.com/repos/conda/conda
|
closed
|
CI tests for pull requests should run when they need to (i.e. not everytime a commit is pushed)
|
source::anaconda tag::performance sprint type::testing
|
### What is the idea?
I recently observed for the following pull requests:
- https://github.com/conda/conda/pull/12132
That all the tests were running in CI even though we were only editing the README file. This should not happen.
### Why is this needed?
To decrease load from test runners and to only use these resources when necessary (i.e. eco-friendly tests 😉 🌲 ).
### What should happen?
We need to add more complex rules to our GitHub action that preferably only run tests when files concerned with the software itself have been changed.
Excluded from this include but are not limited to the following:
- READMEs and other top level documentation (e.g. CONTRIBUTING too)
- Documentation; when this changes we should only run the applicable CI tests which is the documentation building/staging tool.
- Others?
### Tasks
- [x] Come up with a more reasonable way to configure GitHub actions to only run tests when necessary
- [x] https://github.com/conda/conda/pull/12141
- [x] https://github.com/conda/conda/pull/12147
- [x] Figure out what else needs to be excluded, if anything
- [x] Update branch protections' required checks to specify the "Tests / Analyze results" instead of individual jobs
- [x] Test this manually and partner up with a reviewer early to explain how your new changes work (this will be counted as part of the review process).
- [x] https://github.com/conda/conda/pull/12144 **Test succeeded!**
- [x] https://github.com/conda/conda/pull/12145
- [x] https://github.com/conda/conda/pull/12146
- [x] Port this over to conda-build
- [x] https://github.com/conda/conda-build/pull/4664
- [x] Use a GitHub Action instead to determine whether tests are being skipped
- [x] https://github.com/conda/conda/pull/12180
- [x] https://github.com/conda/conda-build/pull/4675
|
1.0
|
CI tests for pull requests should run when they need to (i.e. not everytime a commit is pushed) - ### What is the idea?
I recently observed for the following pull requests:
- https://github.com/conda/conda/pull/12132
That all the tests were running in CI even though we were only editing the README file. This should not happen.
### Why is this needed?
To decrease load from test runners and to only use these resources when necessary (i.e. eco-friendly tests 😉 🌲 ).
### What should happen?
We need to add more complex rules to our GitHub action that preferably only run tests when files concerned with the software itself have been changed.
Excluded from this include but are not limited to the following:
- READMEs and other top level documentation (e.g. CONTRIBUTING too)
- Documentation; when this changes we should only run the applicable CI tests which is the documentation building/staging tool.
- Others?
### Tasks
- [x] Come up with a more reasonable way to configure GitHub actions to only run tests when necessary
- [x] https://github.com/conda/conda/pull/12141
- [x] https://github.com/conda/conda/pull/12147
- [x] Figure out what else needs to be excluded, if anything
- [x] Update branch protections' required checks to specify the "Tests / Analyze results" instead of individual jobs
- [x] Test this manually and partner up with a reviewer early to explain how your new changes work (this will be counted as part of the review process).
- [x] https://github.com/conda/conda/pull/12144 **Test succeeded!**
- [x] https://github.com/conda/conda/pull/12145
- [x] https://github.com/conda/conda/pull/12146
- [x] Port this over to conda-build
- [x] https://github.com/conda/conda-build/pull/4664
- [x] Use a GitHub Action instead to determine whether tests are being skipped
- [x] https://github.com/conda/conda/pull/12180
- [x] https://github.com/conda/conda-build/pull/4675
|
test
|
ci tests for pull requests should run when they need to i e not everytime a commit is pushed what is the idea i recently observed for the following pull requests that all the tests were running in ci even though we were only editing the readme file this should not happen why is this needed to decrease load from test runners and to only use these resources when necessary i e eco friendly tests 😉 🌲 what should happen we need to add more complex rules to our github action that preferably only run tests when files concerned with the software itself have been changed excluded from this include but are not limited to the following readmes and other top level documentation e g contributing too documentation when this changes we should only run the applicable ci tests which is the documentation building staging tool others tasks come up with a more reasonable way to configure github actions to only run tests when necessary figure out what else needs to be excluded if anything update branch protections required checks to specify the tests analyze results instead of individual jobs test this manually and partner up with a reviewer early to explain how your new changes work this will be counted as part of the review process test succeeded port this over to conda build use a github action instead to determine whether tests are being skipped
| 1
|
236,928
| 7,753,678,052
|
IssuesEvent
|
2018-05-31 02:06:57
|
Gloirin/m2gTest
|
https://api.github.com/repos/Gloirin/m2gTest
|
closed
|
0006956:
define default value for boolean db fields
|
Felamimail high priority
|
**Reported by pschuele on 16 Aug 2012 12:45**
**Version:** Joey (2012.10.1~alpha1)
define default value for boolean db fields
**Additional information:** see https://gerrit.tine20.org/tine20/#/c/916
|
1.0
|
0006956:
define default value for boolean db fields - **Reported by pschuele on 16 Aug 2012 12:45**
**Version:** Joey (2012.10.1~alpha1)
define default value for boolean db fields
**Additional information:** see https://gerrit.tine20.org/tine20/#/c/916
|
non_test
|
define default value for boolean db fields reported by pschuele on aug version joey define default value for boolean db fields additional information see
| 0
|
133,316
| 10,816,983,395
|
IssuesEvent
|
2019-11-08 08:41:06
|
prestosql/presto
|
https://api.github.com/repos/prestosql/presto
|
opened
|
Selective tests execution
|
maintenance test
|
Currently we always run all the tests per each build. Either for master branch as well for each pull requests. This is great because it gives high confidence that change is working, however it is very time and resource consuming which becomes more important while Presto is getting bigger and bigger.
The idea here is to run less tests without sacrificing confidence that change is correct. In other words run only tests for components that are affected by the change. Like not running tests for Kudu connector when someone changes authorization code in Hive connector.
Today we are doing something to address that. Not running tests for CDH Hadoop platform when running build for pull requests. These tests are run only for master branch.
This issue is raised to start broader discussion.
Below there is the example implementation design details. The best when it would be possible to implement this only using Travis.
Let say they are:
- complete test build - that runs all tests (some way as it is done today)
- partial test build - that runs only tests for components affected by the change
- storage - that stores information about which git commit hashes passed successfully complete test build. Storage needs to be publically accessible. Maybe we could use private infra to mark successful builds, if that is not possible using Travis.
- component to test mapping - a map that points what tests have to be executed when testing given component. e.g. for presto-hive we would run all unit tests in presto-hive, hive integration tests in presto-hive-hadoop2/bin and set of product tests execution.
1. We still run complete build for each commit (merge) in master branch.
2. Whenever complete build passes successfully we store the git commit hash in the storage.
3. When running partial build (for pull request) we would check in the storage latest successful complete build git commit hash. Then we could check all the commits till HEAD since that git commit hash to find what components were changed. Having a list of changed components we need to find a list of affected components that would be list of changed components with all the components that are using them. e.g. when presto-hive is changed we need to test also presto-iceberg. Then using component to test mapping we could run only tests for components that are affected by the change.
|
1.0
|
Selective tests execution - Currently we always run all the tests per each build. Either for master branch as well for each pull requests. This is great because it gives high confidence that change is working, however it is very time and resource consuming which becomes more important while Presto is getting bigger and bigger.
The idea here is to run less tests without sacrificing confidence that change is correct. In other words run only tests for components that are affected by the change. Like not running tests for Kudu connector when someone changes authorization code in Hive connector.
Today we are doing something to address that. Not running tests for CDH Hadoop platform when running build for pull requests. These tests are run only for master branch.
This issue is raised to start broader discussion.
Below there is the example implementation design details. The best when it would be possible to implement this only using Travis.
Let say they are:
- complete test build - that runs all tests (some way as it is done today)
- partial test build - that runs only tests for components affected by the change
- storage - that stores information about which git commit hashes passed successfully complete test build. Storage needs to be publically accessible. Maybe we could use private infra to mark successful builds, if that is not possible using Travis.
- component to test mapping - a map that points what tests have to be executed when testing given component. e.g. for presto-hive we would run all unit tests in presto-hive, hive integration tests in presto-hive-hadoop2/bin and set of product tests execution.
1. We still run complete build for each commit (merge) in master branch.
2. Whenever complete build passes successfully we store the git commit hash in the storage.
3. When running partial build (for pull request) we would check in the storage latest successful complete build git commit hash. Then we could check all the commits till HEAD since that git commit hash to find what components were changed. Having a list of changed components we need to find a list of affected components that would be list of changed components with all the components that are using them. e.g. when presto-hive is changed we need to test also presto-iceberg. Then using component to test mapping we could run only tests for components that are affected by the change.
|
test
|
selective tests execution currently we always run all the tests per each build either for master branch as well for each pull requests this is great because it gives high confidence that change is working however it is very time and resource consuming which becomes more important while presto is getting bigger and bigger the idea here is to run less tests without sacrificing confidence that change is correct in other words run only tests for components that are affected by the change like not running tests for kudu connector when someone changes authorization code in hive connector today we are doing something to address that not running tests for cdh hadoop platform when running build for pull requests these tests are run only for master branch this issue is raised to start broader discussion below there is the example implementation design details the best when it would be possible to implement this only using travis let say they are complete test build that runs all tests some way as it is done today partial test build that runs only tests for components affected by the change storage that stores information about which git commit hashes passed successfully complete test build storage needs to be publically accessible maybe we could use private infra to mark successful builds if that is not possible using travis component to test mapping a map that points what tests have to be executed when testing given component e g for presto hive we would run all unit tests in presto hive hive integration tests in presto hive bin and set of product tests execution we still run complete build for each commit merge in master branch whenever complete build passes successfully we store the git commit hash in the storage when running partial build for pull request we would check in the storage latest successful complete build git commit hash then we could check all the commits till head since that git commit hash to find what components were changed having a list of changed components we need to find a list of affected components that would be list of changed components with all the components that are using them e g when presto hive is changed we need to test also presto iceberg then using component to test mapping we could run only tests for components that are affected by the change
| 1
|
108,735
| 9,329,731,008
|
IssuesEvent
|
2019-03-28 03:40:09
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
mimxrt1050_evk Fatal fault in thread tests/kernel/mem_protect/stackprot Fatal fault in thread
|
area: Testing area: Testing Suite area: Tests bug platform: NXP priority: medium
|
**Describe the bug**
kernel mem_protect stack prot
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=mimxrt1050_evk ..
3. make
4. See error
**Expected behavior**
test PASS
**Impact**
What impact does this issue have on your progress (e.g., annoyance, showstopper)
**Screenshots or console output**
```
***** Booting Zephyr OS v1.14.0-rc1-1369-gf30a1c03a260 *****
Starts alternate_thread
alternate_thread: Input string is too long and stack overflowed!
***** Stack Check Fail! *****
Current thread ID = 0x80000068
Faulting instruction address = 0x6000ae44
Fatal fault in thread 0x80000068! Aborting.
Running test suite stackprot
===================================================================
starting test - 1?:抋?x?jW焰}屠.
鳃?
糃?栳?b愞w唼?盛畈c诽F嗕籈vN?媮+8D3?物?孯PH!8y"t3?%g榥[S€<D哥\硉s赼.严/r?贿pI$畴f揗谞偤&ps稦?耋?dI虸╟绕€橛尌z劬??夜∪{?忟W博燘撴?畛?崗D硎:?o??燘 侤?
*"$[8F?L>縞嶊<肌庛
[a?娔?!?t鯜駛诋兾阺x瑆饡怙b???惞MG)圩)fⅵ訒€r$kx櫚倒E5爵A孉~氼6K簳瓞廞愯u旑J爥0?
***** MPU FAULT *****
Instruction Access Violation
***** Hardware exception *****
Current thread ID = 0x800001ec
Faulting instruction address = 0xe9e8
Fatal fault in thread 0x800001ec! Aborting.
```
**Environment (please complete the following information):**
- OS: (Linux )
- Toolchain ( Zephyr SDK)
- Commit SHA or Version used: v1.14.0_rc2
|
3.0
|
mimxrt1050_evk Fatal fault in thread tests/kernel/mem_protect/stackprot Fatal fault in thread - **Describe the bug**
kernel mem_protect stack prot
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=mimxrt1050_evk ..
3. make
4. See error
**Expected behavior**
test PASS
**Impact**
What impact does this issue have on your progress (e.g., annoyance, showstopper)
**Screenshots or console output**
```
***** Booting Zephyr OS v1.14.0-rc1-1369-gf30a1c03a260 *****
Starts alternate_thread
alternate_thread: Input string is too long and stack overflowed!
***** Stack Check Fail! *****
Current thread ID = 0x80000068
Faulting instruction address = 0x6000ae44
Fatal fault in thread 0x80000068! Aborting.
Running test suite stackprot
===================================================================
starting test - 1?:抋?x?jW焰}屠.
鳃?
糃?栳?b愞w唼?盛畈c诽F嗕籈vN?媮+8D3?物?孯PH!8y"t3?%g榥[S€<D哥\硉s赼.严/r?贿pI$畴f揗谞偤&ps稦?耋?dI虸╟绕€橛尌z劬??夜∪{?忟W博燘撴?畛?崗D硎:?o??燘 侤?
*"$[8F?L>縞嶊<肌庛
[a?娔?!?t鯜駛诋兾阺x瑆饡怙b???惞MG)圩)fⅵ訒€r$kx櫚倒E5爵A孉~氼6K簳瓞廞愯u旑J爥0?
***** MPU FAULT *****
Instruction Access Violation
***** Hardware exception *****
Current thread ID = 0x800001ec
Faulting instruction address = 0xe9e8
Fatal fault in thread 0x800001ec! Aborting.
```
**Environment (please complete the following information):**
- OS: (Linux )
- Toolchain ( Zephyr SDK)
- Commit SHA or Version used: v1.14.0_rc2
|
test
|
evk fatal fault in thread tests kernel mem protect stackprot fatal fault in thread describe the bug kernel mem protect stack prot to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard evk make see error expected behavior test pass impact what impact does this issue have on your progress e g annoyance showstopper screenshots or console output booting zephyr os starts alternate thread alternate thread input string is too long and stack overflowed stack check fail current thread id faulting instruction address fatal fault in thread aborting running test suite stackprot starting test 抋 x jw焰 屠 鳃 糃 栳 b愞w唼 盛畈c诽f嗕籈vn 媮 物 孯ph g榥 s€ d哥 硉s赼 严 r 贿pi 畴f揗谞偤 ps稦 耋 di虸╟绕€橛尌z劬 夜∪ 忟w博燘撴 畛 崗d硎 o 燘 侤 f l 縞嶊 肌庛 a 娔 t鯜駛诋兾阺x瑆饡怙b 惞mg 圩 fⅵ訒€r kx櫚倒e 氼 簳瓞 mpu fault instruction access violation hardware exception current thread id faulting instruction address fatal fault in thread aborting environment please complete the following information os linux toolchain zephyr sdk commit sha or version used
| 1
|
60,961
| 25,319,146,758
|
IssuesEvent
|
2022-11-18 01:15:05
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
Per tenant reauth is broken
|
:gear: identity service
|
Broken due to TS identity client not respecting the existence of the error store in the newest identity service.
|
1.0
|
Per tenant reauth is broken - Broken due to TS identity client not respecting the existence of the error store in the newest identity service.
|
non_test
|
per tenant reauth is broken broken due to ts identity client not respecting the existence of the error store in the newest identity service
| 0
|
108,968
| 16,825,780,628
|
IssuesEvent
|
2021-06-17 18:21:53
|
gravitational/teleport
|
https://api.github.com/repos/gravitational/teleport
|
closed
|
Roles should not inadvertently grant access
|
c-ca c-fg c-gl security
|
### Description
**What happened**:
When creating a role that only specifies node labels, application and kubernetes access are fully granted.
Take this short example role:
```yaml
version: v3
kind: role
metadata:
name: qa
spec:
allow:
node_labels:
'account_type': 'QA'
```
When this is created via `tctl create -f qa.yaml`, it results in the following `tctl get roles/qa`:
```yaml
kind: role
metadata:
id: 1612814108898420391
name: qa
spec:
allow:
app_labels:
'*': '*'
kubernetes_labels:
'*': '*'
node_labels:
account_type: QA
deny: {}
options:
cert_format: standard
enhanced_recording:
- command
- network
forward_agent: false
max_session_ttl: 30h0m0s
port_forwarding: true
version: v3
```
Since there was no initial mention of the app nor kube services, this could lead to a situation where a user with this QA role will be allowed to access apps/kube clusters when it was only intended for her to get access to nodes with the `account_type: QA` label.
**What you expected to happen**:
Teleport should follow the concept of least privileged access. no access should be granted to kubernetes clusters nor applications.
**How to reproduce it (as minimally and precisely as possible)**:
Copy/paste the example yaml given above and run the commands above.
### Environment
Environment 1
- Teleport version (use `teleport version`): `Teleport Enterprise v5.0.0 git:v5.0.0-0-gac4971801 go1.15.5`
- Tsh version (use `tsh version`): `Teleport v5.0.0 git:v5.0.0-0-gac4971801 go1.15.5`
- OS (e.g. from `/etc/os-release`): Amazon Linux 2
Environment 2
- Teleport version (use `teleport version`): `Teleport Enterprise v5.1.2 git:v5.1.2-0-g822d10b44 go1.15.5`
- Tsh version (use `tsh version`): `Teleport v5.1.2 git:v5.1.2-0-g822d10b44 go1.15.5`
- OS (e.g. from `/etc/os-release`): quay.io/gravitational/teleport-ent:5.1.2 container image on EKS
- Where are you running Teleport? (e.g. AWS, GCP, Dedicated Hardware):
AWS
**Browser environment**
N/A
**Relevant Debug Logs If Applicable**
N/A
|
True
|
Roles should not inadvertently grant access - ### Description
**What happened**:
When creating a role that only specifies node labels, application and kubernetes access are fully granted.
Take this short example role:
```yaml
version: v3
kind: role
metadata:
name: qa
spec:
allow:
node_labels:
'account_type': 'QA'
```
When this is created via `tctl create -f qa.yaml`, it results in the following `tctl get roles/qa`:
```yaml
kind: role
metadata:
id: 1612814108898420391
name: qa
spec:
allow:
app_labels:
'*': '*'
kubernetes_labels:
'*': '*'
node_labels:
account_type: QA
deny: {}
options:
cert_format: standard
enhanced_recording:
- command
- network
forward_agent: false
max_session_ttl: 30h0m0s
port_forwarding: true
version: v3
```
Since there was no initial mention of the app nor kube services, this could lead to a situation where a user with this QA role will be allowed to access apps/kube clusters when it was only intended for her to get access to nodes with the `account_type: QA` label.
**What you expected to happen**:
Teleport should follow the concept of least privileged access. no access should be granted to kubernetes clusters nor applications.
**How to reproduce it (as minimally and precisely as possible)**:
Copy/paste the example yaml given above and run the commands above.
### Environment
Environment 1
- Teleport version (use `teleport version`): `Teleport Enterprise v5.0.0 git:v5.0.0-0-gac4971801 go1.15.5`
- Tsh version (use `tsh version`): `Teleport v5.0.0 git:v5.0.0-0-gac4971801 go1.15.5`
- OS (e.g. from `/etc/os-release`): Amazon Linux 2
Environment 2
- Teleport version (use `teleport version`): `Teleport Enterprise v5.1.2 git:v5.1.2-0-g822d10b44 go1.15.5`
- Tsh version (use `tsh version`): `Teleport v5.1.2 git:v5.1.2-0-g822d10b44 go1.15.5`
- OS (e.g. from `/etc/os-release`): quay.io/gravitational/teleport-ent:5.1.2 container image on EKS
- Where are you running Teleport? (e.g. AWS, GCP, Dedicated Hardware):
AWS
**Browser environment**
N/A
**Relevant Debug Logs If Applicable**
N/A
|
non_test
|
roles should not inadvertently grant access description what happened when creating a role that only specifies node labels application and kubernetes access are fully granted take this short example role yaml version kind role metadata name qa spec allow node labels account type qa when this is created via tctl create f qa yaml it results in the following tctl get roles qa yaml kind role metadata id name qa spec allow app labels kubernetes labels node labels account type qa deny options cert format standard enhanced recording command network forward agent false max session ttl port forwarding true version since there was no initial mention of the app nor kube services this could lead to a situation where a user with this qa role will be allowed to access apps kube clusters when it was only intended for her to get access to nodes with the account type qa label what you expected to happen teleport should follow the concept of least privileged access no access should be granted to kubernetes clusters nor applications how to reproduce it as minimally and precisely as possible copy paste the example yaml given above and run the commands above environment environment teleport version use teleport version teleport enterprise git tsh version use tsh version teleport git os e g from etc os release amazon linux environment teleport version use teleport version teleport enterprise git tsh version use tsh version teleport git os e g from etc os release quay io gravitational teleport ent container image on eks where are you running teleport e g aws gcp dedicated hardware aws browser environment n a relevant debug logs if applicable n a
| 0
|
188,107
| 14,438,981,659
|
IssuesEvent
|
2020-12-07 13:48:28
|
kalexmills/github-vet-tests-dec2020
|
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
|
closed
|
wrapp/go-pcre: regexp/find_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [wrapp/go-pcre](https://www.github.com/wrapp/go-pcre) at [regexp/find_test.go](https://github.com/wrapp/go-pcre/blob/14dac038b5c4e894d769709008268e1968b0969b/regexp/find_test.go#L210-L212)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 211 may store a reference to test
[Click here to see the code in its original context.](https://github.com/wrapp/go-pcre/blob/14dac038b5c4e894d769709008268e1968b0969b/regexp/find_test.go#L210-L212)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range findTests {
testFindIndex(&test, MustCompile(test.pat).FindIndex([]byte(test.text)), t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 14dac038b5c4e894d769709008268e1968b0969b
|
1.0
|
wrapp/go-pcre: regexp/find_test.go; 3 LoC -
Found a possible issue in [wrapp/go-pcre](https://www.github.com/wrapp/go-pcre) at [regexp/find_test.go](https://github.com/wrapp/go-pcre/blob/14dac038b5c4e894d769709008268e1968b0969b/regexp/find_test.go#L210-L212)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 211 may store a reference to test
[Click here to see the code in its original context.](https://github.com/wrapp/go-pcre/blob/14dac038b5c4e894d769709008268e1968b0969b/regexp/find_test.go#L210-L212)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range findTests {
testFindIndex(&test, MustCompile(test.pat).FindIndex([]byte(test.text)), t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 14dac038b5c4e894d769709008268e1968b0969b
|
test
|
wrapp go pcre regexp find test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call at line may store a reference to test click here to show the line s of go which triggered the analyzer go for test range findtests testfindindex test mustcompile test pat findindex byte test text t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
94,604
| 8,506,594,457
|
IssuesEvent
|
2018-10-30 16:56:04
|
poanetwork/hbbft
|
https://api.github.com/repos/poanetwork/hbbft
|
closed
|
Test that the reordering attack on `BinaryAgreement` no longer succeeds
|
good first issue test
|
The reordering attack described in https://github.com/amiller/HoneyBadgerBFT/issues/59#issue-310368284 was avoided by adding `Conf` messages in #44, see also #37. We need a test that replays the attack and verifies that the attack does not succeed.
Closes #37.
Closes #45.
|
1.0
|
Test that the reordering attack on `BinaryAgreement` no longer succeeds - The reordering attack described in https://github.com/amiller/HoneyBadgerBFT/issues/59#issue-310368284 was avoided by adding `Conf` messages in #44, see also #37. We need a test that replays the attack and verifies that the attack does not succeed.
Closes #37.
Closes #45.
|
test
|
test that the reordering attack on binaryagreement no longer succeeds the reordering attack described in was avoided by adding conf messages in see also we need a test that replays the attack and verifies that the attack does not succeed closes closes
| 1
|
165,331
| 13,998,877,004
|
IssuesEvent
|
2020-10-28 10:03:14
|
getsentry/sentry
|
https://api.github.com/repos/getsentry/sentry
|
closed
|
Incorrect inferred title from native callstack
|
Component: Documentation Component: Issues On hold
|
## Important Details
How are you running Sentry?
Saas (sentry.io)
## Description
I'm using the store API to write custom callstack based events to Sentry, however I'm not getting the expected title for the issue. I expect the name would be based on the top most stack frame that belongs to my application, however I always get the top most stack frame regardless. I use the custom grouping enhancements to tell Sentry which stack frames belong to my application and which ones don't.
## Steps to Reproduce
1. Using the store endpoint send the following Event:
```json
{
"platform": "native",
"level": "fatal",
"sentry.interfaces.Exception": {
"values": [
{
"type": "Crash",
"stacktrace": {
"frames": [
{
"function": "Thread_Start",
"package": "System"
},
{
"function": "Mid_Callstack",
"package": "MyApp"
},
{
"function": "Crashing_Function",
"package": "MyApp"
},
{
"function": "DONT_WANT_Exception_Handler",
"package": "System"
}
]
}
}
]
}
}
```
2. Setup a server side custom grouping enhancement like so:
`stack.package:MyApp +app`
3. Check the created issue in Sentry, note the title is "Crash DONT_WANT_Exception_Handler", what I expect is "Crash Crashing_Function".
4. Check the callstack in the issue, Sentry has correctly identified the stack frames from "MyApp" are part of the application and the "System" frames are not.
### Possible Solution
I've done some brief analysis of the Sentry source code, first time so bare with me. It looks like the intended behavior is for the title to be fed via the "culprit" tag. I can see the function `get_culprit` called via `_pull_out_data` from `event_manager.py` takes care of that correctly walking the stack and checking which frames are `in_app`. However the function that applies the server enhancement and ultimately sets `in_app` for the event is called from `normalize_stacktraces_for_grouping`. As far as I can tell these two actions happen around the wrong way, first derive the culprit based on `in_app` then apply the server settings that set `in_app`. I'll leave the solution for someone more familiar with the code, but that's seems to be the core issue.
As for solving my issue, I have the ability to process the stack frames before being sent to Sentry, so I can manually apply the `in_app` tags using my code rather than the server settings. This gives me the desired result, however the above took a bunch of time so I want to leave my notes / findings here for others and potentially for someone to fix.
|
1.0
|
Incorrect inferred title from native callstack - ## Important Details
How are you running Sentry?
Saas (sentry.io)
## Description
I'm using the store API to write custom callstack based events to Sentry, however I'm not getting the expected title for the issue. I expect the name would be based on the top most stack frame that belongs to my application, however I always get the top most stack frame regardless. I use the custom grouping enhancements to tell Sentry which stack frames belong to my application and which ones don't.
## Steps to Reproduce
1. Using the store endpoint send the following Event:
```json
{
"platform": "native",
"level": "fatal",
"sentry.interfaces.Exception": {
"values": [
{
"type": "Crash",
"stacktrace": {
"frames": [
{
"function": "Thread_Start",
"package": "System"
},
{
"function": "Mid_Callstack",
"package": "MyApp"
},
{
"function": "Crashing_Function",
"package": "MyApp"
},
{
"function": "DONT_WANT_Exception_Handler",
"package": "System"
}
]
}
}
]
}
}
```
2. Setup a server side custom grouping enhancement like so:
`stack.package:MyApp +app`
3. Check the created issue in Sentry, note the title is "Crash DONT_WANT_Exception_Handler", what I expect is "Crash Crashing_Function".
4. Check the callstack in the issue, Sentry has correctly identified the stack frames from "MyApp" are part of the application and the "System" frames are not.
### Possible Solution
I've done some brief analysis of the Sentry source code, first time so bare with me. It looks like the intended behavior is for the title to be fed via the "culprit" tag. I can see the function `get_culprit` called via `_pull_out_data` from `event_manager.py` takes care of that correctly walking the stack and checking which frames are `in_app`. However the function that applies the server enhancement and ultimately sets `in_app` for the event is called from `normalize_stacktraces_for_grouping`. As far as I can tell these two actions happen around the wrong way, first derive the culprit based on `in_app` then apply the server settings that set `in_app`. I'll leave the solution for someone more familiar with the code, but that's seems to be the core issue.
As for solving my issue, I have the ability to process the stack frames before being sent to Sentry, so I can manually apply the `in_app` tags using my code rather than the server settings. This gives me the desired result, however the above took a bunch of time so I want to leave my notes / findings here for others and potentially for someone to fix.
|
non_test
|
incorrect inferred title from native callstack important details how are you running sentry saas sentry io description i m using the store api to write custom callstack based events to sentry however i m not getting the expected title for the issue i expect the name would be based on the top most stack frame that belongs to my application however i always get the top most stack frame regardless i use the custom grouping enhancements to tell sentry which stack frames belong to my application and which ones don t steps to reproduce using the store endpoint send the following event json platform native level fatal sentry interfaces exception values type crash stacktrace frames function thread start package system function mid callstack package myapp function crashing function package myapp function dont want exception handler package system setup a server side custom grouping enhancement like so stack package myapp app check the created issue in sentry note the title is crash dont want exception handler what i expect is crash crashing function check the callstack in the issue sentry has correctly identified the stack frames from myapp are part of the application and the system frames are not possible solution i ve done some brief analysis of the sentry source code first time so bare with me it looks like the intended behavior is for the title to be fed via the culprit tag i can see the function get culprit called via pull out data from event manager py takes care of that correctly walking the stack and checking which frames are in app however the function that applies the server enhancement and ultimately sets in app for the event is called from normalize stacktraces for grouping as far as i can tell these two actions happen around the wrong way first derive the culprit based on in app then apply the server settings that set in app i ll leave the solution for someone more familiar with the code but that s seems to be the core issue as for solving my issue i have the ability to process the stack frames before being sent to sentry so i can manually apply the in app tags using my code rather than the server settings this gives me the desired result however the above took a bunch of time so i want to leave my notes findings here for others and potentially for someone to fix
| 0
|
2,232
| 3,353,054,434
|
IssuesEvent
|
2015-11-18 02:27:09
|
mapbox/gl-draw
|
https://api.github.com/repos/mapbox/gl-draw
|
closed
|
Dont render features in feature collection individually
|
performance
|
The code is cleaner as is but this is very bad when there there are a lot of features (even a few thousand breaks it)
https://github.com/mapbox/gl-draw/blob/master/src/draw.js#L441-L469
|
True
|
Dont render features in feature collection individually - The code is cleaner as is but this is very bad when there there are a lot of features (even a few thousand breaks it)
https://github.com/mapbox/gl-draw/blob/master/src/draw.js#L441-L469
|
non_test
|
dont render features in feature collection individually the code is cleaner as is but this is very bad when there there are a lot of features even a few thousand breaks it
| 0
|
430,310
| 30,175,280,442
|
IssuesEvent
|
2023-07-04 03:36:45
|
prodigytech-doc/api-docs
|
https://api.github.com/repos/prodigytech-doc/api-docs
|
opened
|
gravityScale中的Precautions注释字段应该放在set中。
|
documentation
|
### 所属页面地址
https://api-docs.ark.online/classes/Gameplay.CharacterBase.html#forceupdatemovement
### 选中内容
Precautions
范围0~10, 过大和过小的值都会被限制
### 请选择反馈的问题类型
缺少信息 - 我想了解更多信息
### 补充说明
_No response_
|
1.0
|
gravityScale中的Precautions注释字段应该放在set中。 - ### 所属页面地址
https://api-docs.ark.online/classes/Gameplay.CharacterBase.html#forceupdatemovement
### 选中内容
Precautions
范围0~10, 过大和过小的值都会被限制
### 请选择反馈的问题类型
缺少信息 - 我想了解更多信息
### 补充说明
_No response_
|
non_test
|
gravityscale中的precautions注释字段应该放在set中。 所属页面地址 选中内容 precautions 过大和过小的值都会被限制 请选择反馈的问题类型 缺少信息 我想了解更多信息 补充说明 no response
| 0
|
15,523
| 8,952,372,108
|
IssuesEvent
|
2019-01-25 16:21:53
|
labofoz/labofoz.com
|
https://api.github.com/repos/labofoz/labofoz.com
|
opened
|
Use a barcode reader to do Voice to Barcode to type into a computer w/o keyboard
|
🦄 performance
|
Let's say you didn't have a keyboard, but had a mobile device and barcode reader. Let's make it so that we can type into that computer with just your voice!
- [ ] Voice to text
- [ ] Text to barcode
- [ ] Point barcode reader to phone which translates voice to barcode
|
True
|
Use a barcode reader to do Voice to Barcode to type into a computer w/o keyboard - Let's say you didn't have a keyboard, but had a mobile device and barcode reader. Let's make it so that we can type into that computer with just your voice!
- [ ] Voice to text
- [ ] Text to barcode
- [ ] Point barcode reader to phone which translates voice to barcode
|
non_test
|
use a barcode reader to do voice to barcode to type into a computer w o keyboard let s say you didn t have a keyboard but had a mobile device and barcode reader let s make it so that we can type into that computer with just your voice voice to text text to barcode point barcode reader to phone which translates voice to barcode
| 0
|
201,929
| 15,229,578,125
|
IssuesEvent
|
2021-02-18 13:05:55
|
WeiXian042901/fyp_repository
|
https://api.github.com/repos/WeiXian042901/fyp_repository
|
opened
|
QA_007-Account management create button
|
Acceptance Test Admin
|
**Test Case**
- Check if the create button will direct the user to the Registration page for admins where they can also choose the role for the account created.
**Pre-Conditions**
-User must be logged in
-User must be on an admin account
-User must be on the Account Management page
**Test Steps**
1. Click on the “Create Account” button.
**Test Data**
- n/a
**Expected Results**
- The user is directed to the Registration page.
**Actual Results**
- The user is directed to the Registration page.
**Pass/Fail**
-Pass
**Tested by**
- Jun Liang
|
1.0
|
QA_007-Account management create button - **Test Case**
- Check if the create button will direct the user to the Registration page for admins where they can also choose the role for the account created.
**Pre-Conditions**
-User must be logged in
-User must be on an admin account
-User must be on the Account Management page
**Test Steps**
1. Click on the “Create Account” button.
**Test Data**
- n/a
**Expected Results**
- The user is directed to the Registration page.
**Actual Results**
- The user is directed to the Registration page.
**Pass/Fail**
-Pass
**Tested by**
- Jun Liang
|
test
|
qa account management create button test case check if the create button will direct the user to the registration page for admins where they can also choose the role for the account created pre conditions user must be logged in user must be on an admin account user must be on the account management page test steps click on the “create account” button test data n a expected results the user is directed to the registration page actual results the user is directed to the registration page pass fail pass tested by jun liang
| 1
|
567,399
| 16,857,818,419
|
IssuesEvent
|
2021-06-21 09:06:30
|
snowplow/snowplow-objc-tracker
|
https://api.github.com/repos/snowplow/snowplow-objc-tracker
|
closed
|
Bump FMDB to version 2.7
|
priority:medium status:completed type:enhancement
|
As requested on Discourse: https://discourse.snowplowanalytics.com/t/objectivec-tracker-podspec-fmdb-dependency/5236
FMDB version 2.6 has iOS 8.0 as minimum supported version, which is unsupported by Xcode 12 and causes a warning when publishing with Cocoapods.
FMDB version 2.7 supports iOS 9.0 like our iOS tracker v2.
|
1.0
|
Bump FMDB to version 2.7 - As requested on Discourse: https://discourse.snowplowanalytics.com/t/objectivec-tracker-podspec-fmdb-dependency/5236
FMDB version 2.6 has iOS 8.0 as minimum supported version, which is unsupported by Xcode 12 and causes a warning when publishing with Cocoapods.
FMDB version 2.7 supports iOS 9.0 like our iOS tracker v2.
|
non_test
|
bump fmdb to version as requested on discourse fmdb version has ios as minimum supported version which is unsupported by xcode and causes a warning when publishing with cocoapods fmdb version supports ios like our ios tracker
| 0
|
155,693
| 24,504,470,284
|
IssuesEvent
|
2022-10-10 15:13:09
|
carbon-design-system/carbon-for-ibm-dotcom
|
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
|
closed
|
[Figma] Card - Static
|
design design kit stretch project
|
Add the Card - static component to Figma kit. Remember we are only building variants for `xlg`, `md`, and `sm`.
[Design specs](https://ibm.ent.box.com/file/911782495148?s=hyuqbkgi43cij4reru3djh7ztmwp9vlp)
[Website](https://www.ibm.com/standards/carbon/components/cards#card)
**Pre work:** We need a way to swap icon for Link with icon. This might instead be built into the `Card - basic` component.
**Must haves:**
- [ ] Create a flexible base for the `Card - static` component, or use the `Card - basic` base.
- [ ] Should use `Link with icon` component
- [ ] Include variants with image, tags, eyebrow, and body copy. Note that these are all optional, so can appear with different options. Use variants to give designers the ability to toggle what is visible. Use auto layout to automatically set the correct padding.
- [ ] The [Storybook](https://www.ibm.com/standards/carbon/web-components/?path=/story/components-card--static) has an option for `Outlined card` by default, which can be disabled to bring in a gray background. That should be a variant option.
- [ ] Designers should be able to resize the width and have the content condense. Check the specs and use auto layout to handle this.
- [ ] Only the CTA is interactive - make sure the `Card - static` component itself does not have interactive states.

**Resources**
- C4IBM.com Component checklist is available in the Read Me 👀 page of the Figma library
- [Bases demo](https://secure.video.ibm.com/channel/23570833/playlist/641336/video/131117161) from the Figma Fridays (Episode 23)
- Reference the [best practices](https://www.figma.com/file/8n9IJtHnICSydODoYl4ueW/Figma-Best-Practices?node-id=52%3A6706) from Figma guild
- Info on [auto layout](https://www.figma.com/proto/RQhu2NWQVr1JDaU4mpRuZh/Figma-Exercises?node-id=422%3A2366&scaling=min-zoom&page-id=52%3A6706&starting-point-node-id=422%3A2366)
|
2.0
|
[Figma] Card - Static - Add the Card - static component to Figma kit. Remember we are only building variants for `xlg`, `md`, and `sm`.
[Design specs](https://ibm.ent.box.com/file/911782495148?s=hyuqbkgi43cij4reru3djh7ztmwp9vlp)
[Website](https://www.ibm.com/standards/carbon/components/cards#card)
**Pre work:** We need a way to swap icon for Link with icon. This might instead be built into the `Card - basic` component.
**Must haves:**
- [ ] Create a flexible base for the `Card - static` component, or use the `Card - basic` base.
- [ ] Should use `Link with icon` component
- [ ] Include variants with image, tags, eyebrow, and body copy. Note that these are all optional, so can appear with different options. Use variants to give designers the ability to toggle what is visible. Use auto layout to automatically set the correct padding.
- [ ] The [Storybook](https://www.ibm.com/standards/carbon/web-components/?path=/story/components-card--static) has an option for `Outlined card` by default, which can be disabled to bring in a gray background. That should be a variant option.
- [ ] Designers should be able to resize the width and have the content condense. Check the specs and use auto layout to handle this.
- [ ] Only the CTA is interactive - make sure the `Card - static` component itself does not have interactive states.

**Resources**
- C4IBM.com Component checklist is available in the Read Me 👀 page of the Figma library
- [Bases demo](https://secure.video.ibm.com/channel/23570833/playlist/641336/video/131117161) from the Figma Fridays (Episode 23)
- Reference the [best practices](https://www.figma.com/file/8n9IJtHnICSydODoYl4ueW/Figma-Best-Practices?node-id=52%3A6706) from Figma guild
- Info on [auto layout](https://www.figma.com/proto/RQhu2NWQVr1JDaU4mpRuZh/Figma-Exercises?node-id=422%3A2366&scaling=min-zoom&page-id=52%3A6706&starting-point-node-id=422%3A2366)
|
non_test
|
card static add the card static component to figma kit remember we are only building variants for xlg md and sm pre work we need a way to swap icon for link with icon this might instead be built into the card basic component must haves create a flexible base for the card static component or use the card basic base should use link with icon component include variants with image tags eyebrow and body copy note that these are all optional so can appear with different options use variants to give designers the ability to toggle what is visible use auto layout to automatically set the correct padding the has an option for outlined card by default which can be disabled to bring in a gray background that should be a variant option designers should be able to resize the width and have the content condense check the specs and use auto layout to handle this only the cta is interactive make sure the card static component itself does not have interactive states resources com component checklist is available in the read me 👀 page of the figma library from the figma fridays episode reference the from figma guild info on
| 0
|
323,874
| 27,755,856,643
|
IssuesEvent
|
2023-03-16 02:22:18
|
longhorn/longhorn
|
https://api.github.com/repos/longhorn/longhorn
|
opened
|
[TEST] Add an option to run the integration test with controlled network spec
|
kind/test
|
## What's the test to develop? Please describe
We can introduce the network controlling parameter in our testing pipeline, so that we can run our test with desired network spec like 1gps.
We can insert latency in the nodes where the tests are running using test like tc.
|
1.0
|
[TEST] Add an option to run the integration test with controlled network spec - ## What's the test to develop? Please describe
We can introduce the network controlling parameter in our testing pipeline, so that we can run our test with desired network spec like 1gps.
We can insert latency in the nodes where the tests are running using test like tc.
|
test
|
add an option to run the integration test with controlled network spec what s the test to develop please describe we can introduce the network controlling parameter in our testing pipeline so that we can run our test with desired network spec like we can insert latency in the nodes where the tests are running using test like tc
| 1
|
299,089
| 25,880,119,691
|
IssuesEvent
|
2022-12-14 10:40:29
|
02Gqbriel/m150-osCommerce
|
https://api.github.com/repos/02Gqbriel/m150-osCommerce
|
closed
|
2 - Datenbank Migration Testen
|
testing
|
Die Migration testen und Testergebnisse protokolieren.
Aufwand: 1h
Vorgänger: 1
|
1.0
|
2 - Datenbank Migration Testen - Die Migration testen und Testergebnisse protokolieren.
Aufwand: 1h
Vorgänger: 1
|
test
|
datenbank migration testen die migration testen und testergebnisse protokolieren aufwand vorgänger
| 1
|
303,694
| 26,223,905,935
|
IssuesEvent
|
2023-01-04 16:55:41
|
rotki/rotki
|
https://api.github.com/repos/rotki/rotki
|
closed
|
Make caching of cypress and pnpm cache consistent
|
tests
|
## Abstract
```
The cypress npm package is installed, but the Cypress binary is missing.
We expected the binary to be installed here: /home/runner/.cache/Cypress/12.1.0/Cypress/Cypress
Reasons it may be missing:
- You're caching 'node_modules' but are not caching this path: /home/runner/.cache/Cypress
- You ran 'npm install' at an earlier build step but did not persist: /home/runner/.cache/Cypress
Properly caching the binary will fix this error and avoid downloading and unzipping Cypress.
```
## Motivation
With `action/setup-node` we are using it's own cache support.
With Cypress we use caching some times.
There are caches where what is restored from action/setup-node considers cypress installed, but
~/.cache/Cypress is not properly restored
|
1.0
|
Make caching of cypress and pnpm cache consistent - ## Abstract
```
The cypress npm package is installed, but the Cypress binary is missing.
We expected the binary to be installed here: /home/runner/.cache/Cypress/12.1.0/Cypress/Cypress
Reasons it may be missing:
- You're caching 'node_modules' but are not caching this path: /home/runner/.cache/Cypress
- You ran 'npm install' at an earlier build step but did not persist: /home/runner/.cache/Cypress
Properly caching the binary will fix this error and avoid downloading and unzipping Cypress.
```
## Motivation
With `action/setup-node` we are using it's own cache support.
With Cypress we use caching some times.
There are caches where what is restored from action/setup-node considers cypress installed, but
~/.cache/Cypress is not properly restored
|
test
|
make caching of cypress and pnpm cache consistent abstract the cypress npm package is installed but the cypress binary is missing we expected the binary to be installed here home runner cache cypress cypress cypress reasons it may be missing you re caching node modules but are not caching this path home runner cache cypress you ran npm install at an earlier build step but did not persist home runner cache cypress properly caching the binary will fix this error and avoid downloading and unzipping cypress motivation with action setup node we are using it s own cache support with cypress we use caching some times there are caches where what is restored from action setup node considers cypress installed but cache cypress is not properly restored
| 1
|
19,301
| 3,757,910,942
|
IssuesEvent
|
2016-03-14 04:56:22
|
NishantUpadhyay-BTC/BLISS-Issue-Tracking
|
https://api.github.com/repos/NishantUpadhyay-BTC/BLISS-Issue-Tracking
|
closed
|
Guest UI: Availability: Availability Table: Static Headers
|
bug Deployed to Test
|
The headers of the Availability table (lodging types) should be static - should not scroll with the dates.
|
1.0
|
Guest UI: Availability: Availability Table: Static Headers - The headers of the Availability table (lodging types) should be static - should not scroll with the dates.
|
test
|
guest ui availability availability table static headers the headers of the availability table lodging types should be static should not scroll with the dates
| 1
|
177,372
| 14,626,576,109
|
IssuesEvent
|
2020-12-23 10:36:36
|
poikilos/artspatter
|
https://api.github.com/repos/poikilos/artspatter
|
opened
|
Add past install steps to development documentation.
|
documentation
|
from old install.bat:
```batch
@echo off
REM node
REM Most lines are commented since you only have to run them once to
REM create the workspace (such as downloading dependencies to
REM node_modules directories)
cd ..
REM SET PATH=%PATH%;%CD%/node-v14.15.0-win-x64
cd artspatter
echo PATH is %PATH%
echo Running npm in "%CD%"...
SET PACKAGER_RUNALL=npm-run-all
SET PACKAGER=%CD%/../node-v14.15.0-win-x64/npm.cmd
REM SET PACKAGER=npm
REM ^ Requires adding the %CD%/../node-v14.15.0-win-x64 to the system PATH,
REM since create-react-app also tries to call npm.
%PACKAGER% install
REM ^ Download all packages for server.
REM %PACKAGER% init
REM %PACKAGER% install express
REM %PACKAGER% install dotenv
REM %PACKAGER% install dotenv
REM %PACKAGER% install mongoose
SET CLIENT_APP=client
REM Don't do: %CD%/../node-v14.15.0-win-x64/npx.cmd create-react-app %CLIENT_APP%
REM You need the serviceWorker:
REM md client
REM cd client
REM npx create-react-app %CLIENT_APP% --template cra-template-pwa
REM ^ See https://github.com/facebook/create-react-app/issues/10032
REM %PACKAGER% install concurrently --save-dev
REM %PACKAGER% install nodemon --save-dev
REM ^ --save-dev dependencies aren't automatically installed with
REM "npm.cmd install" on the consumer (server owner)'s machine.
REM %PACKAGER% install
REM %PACKAGER% install
REM %PACKAGER% install mongoose-sequence
REM %PACKAGER% install eslint --save-dev
REM ^ Do NOT install eslint in the react client (see issue #13)
REM %PACKAGER% install body-parser
REM %PACKAGER% install express mongoose body-parser cors jsonwebtoken bcryptjs
REM ^ This is for authentication (BezKoder, 2019).
REM %PACKAGER% install mocha
REM %PACKAGER% install multer
REM - installed and removed fs--I don't need io, just file operations.
REM %PACKAGER% install nanoid
REM %PACKAGER% install node-thumbnail
echo CLIENT
echo The following packages only apply to the React.js user-facing page.
REM IF NOT EXIST client md client
REM IF NOT EXIST client echo ERROR: can't create %CD%/client
IF NOT EXIST client echo ERROR: can't find %CD%/client
IF NOT EXIST client exit /b 1
cd client
IF %errorlevel% NEQ 0 EXIT /b %errorlevel%
%PACKAGER% install
REM ^ Download all packages for client.
REM echo Running %PACKAGER% in "%CD%"...
REM pause
REM %PACKAGER% install axios
REM NEVER DO THIS after create-react-app (See issue#13) %PACKAGER% install eslint --save-dev
REM %PACKAGER% install --save-dev eslint-config-react-app @typescript-eslint/eslint-plugin@^4.0.0 @typescript-eslint/parser@^4.0.0 babel-eslint@^10.0.0 eslint@^7.5.0 eslint-plugin-flowtype@^5.2.0 eslint-plugin-import@^2.22.0 eslint-plugin-jsx-a11y@^6.3.1 eslint-plugin-react@^7.20.3 eslint-plugin-react-hooks@^4.0.8
REM ^ as per <https://www.npmjs.com/package/eslint-config-react-app> (and add their .eslintrc.json file)
REM as per https://daveceddia.com/tailwind-create-react-app/:
REM %PACKAGER% install react-router-dom
REM %PACKAGER% install tailwindcss %PACKAGER_RUNALL% chokidar-cli
REM %PACKAGER% install postcss
REM %PACKAGER% install @fortawesome/fontawesome-svg-core ^
REM @fortawesome/free-solid-svg-icons ^
REM @fortawesome/react-fontawesome
REM %PACKAGER% install react-infinite-scroll-component
REM %PACKAGER% install tailwind-navbar-react
REM ^ Preview: https://fontawesome.com/icons?s=solid&m=free
REM as per BezKoder (2019):
REM %PACKAGER% install react-validation validator
pause
REM cmd /k
REM References
REM BezKoder. (2019, October 19). Node.js + MongoDB: User authentication
REM & authorization with JWT. BezKoder.
REM https://bezkoder.com/node-js-mongodb-auth-jwt/
```
|
1.0
|
Add past install steps to development documentation. - from old install.bat:
```batch
@echo off
REM node
REM Most lines are commented since you only have to run them once to
REM create the workspace (such as downloading dependencies to
REM node_modules directories)
cd ..
REM SET PATH=%PATH%;%CD%/node-v14.15.0-win-x64
cd artspatter
echo PATH is %PATH%
echo Running npm in "%CD%"...
SET PACKAGER_RUNALL=npm-run-all
SET PACKAGER=%CD%/../node-v14.15.0-win-x64/npm.cmd
REM SET PACKAGER=npm
REM ^ Requires adding the %CD%/../node-v14.15.0-win-x64 to the system PATH,
REM since create-react-app also tries to call npm.
%PACKAGER% install
REM ^ Download all packages for server.
REM %PACKAGER% init
REM %PACKAGER% install express
REM %PACKAGER% install dotenv
REM %PACKAGER% install dotenv
REM %PACKAGER% install mongoose
SET CLIENT_APP=client
REM Don't do: %CD%/../node-v14.15.0-win-x64/npx.cmd create-react-app %CLIENT_APP%
REM You need the serviceWorker:
REM md client
REM cd client
REM npx create-react-app %CLIENT_APP% --template cra-template-pwa
REM ^ See https://github.com/facebook/create-react-app/issues/10032
REM %PACKAGER% install concurrently --save-dev
REM %PACKAGER% install nodemon --save-dev
REM ^ --save-dev dependencies aren't automatically installed with
REM "npm.cmd install" on the consumer (server owner)'s machine.
REM %PACKAGER% install
REM %PACKAGER% install
REM %PACKAGER% install mongoose-sequence
REM %PACKAGER% install eslint --save-dev
REM ^ Do NOT install eslint in the react client (see issue #13)
REM %PACKAGER% install body-parser
REM %PACKAGER% install express mongoose body-parser cors jsonwebtoken bcryptjs
REM ^ This is for authentication (BezKoder, 2019).
REM %PACKAGER% install mocha
REM %PACKAGER% install multer
REM - installed and removed fs--I don't need io, just file operations.
REM %PACKAGER% install nanoid
REM %PACKAGER% install node-thumbnail
echo CLIENT
echo The following packages only apply to the React.js user-facing page.
REM IF NOT EXIST client md client
REM IF NOT EXIST client echo ERROR: can't create %CD%/client
IF NOT EXIST client echo ERROR: can't find %CD%/client
IF NOT EXIST client exit /b 1
cd client
IF %errorlevel% NEQ 0 EXIT /b %errorlevel%
%PACKAGER% install
REM ^ Download all packages for client.
REM echo Running %PACKAGER% in "%CD%"...
REM pause
REM %PACKAGER% install axios
REM NEVER DO THIS after create-react-app (See issue#13) %PACKAGER% install eslint --save-dev
REM %PACKAGER% install --save-dev eslint-config-react-app @typescript-eslint/eslint-plugin@^4.0.0 @typescript-eslint/parser@^4.0.0 babel-eslint@^10.0.0 eslint@^7.5.0 eslint-plugin-flowtype@^5.2.0 eslint-plugin-import@^2.22.0 eslint-plugin-jsx-a11y@^6.3.1 eslint-plugin-react@^7.20.3 eslint-plugin-react-hooks@^4.0.8
REM ^ as per <https://www.npmjs.com/package/eslint-config-react-app> (and add their .eslintrc.json file)
REM as per https://daveceddia.com/tailwind-create-react-app/:
REM %PACKAGER% install react-router-dom
REM %PACKAGER% install tailwindcss %PACKAGER_RUNALL% chokidar-cli
REM %PACKAGER% install postcss
REM %PACKAGER% install @fortawesome/fontawesome-svg-core ^
REM @fortawesome/free-solid-svg-icons ^
REM @fortawesome/react-fontawesome
REM %PACKAGER% install react-infinite-scroll-component
REM %PACKAGER% install tailwind-navbar-react
REM ^ Preview: https://fontawesome.com/icons?s=solid&m=free
REM as per BezKoder (2019):
REM %PACKAGER% install react-validation validator
pause
REM cmd /k
REM References
REM BezKoder. (2019, October 19). Node.js + MongoDB: User authentication
REM & authorization with JWT. BezKoder.
REM https://bezkoder.com/node-js-mongodb-auth-jwt/
```
|
non_test
|
add past install steps to development documentation from old install bat batch echo off rem node rem most lines are commented since you only have to run them once to rem create the workspace such as downloading dependencies to rem node modules directories cd rem set path path cd node win cd artspatter echo path is path echo running npm in cd set packager runall npm run all set packager cd node win npm cmd rem set packager npm rem requires adding the cd node win to the system path rem since create react app also tries to call npm packager install rem download all packages for server rem packager init rem packager install express rem packager install dotenv rem packager install dotenv rem packager install mongoose set client app client rem don t do cd node win npx cmd create react app client app rem you need the serviceworker rem md client rem cd client rem npx create react app client app template cra template pwa rem see rem packager install concurrently save dev rem packager install nodemon save dev rem save dev dependencies aren t automatically installed with rem npm cmd install on the consumer server owner s machine rem packager install rem packager install rem packager install mongoose sequence rem packager install eslint save dev rem do not install eslint in the react client see issue rem packager install body parser rem packager install express mongoose body parser cors jsonwebtoken bcryptjs rem this is for authentication bezkoder rem packager install mocha rem packager install multer rem installed and removed fs i don t need io just file operations rem packager install nanoid rem packager install node thumbnail echo client echo the following packages only apply to the react js user facing page rem if not exist client md client rem if not exist client echo error can t create cd client if not exist client echo error can t find cd client if not exist client exit b cd client if errorlevel neq exit b errorlevel packager install rem download all packages for client rem echo running packager in cd rem pause rem packager install axios rem never do this after create react app see issue packager install eslint save dev rem packager install save dev eslint config react app typescript eslint eslint plugin typescript eslint parser babel eslint eslint eslint plugin flowtype eslint plugin import eslint plugin jsx eslint plugin react eslint plugin react hooks rem as per and add their eslintrc json file rem as per rem packager install react router dom rem packager install tailwindcss packager runall chokidar cli rem packager install postcss rem packager install fortawesome fontawesome svg core rem fortawesome free solid svg icons rem fortawesome react fontawesome rem packager install react infinite scroll component rem packager install tailwind navbar react rem preview rem as per bezkoder rem packager install react validation validator pause rem cmd k rem references rem bezkoder october node js mongodb user authentication rem authorization with jwt bezkoder rem
| 0
|
247,481
| 26,711,644,008
|
IssuesEvent
|
2023-01-28 01:17:40
|
Trinadh465/linux-3.0.35_CVE-2019-10220
|
https://api.github.com/repos/Trinadh465/linux-3.0.35_CVE-2019-10220
|
opened
|
CVE-2023-23455 (Medium) detected in linuxlinux-3.0.40
|
security vulnerability
|
## CVE-2023-23455 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-3.0.35_CVE-2019-10220/commit/1ad216eff9de6bc4364330a52955ad1743ba56e2">1ad216eff9de6bc4364330a52955ad1743ba56e2</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/sch_atm.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
atm_tc_enqueue in net/sched/sch_atm.c in the Linux kernel through 6.1.4 allows attackers to cause a denial of service because of type confusion (non-negative numbers can sometimes indicate a TC_ACT_SHOT condition rather than valid classification results).
<p>Publish Date: 2023-01-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23455>CVE-2023-23455</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-23455">https://www.linuxkernelcves.com/cves/CVE-2023-23455</a></p>
<p>Release Date: 2023-01-12</p>
<p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.87,v6.0.19,v6.1.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-23455 (Medium) detected in linuxlinux-3.0.40 - ## CVE-2023-23455 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-3.0.35_CVE-2019-10220/commit/1ad216eff9de6bc4364330a52955ad1743ba56e2">1ad216eff9de6bc4364330a52955ad1743ba56e2</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/sch_atm.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
atm_tc_enqueue in net/sched/sch_atm.c in the Linux kernel through 6.1.4 allows attackers to cause a denial of service because of type confusion (non-negative numbers can sometimes indicate a TC_ACT_SHOT condition rather than valid classification results).
<p>Publish Date: 2023-01-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23455>CVE-2023-23455</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-23455">https://www.linuxkernelcves.com/cves/CVE-2023-23455</a></p>
<p>Release Date: 2023-01-12</p>
<p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.87,v6.0.19,v6.1.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files net sched sch atm c vulnerability details atm tc enqueue in net sched sch atm c in the linux kernel through allows attackers to cause a denial of service because of type confusion non negative numbers can sometimes indicate a tc act shot condition rather than valid classification results publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
419,408
| 12,223,169,168
|
IssuesEvent
|
2020-05-02 16:25:19
|
minj/foxtrick
|
https://api.github.com/repos/minj/foxtrick
|
closed
|
Add player notes as an icon tooltip on the team players page
|
Priority-Low SkillTable chpp-granted feature wip
|
Original [issue 1229](https://code.google.com/p/foxtrick/issues/detail?id=1229) created by [kosio.petkov](mailto:kosio.petkov@gmail.com) on 2014-08-15T19:13:23.000Z:
Hello,
I often use the Notes for players, especially those in the academy. When I open the players list I have to open each player's page to see the notes I've added. Instead, it would be useful to have a "Notes" link for every player with notes so that I can see them on hover (tooltip like).
Thanks!
|
1.0
|
Add player notes as an icon tooltip on the team players page - Original [issue 1229](https://code.google.com/p/foxtrick/issues/detail?id=1229) created by [kosio.petkov](mailto:kosio.petkov@gmail.com) on 2014-08-15T19:13:23.000Z:
Hello,
I often use the Notes for players, especially those in the academy. When I open the players list I have to open each player's page to see the notes I've added. Instead, it would be useful to have a "Notes" link for every player with notes so that I can see them on hover (tooltip like).
Thanks!
|
non_test
|
add player notes as an icon tooltip on the team players page original created by mailto kosio petkov gmail com on hello i often use the notes for players especially those in the academy when i open the players list i have to open each player s page to see the notes i ve added instead it would be useful to have a quot notes quot link for every player with notes so that i can see them on hover tooltip like thanks
| 0
|
504,213
| 14,615,066,409
|
IssuesEvent
|
2020-12-22 10:55:13
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
sports.ndtv.com - see bug description
|
browser-focus-geckoview engine-gecko priority-important type-tracking-protection-strict
|
<!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64089 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://sports.ndtv.com/australia-vs-india-2020-21/prithvi-shaw-posts-cryptic-message-on-social-media-after-adelaide-failures-2341612?amp=1
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: website contents not visible
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
sports.ndtv.com - see bug description - <!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64089 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://sports.ndtv.com/australia-vs-india-2020-21/prithvi-shaw-posts-cryptic-message-on-social-media-after-adelaide-failures-2341612?amp=1
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: website contents not visible
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
sports ndtv com see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description website contents not visible steps to reproduce browser configuration none from with ❤️
| 0
|
238,911
| 19,785,783,449
|
IssuesEvent
|
2022-01-18 06:28:28
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
ccl/serverccl/diagnosticsccl: TestTenantReport failed
|
C-test-failure O-robot branch-master
|
ccl/serverccl/diagnosticsccl.TestTenantReport [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4140537&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4140537&tab=artifacts#/) on master @ [365b4da8bd02c06ee59d2130a56dec74ffc9ce21](https://github.com/cockroachdb/cockroach/commits/365b4da8bd02c06ee59d2130a56dec74ffc9ce21):
```
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/auth.go:97 +0x2a1
github.com/cockroachdb/cockroach/pkg/rpc.kvAuth.streamInterceptor-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/auth.go:74 +0x92
google.golang.org/grpc.chainStreamInterceptors.func1.1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1409 +0x1e2
github.com/cockroachdb/cockroach/pkg/rpc.NewServer.func2.1()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:213 +0x6f
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:344 +0x161
github.com/cockroachdb/cockroach/pkg/rpc.NewServer.func2()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:212 +0x152
google.golang.org/grpc.chainStreamInterceptors.func1.1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1409 +0x1e2
google.golang.org/grpc.chainStreamInterceptors.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1411 +0x261
google.golang.org/grpc.(*Server).processStreamingRPC()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1547 +0x1a6d
google.golang.org/grpc.(*Server).handleStream()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1620 +0xf89
google.golang.org/grpc.(*Server).serveStreams.func1.2()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:921 +0xfd
Goroutine 345 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:478 +0x615
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:354 +0x13d
github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber.(*KVSubscriber).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber/kvsubscriber.go:177 +0x26
github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1801 +0x41fa
github.com/cockroachdb/cockroach/pkg/server.(*Server).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44
github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x64
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:241 +0x17b
github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl_test.startReporterTest()
/go/src/github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl/reporter_test.go:430 +0xa64
github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl_test.TestTenantReport()
/go/src/github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl/reporter_test.go:51 +0x124
testing.tRunner()
/usr/local/go/src/testing/testing.go:1259 +0x22f
testing.(*T).Run·dwrap·21()
/usr/local/go/src/testing/testing.go:1306 +0x47
==================
reporter_test.go:95: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantReport4183284647
testing.go:1152: race detected during execution of test
--- FAIL: TestTenantReport (3.55s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- GOFLAGS=-race -parallel=4
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantReport.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
ccl/serverccl/diagnosticsccl: TestTenantReport failed - ccl/serverccl/diagnosticsccl.TestTenantReport [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4140537&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4140537&tab=artifacts#/) on master @ [365b4da8bd02c06ee59d2130a56dec74ffc9ce21](https://github.com/cockroachdb/cockroach/commits/365b4da8bd02c06ee59d2130a56dec74ffc9ce21):
```
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/auth.go:97 +0x2a1
github.com/cockroachdb/cockroach/pkg/rpc.kvAuth.streamInterceptor-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/auth.go:74 +0x92
google.golang.org/grpc.chainStreamInterceptors.func1.1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1409 +0x1e2
github.com/cockroachdb/cockroach/pkg/rpc.NewServer.func2.1()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:213 +0x6f
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:344 +0x161
github.com/cockroachdb/cockroach/pkg/rpc.NewServer.func2()
/go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:212 +0x152
google.golang.org/grpc.chainStreamInterceptors.func1.1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1409 +0x1e2
google.golang.org/grpc.chainStreamInterceptors.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1411 +0x261
google.golang.org/grpc.(*Server).processStreamingRPC()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1547 +0x1a6d
google.golang.org/grpc.(*Server).handleStream()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1620 +0xf89
google.golang.org/grpc.(*Server).serveStreams.func1.2()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:921 +0xfd
Goroutine 345 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:478 +0x615
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:354 +0x13d
github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber.(*KVSubscriber).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber/kvsubscriber.go:177 +0x26
github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1801 +0x41fa
github.com/cockroachdb/cockroach/pkg/server.(*Server).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44
github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x64
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:241 +0x17b
github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl_test.startReporterTest()
/go/src/github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl/reporter_test.go:430 +0xa64
github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl_test.TestTenantReport()
/go/src/github.com/cockroachdb/cockroach/pkg/ccl/serverccl/diagnosticsccl/reporter_test.go:51 +0x124
testing.tRunner()
/usr/local/go/src/testing/testing.go:1259 +0x22f
testing.(*T).Run·dwrap·21()
/usr/local/go/src/testing/testing.go:1306 +0x47
==================
reporter_test.go:95: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantReport4183284647
testing.go:1152: race detected during execution of test
--- FAIL: TestTenantReport (3.55s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- GOFLAGS=-race -parallel=4
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantReport.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
ccl serverccl diagnosticsccl testtenantreport failed ccl serverccl diagnosticsccl testtenantreport with on master go src github com cockroachdb cockroach pkg rpc auth go github com cockroachdb cockroach pkg rpc kvauth streaminterceptor fm go src github com cockroachdb cockroach pkg rpc auth go google golang org grpc chainstreaminterceptors go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach pkg rpc newserver go src github com cockroachdb cockroach pkg rpc context go github com cockroachdb cockroach pkg util stop stopper runtaskwitherr go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg rpc newserver go src github com cockroachdb cockroach pkg rpc context go google golang org grpc chainstreaminterceptors go src github com cockroachdb cockroach vendor google golang org grpc server go google golang org grpc chainstreaminterceptors go src github com cockroachdb cockroach vendor google golang org grpc server go google golang org grpc server processstreamingrpc go src github com cockroachdb cockroach vendor google golang org grpc server go google golang org grpc server handlestream go src github com cockroachdb cockroach vendor google golang org grpc server go google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctaskex go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg spanconfig spanconfigkvsubscriber kvsubscriber start go src github com cockroachdb cockroach pkg spanconfig spanconfigkvsubscriber kvsubscriber go github com cockroachdb cockroach pkg server server prestart go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server testserver start go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg testutils serverutils startserver go src github com cockroachdb cockroach pkg testutils serverutils test server shim go github com cockroachdb cockroach pkg ccl serverccl diagnosticsccl test startreportertest go src github com cockroachdb cockroach pkg ccl serverccl diagnosticsccl reporter test go github com cockroachdb cockroach pkg ccl serverccl diagnosticsccl test testtenantreport go src github com cockroachdb cockroach pkg ccl serverccl diagnosticsccl reporter test go testing trunner usr local go src testing testing go testing t run·dwrap· usr local go src testing testing go reporter test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts testing go race detected during execution of test fail testtenantreport help see also parameters in this failure goflags race parallel cc cockroachdb server
| 1
|
25,385
| 4,156,219,606
|
IssuesEvent
|
2016-06-16 17:14:37
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
e2e test failure: Kubectl apply should apply a new configuration to an existing RC
|
kind/flake team/node team/test-infra
|
[Fail] [k8s.io] Kubectl client [AfterEach] [k8s.io] Kubectl apply should apply a new configuration to an existing RC
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:230
Assigning to test-infra team to triage.
|
1.0
|
e2e test failure: Kubectl apply should apply a new configuration to an existing RC - [Fail] [k8s.io] Kubectl client [AfterEach] [k8s.io] Kubectl apply should apply a new configuration to an existing RC
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:230
Assigning to test-infra team to triage.
|
test
|
test failure kubectl apply should apply a new configuration to an existing rc kubectl client kubectl apply should apply a new configuration to an existing rc go src io kubernetes output dockerized go src io kubernetes test framework framework go assigning to test infra team to triage
| 1
|
199,434
| 15,038,767,179
|
IssuesEvent
|
2021-02-02 17:50:00
|
livepeer/streamer
|
https://api.github.com/repos/livepeer/streamer
|
opened
|
Page on excessive renames
|
testing
|
Sometimes a region's streams will experience thrashing transcoding and the streams will experience many renames in a row.
Let's page on rename.
|
1.0
|
Page on excessive renames - Sometimes a region's streams will experience thrashing transcoding and the streams will experience many renames in a row.
Let's page on rename.
|
test
|
page on excessive renames sometimes a region s streams will experience thrashing transcoding and the streams will experience many renames in a row let s page on rename
| 1
|
76,087
| 26,229,246,351
|
IssuesEvent
|
2023-01-04 21:57:33
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[FE] 10-10CG - File Upload: 508-defect-1 ❗ Launchblocker: Representative documentation appears under supporting documentation, unable to edit.
|
508/Accessibility 508-defect-1
|
# [508-defect-1 :exclamation: Launchblocker](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-1)
Issue Title: 508 Defect 1, Launchblocker! Representative documentation appears under supporting documentation, unable to edit.
Point of Contact: Angela Fowler
**VFS Point of Contact:** _First name only: Angela
<h2>User Story or Problem Statement:</h2>
As a user, I expect to be able to upload all necessary documents. I expect that once uploaded, documents appear in their assigned place and that I can edit them when necessary.
<h2>Details</h2>
When uploading documentation, I am presented with only the option to upload representative documentation. Once I do, the document is presented under "supporting documentation" and I am unable to delete it. I am also unable to edit representative documentation so that the form appears in the right place.
**NOTE: This issue may already be resolved. Please check the steps and confirm**
<h2>Acceptance Criteria</h2>
- [ ] The user is able to upload both representative and supporting documentation and they appear in their proper places on review.
- [ ] Alternatively, the option to upload supporting documentation is removed completely.
Steps to Recreate
<ul>
<li> Access https://staging.va.gov/family-member-benefits/apply-for-caregiver-assistance-form-10-10cg/introduction</li>
<li> Choose "Begin your application."</li>
<li> complete steps 1-3 of the application</li>
<li>upload a PDF which is less than 25MB per the submission guidelines.</li>
<li>Proceed to step 5.</li>
<li>Attempt to edit the representative documentation.</li>
</ul>
|
1.0
|
[FE] 10-10CG - File Upload: 508-defect-1 ❗ Launchblocker: Representative documentation appears under supporting documentation, unable to edit. - # [508-defect-1 :exclamation: Launchblocker](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-1)
Issue Title: 508 Defect 1, Launchblocker! Representative documentation appears under supporting documentation, unable to edit.
Point of Contact: Angela Fowler
**VFS Point of Contact:** _First name only: Angela
<h2>User Story or Problem Statement:</h2>
As a user, I expect to be able to upload all necessary documents. I expect that once uploaded, documents appear in their assigned place and that I can edit them when necessary.
<h2>Details</h2>
When uploading documentation, I am presented with only the option to upload representative documentation. Once I do, the document is presented under "supporting documentation" and I am unable to delete it. I am also unable to edit representative documentation so that the form appears in the right place.
**NOTE: This issue may already be resolved. Please check the steps and confirm**
<h2>Acceptance Criteria</h2>
- [ ] The user is able to upload both representative and supporting documentation and they appear in their proper places on review.
- [ ] Alternatively, the option to upload supporting documentation is removed completely.
Steps to Recreate
<ul>
<li> Access https://staging.va.gov/family-member-benefits/apply-for-caregiver-assistance-form-10-10cg/introduction</li>
<li> Choose "Begin your application."</li>
<li> complete steps 1-3 of the application</li>
<li>upload a PDF which is less than 25MB per the submission guidelines.</li>
<li>Proceed to step 5.</li>
<li>Attempt to edit the representative documentation.</li>
</ul>
|
non_test
|
file upload defect ❗ launchblocker representative documentation appears under supporting documentation unable to edit issue title defect launchblocker representative documentation appears under supporting documentation unable to edit point of contact angela fowler vfs point of contact first name only angela user story or problem statement as a user i expect to be able to upload all necessary documents i expect that once uploaded documents appear in their assigned place and that i can edit them when necessary details when uploading documentation i am presented with only the option to upload representative documentation once i do the document is presented under supporting documentation and i am unable to delete it i am also unable to edit representative documentation so that the form appears in the right place note this issue may already be resolved please check the steps and confirm acceptance criteria the user is able to upload both representative and supporting documentation and they appear in their proper places on review alternatively the option to upload supporting documentation is removed completely steps to recreate access choose begin your application complete steps of the application upload a pdf which is less than per the submission guidelines proceed to step attempt to edit the representative documentation
| 0
|
1,022
| 2,677,681,653
|
IssuesEvent
|
2015-03-26 02:24:33
|
rethinkdb/rethinkdb
|
https://api.github.com/repos/rethinkdb/rethinkdb
|
closed
|
Use real dependency management for the webui code like browserify/require.js
|
cp:build cp:web-ui
|
Right now the order of our js files loading is done in mk/webui.mk. All the web ui coffee files must be explicitly mentioned there in the correct order so their dependencies are concatenated earlier in the file.
This is pretty brittle, annoying and unnecessary because there are tools like browserify and require.js
|
1.0
|
Use real dependency management for the webui code like browserify/require.js - Right now the order of our js files loading is done in mk/webui.mk. All the web ui coffee files must be explicitly mentioned there in the correct order so their dependencies are concatenated earlier in the file.
This is pretty brittle, annoying and unnecessary because there are tools like browserify and require.js
|
non_test
|
use real dependency management for the webui code like browserify require js right now the order of our js files loading is done in mk webui mk all the web ui coffee files must be explicitly mentioned there in the correct order so their dependencies are concatenated earlier in the file this is pretty brittle annoying and unnecessary because there are tools like browserify and require js
| 0
|
93,257
| 19,143,434,754
|
IssuesEvent
|
2021-12-02 03:19:04
|
KeyWorksRW/wxUiEditor
|
https://api.github.com/repos/KeyWorksRW/wxUiEditor
|
closed
|
Review default settings for wxFrame window
|
code review/refactor
|
### Description:
<!-- Provide a description of of what you want to happen here -->
The default settings for a wxFrame window need review. Setting **wxTAB_TRAVERSAL** as the default style doesn't really make much sense -- particularly if we support adding a wxPanel where that setting would make a lot more sense. The default size won't make sense once we add a persist option. It might make more sense to ask the user if they want to set a default size rather than forcing one on them and requiring them to remove it if they don't actually want it.
|
1.0
|
Review default settings for wxFrame window - ### Description:
<!-- Provide a description of of what you want to happen here -->
The default settings for a wxFrame window need review. Setting **wxTAB_TRAVERSAL** as the default style doesn't really make much sense -- particularly if we support adding a wxPanel where that setting would make a lot more sense. The default size won't make sense once we add a persist option. It might make more sense to ask the user if they want to set a default size rather than forcing one on them and requiring them to remove it if they don't actually want it.
|
non_test
|
review default settings for wxframe window description the default settings for a wxframe window need review setting wxtab traversal as the default style doesn t really make much sense particularly if we support adding a wxpanel where that setting would make a lot more sense the default size won t make sense once we add a persist option it might make more sense to ask the user if they want to set a default size rather than forcing one on them and requiring them to remove it if they don t actually want it
| 0
|
249,506
| 21,163,993,013
|
IssuesEvent
|
2022-04-07 12:02:29
|
kubernetes/test-infra
|
https://api.github.com/repos/kubernetes/test-infra
|
closed
|
Trigger the milestone applier on branch change in a PR
|
area/prow kind/bug sig/testing lifecycle/rotten
|
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened**:
The milestone applier is set to be applied for[ all PRs targeting the 1.22 release branch for k/website](https://github.com/kubernetes/test-infra/blob/master/config/prow/plugins.yaml#L389), however, this [PR targeted master by mistake](https://github.com/kubernetes/website/pull/27810). The PR was fixed, and is now targeting dev1.22 branch, however, the bot didn’t automatically apply the 1.22 milestone when the fix was done.
**What you expected to happen**:
The bot should retroactively apply the milestone applier if the branch has changed from a non-tracked or tracked one to the one reflected in the config.
**How to reproduce it (as minimally and precisely as possible)**:
I haven't tested this out myself outside of the example provided above, but i think that if you open a PR targeting `master` branch and then switch to `dev-1.22` the bot should do nothing at all.
**Please provide links to example occurrences, if any**:
https://github.com/kubernetes/website/pull/27810
**Anything else we need to know?**:
Maybe this is intentional so that prow doesn't go around applying milestones creating side effects, would like to know if that's the case. if not, i'd be happy to fix this in an upcoming PR.
|
1.0
|
Trigger the milestone applier on branch change in a PR - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened**:
The milestone applier is set to be applied for[ all PRs targeting the 1.22 release branch for k/website](https://github.com/kubernetes/test-infra/blob/master/config/prow/plugins.yaml#L389), however, this [PR targeted master by mistake](https://github.com/kubernetes/website/pull/27810). The PR was fixed, and is now targeting dev1.22 branch, however, the bot didn’t automatically apply the 1.22 milestone when the fix was done.
**What you expected to happen**:
The bot should retroactively apply the milestone applier if the branch has changed from a non-tracked or tracked one to the one reflected in the config.
**How to reproduce it (as minimally and precisely as possible)**:
I haven't tested this out myself outside of the example provided above, but i think that if you open a PR targeting `master` branch and then switch to `dev-1.22` the bot should do nothing at all.
**Please provide links to example occurrences, if any**:
https://github.com/kubernetes/website/pull/27810
**Anything else we need to know?**:
Maybe this is intentional so that prow doesn't go around applying milestones creating side effects, would like to know if that's the case. if not, i'd be happy to fix this in an upcoming PR.
|
test
|
trigger the milestone applier on branch change in a pr what happened the milestone applier is set to be applied for however this the pr was fixed and is now targeting branch however the bot didn’t automatically apply the milestone when the fix was done what you expected to happen the bot should retroactively apply the milestone applier if the branch has changed from a non tracked or tracked one to the one reflected in the config how to reproduce it as minimally and precisely as possible i haven t tested this out myself outside of the example provided above but i think that if you open a pr targeting master branch and then switch to dev the bot should do nothing at all please provide links to example occurrences if any anything else we need to know maybe this is intentional so that prow doesn t go around applying milestones creating side effects would like to know if that s the case if not i d be happy to fix this in an upcoming pr
| 1
|
174,256
| 13,463,983,144
|
IssuesEvent
|
2020-09-09 18:29:05
|
rancher/cis-operator
|
https://api.github.com/repos/rancher/cis-operator
|
closed
|
CIS charts are not available even after adding the catalog/repo
|
[zube]: To Test kind/bug-qa
|
**What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Navigate to Apps --> Repositories -->Add a repo in dashboard URL: https://github.com/prachidamle/charts.git and branch: `cis-v2-chart`
- Go to charts
- CIS benchmark charts are NOT available.
**Expected Result:**
CIS benchmark charts should be available
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - commit id: `9b6984d5f`
- Installation option (single install/HA): HA
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): imported rks k8s
- Kubernetes version (use `kubectl version`):
```
1.18
```
|
1.0
|
CIS charts are not available even after adding the catalog/repo - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Navigate to Apps --> Repositories -->Add a repo in dashboard URL: https://github.com/prachidamle/charts.git and branch: `cis-v2-chart`
- Go to charts
- CIS benchmark charts are NOT available.
**Expected Result:**
CIS benchmark charts should be available
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - commit id: `9b6984d5f`
- Installation option (single install/HA): HA
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): imported rks k8s
- Kubernetes version (use `kubectl version`):
```
1.18
```
|
test
|
cis charts are not available even after adding the catalog repo what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible navigate to apps repositories add a repo in dashboard url and branch cis chart go to charts cis benchmark charts are not available expected result cis benchmark charts should be available environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head commit id installation option single install ha ha if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported imported rks kubernetes version use kubectl version
| 1
|
243,116
| 20,363,269,448
|
IssuesEvent
|
2022-02-21 00:17:44
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
Frequent test failures of `TestFunctional/parallel/AddonsCmd`
|
priority/backlog kind/failing-test
|
This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[Docker_Linux_crio_arm64](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=Docker_Linux_crio_arm64&test=TestFunctional/parallel/AddonsCmd)|23.53|
|
1.0
|
Frequent test failures of `TestFunctional/parallel/AddonsCmd` - This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[Docker_Linux_crio_arm64](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=Docker_Linux_crio_arm64&test=TestFunctional/parallel/AddonsCmd)|23.53|
|
test
|
frequent test failures of testfunctional parallel addonscmd this test has high flake rates for the following environments environment flake rate
| 1
|
4,834
| 2,752,019,274
|
IssuesEvent
|
2015-04-24 13:18:03
|
NativeScript/NativeScript
|
https://api.github.com/repos/NativeScript/NativeScript
|
closed
|
iOS - Swipe Backwards Navigation - Half Swipe
|
4 - Ready For Test S:Medium T:Bug
|
_From @ligaz on March 30, 2015 18:50_
_From @papaytl185 on March 24, 2015 13:47_
If you begin to swipe backward in ios to go to the previous page, but then stay on the current page it will lock / freeze the app. This happens for all the sample applications as well. I am using appbuilder companion app, iphone 6 - ios 8.2
_Copied from original issue: NativeScript/NativeScript#20_
_Copied from original issue: NativeScript/cross-platform-modules#234_
|
1.0
|
iOS - Swipe Backwards Navigation - Half Swipe - _From @ligaz on March 30, 2015 18:50_
_From @papaytl185 on March 24, 2015 13:47_
If you begin to swipe backward in ios to go to the previous page, but then stay on the current page it will lock / freeze the app. This happens for all the sample applications as well. I am using appbuilder companion app, iphone 6 - ios 8.2
_Copied from original issue: NativeScript/NativeScript#20_
_Copied from original issue: NativeScript/cross-platform-modules#234_
|
test
|
ios swipe backwards navigation half swipe from ligaz on march from on march if you begin to swipe backward in ios to go to the previous page but then stay on the current page it will lock freeze the app this happens for all the sample applications as well i am using appbuilder companion app iphone ios copied from original issue nativescript nativescript copied from original issue nativescript cross platform modules
| 1
|
44,289
| 5,621,285,531
|
IssuesEvent
|
2017-04-04 09:32:59
|
OAButton/discussion
|
https://api.github.com/repos/OAButton/discussion
|
closed
|
Allow DOIs / PMC / PMID (and maybe titles) to be put into our url box on /request/new
|
Blocked: Copy Blocked: Development Blocked: Test enhancement JISC Website
|
Like http://cx.ctg.li/ & oadoi.org
|
1.0
|
Allow DOIs / PMC / PMID (and maybe titles) to be put into our url box on /request/new - Like http://cx.ctg.li/ & oadoi.org
|
test
|
allow dois pmc pmid and maybe titles to be put into our url box on request new like oadoi org
| 1
|
305,761
| 23,129,831,034
|
IssuesEvent
|
2022-07-28 09:22:32
|
serenity-rs/songbird
|
https://api.github.com/repos/serenity-rs/songbird
|
closed
|
(next) `tracks::create_player` missing
|
bug documentation tracks
|
Although [documented](https://serenity-rs.github.io/songbird/next/songbird/tracks/fn.create_player.html), this function does not exists in the source code, and is impossible to import.
|
1.0
|
(next) `tracks::create_player` missing - Although [documented](https://serenity-rs.github.io/songbird/next/songbird/tracks/fn.create_player.html), this function does not exists in the source code, and is impossible to import.
|
non_test
|
next tracks create player missing although this function does not exists in the source code and is impossible to import
| 0
|
435,900
| 30,525,988,525
|
IssuesEvent
|
2023-07-19 11:18:06
|
mindsdb/mindsdb
|
https://api.github.com/repos/mindsdb/mindsdb
|
closed
|
[Data Handler][Testing][Docs] Test and document EdgelessDB connection to MindsDB
|
documentation
|
## Testing
1. Go to the handler's folder: https://github.com/mindsdb/mindsdb/tree/staging/mindsdb/integrations/handlers/edgelessdb_handler
2. Test the handler following instructions in the README file.
3. Create `Manual_QA.md` file to store test results, including execution of `CREATE DATABASE` and `SELECT`. If you encounter issues, you can [report a bug issue here](https://github.com/mindsdb/mindsdb/issues/new?assignees=&labels=bug&projects=&template=bug_report_v2.yaml&title=%5BBug%5D%3A+).
## Docs
1. Go to the `/mindsdb/docs/data-integrations` directory and create `edgelessdb.mdx` file.
2. Create content following [this template](https://docs.mindsdb.com/data-integrations/postgresql).
3. Go to the `/mindsdb/docs/mint.json` file and add `"data-integrations/edgelessdb"` (alphabetically) under `Data Sources` -> `Databases` groups.
|
1.0
|
[Data Handler][Testing][Docs] Test and document EdgelessDB connection to MindsDB - ## Testing
1. Go to the handler's folder: https://github.com/mindsdb/mindsdb/tree/staging/mindsdb/integrations/handlers/edgelessdb_handler
2. Test the handler following instructions in the README file.
3. Create `Manual_QA.md` file to store test results, including execution of `CREATE DATABASE` and `SELECT`. If you encounter issues, you can [report a bug issue here](https://github.com/mindsdb/mindsdb/issues/new?assignees=&labels=bug&projects=&template=bug_report_v2.yaml&title=%5BBug%5D%3A+).
## Docs
1. Go to the `/mindsdb/docs/data-integrations` directory and create `edgelessdb.mdx` file.
2. Create content following [this template](https://docs.mindsdb.com/data-integrations/postgresql).
3. Go to the `/mindsdb/docs/mint.json` file and add `"data-integrations/edgelessdb"` (alphabetically) under `Data Sources` -> `Databases` groups.
|
non_test
|
test and document edgelessdb connection to mindsdb testing go to the handler s folder test the handler following instructions in the readme file create manual qa md file to store test results including execution of create database and select if you encounter issues you can docs go to the mindsdb docs data integrations directory and create edgelessdb mdx file create content following go to the mindsdb docs mint json file and add data integrations edgelessdb alphabetically under data sources databases groups
| 0
|
208,143
| 15,877,670,655
|
IssuesEvent
|
2021-04-09 09:56:41
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
[Improvement] Provide a user confirmation window before deleting the auth application in a single step
|
Affected/5.7.0 Scenario Test improvement ui
|
Currently when we're trying to delete an existing auth application of a Service provider, it doesn't give a confirmation box to get the approval of the user before deleting it. Without having this, a configured auth application can be deleted in a single click (probably by accident) and it's better to add a confirmation window before deleting it.
Affected Product = IS 5.7.0
== Auth App ==
<img width="1116" alt="created_auth_app" src="https://user-images.githubusercontent.com/8849472/49716167-e3079480-fc78-11e8-8a8b-976a94ae529d.png">
== Deletion Successful Window without a Confirmation ==
<img width="1174" alt="auth_app_deletion" src="https://user-images.githubusercontent.com/8849472/49716186-f9adeb80-fc78-11e8-9618-59a4c6d3c5ad.png">
|
1.0
|
[Improvement] Provide a user confirmation window before deleting the auth application in a single step - Currently when we're trying to delete an existing auth application of a Service provider, it doesn't give a confirmation box to get the approval of the user before deleting it. Without having this, a configured auth application can be deleted in a single click (probably by accident) and it's better to add a confirmation window before deleting it.
Affected Product = IS 5.7.0
== Auth App ==
<img width="1116" alt="created_auth_app" src="https://user-images.githubusercontent.com/8849472/49716167-e3079480-fc78-11e8-8a8b-976a94ae529d.png">
== Deletion Successful Window without a Confirmation ==
<img width="1174" alt="auth_app_deletion" src="https://user-images.githubusercontent.com/8849472/49716186-f9adeb80-fc78-11e8-9618-59a4c6d3c5ad.png">
|
test
|
provide a user confirmation window before deleting the auth application in a single step currently when we re trying to delete an existing auth application of a service provider it doesn t give a confirmation box to get the approval of the user before deleting it without having this a configured auth application can be deleted in a single click probably by accident and it s better to add a confirmation window before deleting it affected product is auth app img width alt created auth app src deletion successful window without a confirmation img width alt auth app deletion src
| 1
|
186,282
| 14,394,659,946
|
IssuesEvent
|
2020-12-03 01:49:22
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
iilness2/jenkins_jnlp_base_go1_11: go/src/go/internal/gccgoimporter/importer_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [iilness2/jenkins_jnlp_base_go1_11](https://www.github.com/iilness2/jenkins_jnlp_base_go1_11) at [go/src/go/internal/gccgoimporter/importer_test.go](https://github.com/iilness2/jenkins_jnlp_base_go1_11/blob/5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d/go/src/go/internal/gccgoimporter/importer_test.go#L115-L117)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/iilness2/jenkins_jnlp_base_go1_11/blob/5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d/go/src/go/internal/gccgoimporter/importer_test.go#L115-L117)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range importerTests {
runImporterTest(t, imp, initmap, &test)
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 116 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d
|
1.0
|
iilness2/jenkins_jnlp_base_go1_11: go/src/go/internal/gccgoimporter/importer_test.go; 3 LoC -
Found a possible issue in [iilness2/jenkins_jnlp_base_go1_11](https://www.github.com/iilness2/jenkins_jnlp_base_go1_11) at [go/src/go/internal/gccgoimporter/importer_test.go](https://github.com/iilness2/jenkins_jnlp_base_go1_11/blob/5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d/go/src/go/internal/gccgoimporter/importer_test.go#L115-L117)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/iilness2/jenkins_jnlp_base_go1_11/blob/5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d/go/src/go/internal/gccgoimporter/importer_test.go#L115-L117)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range importerTests {
runImporterTest(t, imp, initmap, &test)
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 116 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 5a6bd7dfabe7fb437f5c5df059e50fcfc874f40d
|
test
|
jenkins jnlp base go src go internal gccgoimporter importer test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for test range importertests runimportertest t imp initmap test below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
59,442
| 6,651,731,323
|
IssuesEvent
|
2017-09-28 21:12:23
|
Microsoft/PTVS
|
https://api.github.com/repos/Microsoft/PTVS
|
closed
|
UI test framework: Inconclusive is turned into a failure
|
Tests
|
```TestRunnerInterop.TestFailedException: AssertInconclusiveException: Assert.Inconclusive failed. Python interpreter not installed```
|
1.0
|
UI test framework: Inconclusive is turned into a failure - ```TestRunnerInterop.TestFailedException: AssertInconclusiveException: Assert.Inconclusive failed. Python interpreter not installed```
|
test
|
ui test framework inconclusive is turned into a failure testrunnerinterop testfailedexception assertinconclusiveexception assert inconclusive failed python interpreter not installed
| 1
|
263,166
| 23,039,361,400
|
IssuesEvent
|
2022-07-23 00:14:52
|
dask/distributed
|
https://api.github.com/repos/dask/distributed
|
closed
|
`Unused "type: ignore" comment` in CI linting
|
tests
|
I'm commonly seeing a linter error in CI about shuffle extension code when modifying unrelated code:
```
absolufy-imports.........................................................Passed
isort....................................................................Passed
pyupgrade................................................................Passed
black....................................................................Passed
flake8...................................................................Passed
mypy.....................................................................Failed
- hook id: mypy
- exit code: 1
distributed/shuffle/shuffle_extension.py:534: error: Unused "type: ignore" comment
```
I can't reproduce the error locally, and when I remove the comment, it then does trigger an actual mypy error.
|
1.0
|
`Unused "type: ignore" comment` in CI linting - I'm commonly seeing a linter error in CI about shuffle extension code when modifying unrelated code:
```
absolufy-imports.........................................................Passed
isort....................................................................Passed
pyupgrade................................................................Passed
black....................................................................Passed
flake8...................................................................Passed
mypy.....................................................................Failed
- hook id: mypy
- exit code: 1
distributed/shuffle/shuffle_extension.py:534: error: Unused "type: ignore" comment
```
I can't reproduce the error locally, and when I remove the comment, it then does trigger an actual mypy error.
|
test
|
unused type ignore comment in ci linting i m commonly seeing a linter error in ci about shuffle extension code when modifying unrelated code absolufy imports passed isort passed pyupgrade passed black passed passed mypy failed hook id mypy exit code distributed shuffle shuffle extension py error unused type ignore comment i can t reproduce the error locally and when i remove the comment it then does trigger an actual mypy error
| 1
|
116,095
| 4,696,679,280
|
IssuesEvent
|
2016-10-12 06:05:55
|
Kaezon/Unreal-SCP
|
https://api.github.com/repos/Kaezon/Unreal-SCP
|
opened
|
Complete the room_part assets
|
Art Low Priority
|
These need to be done from scratch as well since I changed the base template.
- [] room_part_corner
- [] room_part_wall
- [] room_part_wall_door
- [] room_part_wall_double
- [] room_part_wall_double_door
- [] room_part_dead_end
|
1.0
|
Complete the room_part assets - These need to be done from scratch as well since I changed the base template.
- [] room_part_corner
- [] room_part_wall
- [] room_part_wall_door
- [] room_part_wall_double
- [] room_part_wall_double_door
- [] room_part_dead_end
|
non_test
|
complete the room part assets these need to be done from scratch as well since i changed the base template room part corner room part wall room part wall door room part wall double room part wall double door room part dead end
| 0
|
66,659
| 7,009,502,455
|
IssuesEvent
|
2017-12-19 19:23:44
|
infiniteautomation/ma-core-public
|
https://api.github.com/repos/infiniteautomation/ma-core-public
|
closed
|
Missing Event Types Causes Mango to fail to start
|
Enhancement Ready for Testing
|
This is due to the Event Manager failing to startup when querying for the recent active events.
We will add a new Event Type to use for any event type definitions that are not found but an entry exists in the database.
|
1.0
|
Missing Event Types Causes Mango to fail to start - This is due to the Event Manager failing to startup when querying for the recent active events.
We will add a new Event Type to use for any event type definitions that are not found but an entry exists in the database.
|
test
|
missing event types causes mango to fail to start this is due to the event manager failing to startup when querying for the recent active events we will add a new event type to use for any event type definitions that are not found but an entry exists in the database
| 1
|
207,664
| 15,830,018,296
|
IssuesEvent
|
2021-04-06 11:58:33
|
LiskHQ/lisk-sdk
|
https://api.github.com/repos/LiskHQ/lisk-sdk
|
closed
|
Add functional tests for faucet plugin using testing utils
|
framework-plugins/faucet type: test
|
### Description
Faucet plugin requires testing which use the new framework testing utils.
### Motivation
Framework testing utils should be used to write tests once they are merged into development branch to write tests.
### Acceptance Criteria
Write tests using new utils
|
1.0
|
Add functional tests for faucet plugin using testing utils - ### Description
Faucet plugin requires testing which use the new framework testing utils.
### Motivation
Framework testing utils should be used to write tests once they are merged into development branch to write tests.
### Acceptance Criteria
Write tests using new utils
|
test
|
add functional tests for faucet plugin using testing utils description faucet plugin requires testing which use the new framework testing utils motivation framework testing utils should be used to write tests once they are merged into development branch to write tests acceptance criteria write tests using new utils
| 1
|
179,780
| 13,902,106,111
|
IssuesEvent
|
2020-10-20 04:34:28
|
chef/chef
|
https://api.github.com/repos/chef/chef
|
closed
|
Add functional test for windows_package with remote file
|
Aspect: Testing Status: Sustaining Backlog
|
We introduced a regression in 16.0 where you couldn't specify the path within the remote_file_attributes property of windows_pakage and have it use that path to install the package. This is a simple scenario we should be testing for. Here's the custom code that failed with the custom path set.
```ruby
windows_package node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['name'] do
checksum node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['sha256sum']
source node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['url']
installer_type :custom
returns [0, 3010]
remote_file_attributes ({
path: "#{Chef::Config[:file_cache_path]}\\package\\#{node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['name']} - #{node['vcruntime']['vc6']['version']}.exe",
checksum: node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['sha256sum'],
})
options '/q'
end
```
|
1.0
|
Add functional test for windows_package with remote file - We introduced a regression in 16.0 where you couldn't specify the path within the remote_file_attributes property of windows_pakage and have it use that path to install the package. This is a simple scenario we should be testing for. Here's the custom code that failed with the custom path set.
```ruby
windows_package node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['name'] do
checksum node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['sha256sum']
source node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['url']
installer_type :custom
returns [0, 3010]
remote_file_attributes ({
path: "#{Chef::Config[:file_cache_path]}\\package\\#{node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['name']} - #{node['vcruntime']['vc6']['version']}.exe",
checksum: node['vcruntime']['vc6']['x86'][node['vcruntime']['vc6']['version']]['sha256sum'],
})
options '/q'
end
```
|
test
|
add functional test for windows package with remote file we introduced a regression in where you couldn t specify the path within the remote file attributes property of windows pakage and have it use that path to install the package this is a simple scenario we should be testing for here s the custom code that failed with the custom path set ruby windows package node do checksum node source node installer type custom returns remote file attributes path chef config package node node exe checksum node options q end
| 1
|
486,277
| 14,006,793,099
|
IssuesEvent
|
2020-10-28 20:31:17
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
[consoleApp] Unable to add single logout request url via the console app for a saml web application
|
Priority/High Severity/Critical bug console ux
|
**Describe the issue:**
Unable to add single logout request url via the console app for a saml web application
**How to reproduce:**
1. Access https://localhost:9443/console
2. Login as super admin
3. Navigate to Develop > Applications > New Application > Web Application > SAML and add a saml application
4. When the application is created navigate to the Access tab > SAML > enable single logout profile
5. Select type as BACK CHANNEL
6. Enter a single logout request url (http://testlogout)
7. Update
**Observation**
Even though it says updated successfully.the single logout request url does not get updated
**Environment information**
Mssql 2019
uniqueid jdbc primary US
IS 5.11.0 beta3 snapshot
ubuntu 16.04
Chrome Version 85.0.4183.121

|
1.0
|
[consoleApp] Unable to add single logout request url via the console app for a saml web application - **Describe the issue:**
Unable to add single logout request url via the console app for a saml web application
**How to reproduce:**
1. Access https://localhost:9443/console
2. Login as super admin
3. Navigate to Develop > Applications > New Application > Web Application > SAML and add a saml application
4. When the application is created navigate to the Access tab > SAML > enable single logout profile
5. Select type as BACK CHANNEL
6. Enter a single logout request url (http://testlogout)
7. Update
**Observation**
Even though it says updated successfully.the single logout request url does not get updated
**Environment information**
Mssql 2019
uniqueid jdbc primary US
IS 5.11.0 beta3 snapshot
ubuntu 16.04
Chrome Version 85.0.4183.121

|
non_test
|
unable to add single logout request url via the console app for a saml web application describe the issue unable to add single logout request url via the console app for a saml web application how to reproduce access login as super admin navigate to develop applications new application web application saml and add a saml application when the application is created navigate to the access tab saml enable single logout profile select type as back channel enter a single logout request url update observation even though it says updated successfully the single logout request url does not get updated environment information mssql uniqueid jdbc primary us is snapshot ubuntu chrome version
| 0
|
32,656
| 7,567,974,909
|
IssuesEvent
|
2018-04-22 15:35:31
|
ccpgames/esi-issues
|
https://api.github.com/repos/ccpgames/esi-issues
|
closed
|
/characters/{character_id}/wallet/ does not return json
|
codegen invalid
|
# ATTENTION: DON'T SUBMIT SSO/CREST/XML RELATED ISSUES
# Bug
/characters/{character_id}/wallet/ does not return json, it returns a plain text body. This breaks swagger codegen as one expects all calls to return json.
### Request
/latest/characters/{character_id}/wallet/
### Response
1234.23
### Expected
{
"balance": 1234.23
}
|
1.0
|
/characters/{character_id}/wallet/ does not return json - # ATTENTION: DON'T SUBMIT SSO/CREST/XML RELATED ISSUES
# Bug
/characters/{character_id}/wallet/ does not return json, it returns a plain text body. This breaks swagger codegen as one expects all calls to return json.
### Request
/latest/characters/{character_id}/wallet/
### Response
1234.23
### Expected
{
"balance": 1234.23
}
|
non_test
|
characters character id wallet does not return json attention don t submit sso crest xml related issues bug characters character id wallet does not return json it returns a plain text body this breaks swagger codegen as one expects all calls to return json request latest characters character id wallet response expected balance
| 0
|
278,729
| 24,171,049,672
|
IssuesEvent
|
2022-09-22 19:15:19
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
closed
|
OutOfMemory: OutOfMemoryError: Java heap space
|
comp:gc test failure
|
https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_extended.functional_x86-64_linux_Nightly_testList_1/353
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk8_j9_extended.functional_x86-64_linux_Nightly_testList_1/353/functional_test_output.tar.gz
```
Test handling of other arguments
Testing: -verbose:dynload,sizes,stack,debug
Test start time: 2022/09/21 22:54:08 Eastern Standard Time
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 22:54:08 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
at java.lang.String.<init>(String.java:942)
at TestSuite.evaluateVariables(TestSuite.java:253)
at TestSuite.expandVariable(TestSuite.java:427)
at TestSuite.evaluateVariables(TestSuite.java:249)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Test.run(Test.java:159)
at TestSuite.runTest(TestSuite.java:84)
at TestIterator.addTest(TestIterator.java:84)
at TestConfigParser$TestConfigDocumentHandler.xmlEndElement(TestConfigParser.java:309)
at com.oti.j9.exclude.XMLParser._scan_element_close(XMLParser.java:370)
at com.oti.j9.exclude.XMLParser._scan_element_or_instruction(XMLParser.java:391)
at com.oti.j9.exclude.XMLParser.parseXML(XMLParser.java:442)
at com.oti.j9.exclude.XMLParser.parse(XMLParser.java:470)
at com.oti.j9.exclude.XMLParser.parse(XMLParser.java:459)
at TestConfigParser.runTests(TestConfigParser.java:78)
at MainTester.main(MainTester.java:106)
cmdLineTester_verbosetest_6_FAILED
```
https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly/367/
No artifacts.
```
Testing: -verbose:class -verbose:Xgccon
Test start time: 2022/09/21 20:32:15 Pacific Standard Time
Running command: "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/openjdkbinary/j2sdk-image/bin/java" -XX:+UseCompressedOops -Xjit -Xgcpolicy:balanced -Xdump -verbose:class -verbose:Xgccon -cp "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/TestConfig/resources:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/testng.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/jcommander.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/jitt.jar" org.testng.TestNG -d "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/output_16638151285423/cmdLineTester_verbosetest_6" "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/testng.xml" -testnames AllocationTest -groups level.extended
Time spent starting: 5 milliseconds
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 20:32:16 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "Thread-38" java.lang.OutOfMemoryError: Java heap space
at TestSuite.evaluateVariables(TestSuite.java:235)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Output.match(Output.java:62)
at Test$StreamMatcher.run(Test.java:497)
Time spent executing: 36971 milliseconds
Test result: PASSED
Testing: -verbose:gc -verbose
Test start time: 2022/09/21 20:32:52 Pacific Standard Time
Running command: "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/openjdkbinary/j2sdk-image/bin/java" -XX:+UseCompressedOops -Xjit -Xgcpolicy:balanced -Xdump -verbose:gc -verbose -cp "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/TestConfig/resources:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/testng.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/jcommander.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/jitt.jar" org.testng.TestNG -d "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/output_16638151285423/cmdLineTester_verbosetest_6" "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/testng.xml" -testnames AllocationTest -groups level.extended
Time spent starting: 6 milliseconds
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 20:32:53 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "Thread-42" java.lang.OutOfMemoryError: Java heap space
at java.util.TreeSet.<init>(TreeSet.java:124)
at TestSuite.evaluateVariables(TestSuite.java:222)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Output.match(Output.java:62)
at Test$StreamMatcher.run(Test.java:497)
***[TEST INFO 2022/09/21 20:47:52] ProcessKiller detected a timeout after 900000 milliseconds!***
INFO: The current OS is 'Mac OS X'. 'Debug on timeout' is currently only supported on Linux.
***[TEST INFO 2022/09/21 20:47:52] executing kill -ABRT 16308***
***[TEST INFO 2022/09/21 20:47:52] kill -ABRT signal sent***
***[TEST INFO 2022/09/21 20:52:52] stdout timed out***
***[TEST INFO 2022/09/21 20:52:52] ABRT timed out***
***[TEST INFO 2022/09/21 20:52:52] executing kill -9 16308***
***[TEST INFO 2022/09/21 20:52:52] kill -9 signal sent***
***[TEST INFO 2022/09/21 20:52:52] ProcessKiller destroy test process after timeout***
Output from test:
[OUT] [IncludeExcludeTestAnnotationTransformer] [INFO] EXCLUDE_FILE environment variable: /Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TestConfig/resources/excludes/latest_exclude_8.txt
[OUT] [IncludeExcludeTestAnnotationTransformer] [INFO] Processing exclude file: /Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TestConfig/resources/excludes/latest_exclude_8.txt
[OUT] ...
[OUT] ... TestNG 6.14.2 by Cédric Beust (cedric@beust.com)
[OUT] ...
[OUT]
[OUT] [Allocation] [INFO] 100000 x ObjectAllocation.allocArrays (0) = 3
[OUT] [Allocation] [INFO] 100000 x ObjectAllocation.allocArrays (100) = 20
...
[ERR] <gc-op id="222" type="copy forward" timems="1.101" contextid="219" timestamp="2022-09-21T20:32:53.501">
[ERR] <memory-copied type="eden" objects="1" bytes="2016" bytesdiscarded="0" />
[ERR] <memory-copied type="other" objects="0" bytes="0" bytesdiscarded="0" />
[ERR] <memory-cardclean objects="0" bytes="0" />
[ERR] <regions eden="32" other="1" />
[ERR] <remembered-set-cleared processed="348" cleared="0" durationms="0.045" />
[ERR] <stringconstants candidates="2825" cleared="0" />
Time spent executing: 1200087 milliseconds
Test result: FAILED
```
[Internal build](http://vmfarm.rtp.raleigh.ibm.com/job_output.php?id=43415257)
[Linux S390 64bit Compressed Pointers] 80 Load_Level_2.mauve.5mins.Mode557
`-XX:+UseCompressedOops -Xgcpolicy:balanced -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:count=0`
Ran 60x grinders, no repeats.
```
j> 02:59:23 20220922 02:59:22 Execution Timer INFO: Execution Timer started
j> 02:59:23 20220922 02:59:23 Thread Control Engine INFO: Starting JobSet Primary
j> 02:59:36 JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/22 02:59:35 - please wait.
...
j> 02:59:36 JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/core.20220922.025935.61402.0001.dmp spotted.
p> 20220922-02:59:36 Attempted to collect /proc/pid/maps and /proc/pid/smaps
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/heapdump.20220922.025935.61402.0002.phd spotted.
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/javacore.20220922.025935.61402.0003.txt spotted.
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/Snap.20220922.025935.61402.0004.trc spotted.
j> 02:59:37 20220922 02:59:36 <Primary|SimpleDriver|gnu.testlet.java.text.DecimalFormatSymbols.DumpDefault11|192|Default Invocant> ERROR: OutOfMemoryError thrown from arbitrary java method public static void gnu.testlet.SingleTestHarness.main(java.lang.String[]) throws java.lang.Exception
j> 02:59:37 Throwable trace:
j> 02:59:37 java.lang.OutOfMemoryError: Java heap space
j> 02:59:37 at java.lang.StringBuffer.toString(StringBuffer.java:1783)
j> 02:59:37 at gnu.testlet.java.text.DecimalFormatSymbols.DumpDefault11.test(Unknown Source)
j> 02:59:37 at gnu.testlet.SingleTestHarness.main(Unknown Source)
j> 02:59:37 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
j> 02:59:37 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:90)
j> 02:59:37 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:55)
j> 02:59:37 at java.lang.reflect.Method.invoke(Method.java:508)
j> 02:59:37 at com.ibm.jtc.test.jltf.workunitadapters.ArbitraryJavaWorkUnit.runWork(ArbitraryJavaWorkUnit.java:445)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.WorkUnitRunner.run(WorkUnitRunner.java:113)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.runWork(PooledRunnable.java:170)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.getAndRunWork(PooledRunnable.java:136)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.run(PooledRunnable.java:93)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.PooledLangThreadFactory$JavaLangPooledThread.run(PooledLangThreadFactory.java:152)
j> 02:59:54 20220922 02:59:53 Runtime State Reporter IMPORTANT: 22705 tests complete, 0 currently running
...
j> 03:03:54 20220922 03:03:53 Runtime State Reporter IMPORTANT: 364405 tests complete, 0 currently running
j> 03:04:24 20220922 03:04:23 Thread Control Engine INFO: Stopping JobSet Primary
```
|
1.0
|
OutOfMemory: OutOfMemoryError: Java heap space - https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_extended.functional_x86-64_linux_Nightly_testList_1/353
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk8_j9_extended.functional_x86-64_linux_Nightly_testList_1/353/functional_test_output.tar.gz
```
Test handling of other arguments
Testing: -verbose:dynload,sizes,stack,debug
Test start time: 2022/09/21 22:54:08 Eastern Standard Time
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 22:54:08 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
at java.lang.String.<init>(String.java:942)
at TestSuite.evaluateVariables(TestSuite.java:253)
at TestSuite.expandVariable(TestSuite.java:427)
at TestSuite.evaluateVariables(TestSuite.java:249)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Test.run(Test.java:159)
at TestSuite.runTest(TestSuite.java:84)
at TestIterator.addTest(TestIterator.java:84)
at TestConfigParser$TestConfigDocumentHandler.xmlEndElement(TestConfigParser.java:309)
at com.oti.j9.exclude.XMLParser._scan_element_close(XMLParser.java:370)
at com.oti.j9.exclude.XMLParser._scan_element_or_instruction(XMLParser.java:391)
at com.oti.j9.exclude.XMLParser.parseXML(XMLParser.java:442)
at com.oti.j9.exclude.XMLParser.parse(XMLParser.java:470)
at com.oti.j9.exclude.XMLParser.parse(XMLParser.java:459)
at TestConfigParser.runTests(TestConfigParser.java:78)
at MainTester.main(MainTester.java:106)
cmdLineTester_verbosetest_6_FAILED
```
https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly/367/
No artifacts.
```
Testing: -verbose:class -verbose:Xgccon
Test start time: 2022/09/21 20:32:15 Pacific Standard Time
Running command: "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/openjdkbinary/j2sdk-image/bin/java" -XX:+UseCompressedOops -Xjit -Xgcpolicy:balanced -Xdump -verbose:class -verbose:Xgccon -cp "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/TestConfig/resources:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/testng.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/jcommander.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/jitt.jar" org.testng.TestNG -d "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/output_16638151285423/cmdLineTester_verbosetest_6" "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/testng.xml" -testnames AllocationTest -groups level.extended
Time spent starting: 5 milliseconds
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 20:32:16 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "Thread-38" java.lang.OutOfMemoryError: Java heap space
at TestSuite.evaluateVariables(TestSuite.java:235)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Output.match(Output.java:62)
at Test$StreamMatcher.run(Test.java:497)
Time spent executing: 36971 milliseconds
Test result: PASSED
Testing: -verbose:gc -verbose
Test start time: 2022/09/21 20:32:52 Pacific Standard Time
Running command: "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/openjdkbinary/j2sdk-image/bin/java" -XX:+UseCompressedOops -Xjit -Xgcpolicy:balanced -Xdump -verbose:gc -verbose -cp "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/TestConfig/resources:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/testng.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/lib/jcommander.jar:/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/jitt.jar" org.testng.TestNG -d "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TKG/output_16638151285423/cmdLineTester_verbosetest_6" "/Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../../jvmtest/functional/JIT_Test/testng.xml" -testnames AllocationTest -groups level.extended
Time spent starting: 6 milliseconds
JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/21 20:32:53 - please wait.
...
JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
Exception in thread "Thread-42" java.lang.OutOfMemoryError: Java heap space
at java.util.TreeSet.<init>(TreeSet.java:124)
at TestSuite.evaluateVariables(TestSuite.java:222)
at TestSuite.evaluateVariables(TestSuite.java:146)
at Output.match(Output.java:62)
at Test$StreamMatcher.run(Test.java:497)
***[TEST INFO 2022/09/21 20:47:52] ProcessKiller detected a timeout after 900000 milliseconds!***
INFO: The current OS is 'Mac OS X'. 'Debug on timeout' is currently only supported on Linux.
***[TEST INFO 2022/09/21 20:47:52] executing kill -ABRT 16308***
***[TEST INFO 2022/09/21 20:47:52] kill -ABRT signal sent***
***[TEST INFO 2022/09/21 20:52:52] stdout timed out***
***[TEST INFO 2022/09/21 20:52:52] ABRT timed out***
***[TEST INFO 2022/09/21 20:52:52] executing kill -9 16308***
***[TEST INFO 2022/09/21 20:52:52] kill -9 signal sent***
***[TEST INFO 2022/09/21 20:52:52] ProcessKiller destroy test process after timeout***
Output from test:
[OUT] [IncludeExcludeTestAnnotationTransformer] [INFO] EXCLUDE_FILE environment variable: /Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TestConfig/resources/excludes/latest_exclude_8.txt
[OUT] [IncludeExcludeTestAnnotationTransformer] [INFO] Processing exclude file: /Users/jenkins/workspace/Test_openjdk8_j9_extended.functional_x86-64_mac_Nightly_testList_0/aqa-tests/TKG/../TestConfig/resources/excludes/latest_exclude_8.txt
[OUT] ...
[OUT] ... TestNG 6.14.2 by Cédric Beust (cedric@beust.com)
[OUT] ...
[OUT]
[OUT] [Allocation] [INFO] 100000 x ObjectAllocation.allocArrays (0) = 3
[OUT] [Allocation] [INFO] 100000 x ObjectAllocation.allocArrays (100) = 20
...
[ERR] <gc-op id="222" type="copy forward" timems="1.101" contextid="219" timestamp="2022-09-21T20:32:53.501">
[ERR] <memory-copied type="eden" objects="1" bytes="2016" bytesdiscarded="0" />
[ERR] <memory-copied type="other" objects="0" bytes="0" bytesdiscarded="0" />
[ERR] <memory-cardclean objects="0" bytes="0" />
[ERR] <regions eden="32" other="1" />
[ERR] <remembered-set-cleared processed="348" cleared="0" durationms="0.045" />
[ERR] <stringconstants candidates="2825" cleared="0" />
Time spent executing: 1200087 milliseconds
Test result: FAILED
```
[Internal build](http://vmfarm.rtp.raleigh.ibm.com/job_output.php?id=43415257)
[Linux S390 64bit Compressed Pointers] 80 Load_Level_2.mauve.5mins.Mode557
`-XX:+UseCompressedOops -Xgcpolicy:balanced -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:count=0`
Ran 60x grinders, no repeats.
```
j> 02:59:23 20220922 02:59:22 Execution Timer INFO: Execution Timer started
j> 02:59:23 20220922 02:59:23 Thread Control Engine INFO: Starting JobSet Primary
j> 02:59:36 JVMDUMP039I Processing dump event "systhrow", detail "java/lang/OutOfMemoryError" at 2022/09/22 02:59:35 - please wait.
...
j> 02:59:36 JVMDUMP013I Processed dump event "systhrow", detail "java/lang/OutOfMemoryError".
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/core.20220922.025935.61402.0001.dmp spotted.
p> 20220922-02:59:36 Attempted to collect /proc/pid/maps and /proc/pid/smaps
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/heapdump.20220922.025935.61402.0002.phd spotted.
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/javacore.20220922.025935.61402.0003.txt spotted.
p> 20220922-02:59:36 Dump /tmp/bld_37077/mauve.5mins/43415257/JLTF-mauve-mode557-20220922-0258/Snap.20220922.025935.61402.0004.trc spotted.
j> 02:59:37 20220922 02:59:36 <Primary|SimpleDriver|gnu.testlet.java.text.DecimalFormatSymbols.DumpDefault11|192|Default Invocant> ERROR: OutOfMemoryError thrown from arbitrary java method public static void gnu.testlet.SingleTestHarness.main(java.lang.String[]) throws java.lang.Exception
j> 02:59:37 Throwable trace:
j> 02:59:37 java.lang.OutOfMemoryError: Java heap space
j> 02:59:37 at java.lang.StringBuffer.toString(StringBuffer.java:1783)
j> 02:59:37 at gnu.testlet.java.text.DecimalFormatSymbols.DumpDefault11.test(Unknown Source)
j> 02:59:37 at gnu.testlet.SingleTestHarness.main(Unknown Source)
j> 02:59:37 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
j> 02:59:37 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:90)
j> 02:59:37 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:55)
j> 02:59:37 at java.lang.reflect.Method.invoke(Method.java:508)
j> 02:59:37 at com.ibm.jtc.test.jltf.workunitadapters.ArbitraryJavaWorkUnit.runWork(ArbitraryJavaWorkUnit.java:445)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.WorkUnitRunner.run(WorkUnitRunner.java:113)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.runWork(PooledRunnable.java:170)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.getAndRunWork(PooledRunnable.java:136)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.pooling.PooledRunnable.run(PooledRunnable.java:93)
j> 02:59:37 at com.ibm.jtc.test.jltf.threads.PooledLangThreadFactory$JavaLangPooledThread.run(PooledLangThreadFactory.java:152)
j> 02:59:54 20220922 02:59:53 Runtime State Reporter IMPORTANT: 22705 tests complete, 0 currently running
...
j> 03:03:54 20220922 03:03:53 Runtime State Reporter IMPORTANT: 364405 tests complete, 0 currently running
j> 03:04:24 20220922 03:04:23 Thread Control Engine INFO: Stopping JobSet Primary
```
|
test
|
outofmemory outofmemoryerror java heap space test handling of other arguments testing verbose dynload sizes stack debug test start time eastern standard time processing dump event systhrow detail java lang outofmemoryerror at please wait processed dump event systhrow detail java lang outofmemoryerror exception in thread main java lang outofmemoryerror java heap space at java lang string string java at testsuite evaluatevariables testsuite java at testsuite expandvariable testsuite java at testsuite evaluatevariables testsuite java at testsuite evaluatevariables testsuite java at test run test java at testsuite runtest testsuite java at testiterator addtest testiterator java at testconfigparser testconfigdocumenthandler xmlendelement testconfigparser java at com oti exclude xmlparser scan element close xmlparser java at com oti exclude xmlparser scan element or instruction xmlparser java at com oti exclude xmlparser parsexml xmlparser java at com oti exclude xmlparser parse xmlparser java at com oti exclude xmlparser parse xmlparser java at testconfigparser runtests testconfigparser java at maintester main maintester java cmdlinetester verbosetest failed no artifacts testing verbose class verbose xgccon test start time pacific standard time running command users jenkins workspace test extended functional mac nightly testlist openjdkbinary image bin java xx usecompressedoops xjit xgcpolicy balanced xdump verbose class verbose xgccon cp users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest testconfig resources users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg lib testng jar users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg lib jcommander jar users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest functional jit test jitt jar org testng testng d users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg output cmdlinetester verbosetest users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest functional jit test testng xml testnames allocationtest groups level extended time spent starting milliseconds processing dump event systhrow detail java lang outofmemoryerror at please wait processed dump event systhrow detail java lang outofmemoryerror exception in thread thread java lang outofmemoryerror java heap space at testsuite evaluatevariables testsuite java at testsuite evaluatevariables testsuite java at output match output java at test streammatcher run test java time spent executing milliseconds test result passed testing verbose gc verbose test start time pacific standard time running command users jenkins workspace test extended functional mac nightly testlist openjdkbinary image bin java xx usecompressedoops xjit xgcpolicy balanced xdump verbose gc verbose cp users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest testconfig resources users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg lib testng jar users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg lib jcommander jar users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest functional jit test jitt jar org testng testng d users jenkins workspace test extended functional mac nightly testlist aqa tests tkg tkg output cmdlinetester verbosetest users jenkins workspace test extended functional mac nightly testlist aqa tests tkg jvmtest functional jit test testng xml testnames allocationtest groups level extended time spent starting milliseconds processing dump event systhrow detail java lang outofmemoryerror at please wait processed dump event systhrow detail java lang outofmemoryerror exception in thread thread java lang outofmemoryerror java heap space at java util treeset treeset java at testsuite evaluatevariables testsuite java at testsuite evaluatevariables testsuite java at output match output java at test streammatcher run test java processkiller detected a timeout after milliseconds info the current os is mac os x debug on timeout is currently only supported on linux executing kill abrt kill abrt signal sent stdout timed out abrt timed out executing kill kill signal sent processkiller destroy test process after timeout output from test exclude file environment variable users jenkins workspace test extended functional mac nightly testlist aqa tests tkg testconfig resources excludes latest exclude txt processing exclude file users jenkins workspace test extended functional mac nightly testlist aqa tests tkg testconfig resources excludes latest exclude txt testng by cédric beust cedric beust com x objectallocation allocarrays x objectallocation allocarrays time spent executing milliseconds test result failed load level mauve xx usecompressedoops xgcpolicy balanced xdebug xrunjdwp transport dt socket address server y onthrow no pkg foo launch echo xjit count ran grinders no repeats j execution timer info execution timer started j thread control engine info starting jobset primary j processing dump event systhrow detail java lang outofmemoryerror at please wait j processed dump event systhrow detail java lang outofmemoryerror p dump tmp bld mauve jltf mauve core dmp spotted p attempted to collect proc pid maps and proc pid smaps p dump tmp bld mauve jltf mauve heapdump phd spotted p dump tmp bld mauve jltf mauve javacore txt spotted p dump tmp bld mauve jltf mauve snap trc spotted j error outofmemoryerror thrown from arbitrary java method public static void gnu testlet singletestharness main java lang string throws java lang exception j throwable trace j java lang outofmemoryerror java heap space j at java lang stringbuffer tostring stringbuffer java j at gnu testlet java text decimalformatsymbols test unknown source j at gnu testlet singletestharness main unknown source j at sun reflect nativemethodaccessorimpl native method j at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java j at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java j at java lang reflect method invoke method java j at com ibm jtc test jltf workunitadapters arbitraryjavaworkunit runwork arbitraryjavaworkunit java j at com ibm jtc test jltf threads workunitrunner run workunitrunner java j at com ibm jtc test jltf threads pooling pooledrunnable runwork pooledrunnable java j at com ibm jtc test jltf threads pooling pooledrunnable getandrunwork pooledrunnable java j at com ibm jtc test jltf threads pooling pooledrunnable run pooledrunnable java j at com ibm jtc test jltf threads pooledlangthreadfactory javalangpooledthread run pooledlangthreadfactory java j runtime state reporter important tests complete currently running j runtime state reporter important tests complete currently running j thread control engine info stopping jobset primary
| 1
|
131,814
| 12,491,161,866
|
IssuesEvent
|
2020-06-01 03:03:46
|
iaincollins/next-auth
|
https://api.github.com/repos/iaincollins/next-auth
|
closed
|
Refactor and document Apple provider
|
documentation enhancement
|
The new Apple provider is great, but it is slightly trickier provider to work with than most and there is some scope for a small amount of refactoring to make it easier to use.
We should also pull in the very helpful documentation in #174 and add it to the website.
Suggestion for refactoring the provider, so it can be an object (as well as string):
```javascript
Providers.Apple({
clientId: process.env.APPLE_ID,
clientSecret: {
appleId: process.env.APPLE_ID,
teamId: process.env.APPLE_TEAM_ID,
privateKey: process.env.APPLE_PRIVATE_KEY,
keyId: process.env.APPLE_KEY_ID,
}
})
```
In `src/providers/apple.js` we could add an option like this:
```javascript
clientSecretCallback: ({appleId, keyId, teamId, privateKey}) => {
return jwt.sign({
iss: teamId,
iat: Math.floor(Date.now() / 1000),
exp: Math.floor(Date.now() / 1000) + ( 86400 * 180 ), // 6 months
aud: 'https://appleid.apple.com',
sub: appleId
},
privateKey,
{
algorithm: 'ES256',
keyid
})
}
```
Refactoring the OAuth callback behaviour to support this doesn't seem like it would be too hard.
_Originally posted by @iaincollins in https://github.com/iaincollins/next-auth/pull/174#issuecomment-634801822_
|
1.0
|
Refactor and document Apple provider - The new Apple provider is great, but it is slightly trickier provider to work with than most and there is some scope for a small amount of refactoring to make it easier to use.
We should also pull in the very helpful documentation in #174 and add it to the website.
Suggestion for refactoring the provider, so it can be an object (as well as string):
```javascript
Providers.Apple({
clientId: process.env.APPLE_ID,
clientSecret: {
appleId: process.env.APPLE_ID,
teamId: process.env.APPLE_TEAM_ID,
privateKey: process.env.APPLE_PRIVATE_KEY,
keyId: process.env.APPLE_KEY_ID,
}
})
```
In `src/providers/apple.js` we could add an option like this:
```javascript
clientSecretCallback: ({appleId, keyId, teamId, privateKey}) => {
return jwt.sign({
iss: teamId,
iat: Math.floor(Date.now() / 1000),
exp: Math.floor(Date.now() / 1000) + ( 86400 * 180 ), // 6 months
aud: 'https://appleid.apple.com',
sub: appleId
},
privateKey,
{
algorithm: 'ES256',
keyid
})
}
```
Refactoring the OAuth callback behaviour to support this doesn't seem like it would be too hard.
_Originally posted by @iaincollins in https://github.com/iaincollins/next-auth/pull/174#issuecomment-634801822_
|
non_test
|
refactor and document apple provider the new apple provider is great but it is slightly trickier provider to work with than most and there is some scope for a small amount of refactoring to make it easier to use we should also pull in the very helpful documentation in and add it to the website suggestion for refactoring the provider so it can be an object as well as string javascript providers apple clientid process env apple id clientsecret appleid process env apple id teamid process env apple team id privatekey process env apple private key keyid process env apple key id in src providers apple js we could add an option like this javascript clientsecretcallback appleid keyid teamid privatekey return jwt sign iss teamid iat math floor date now exp math floor date now months aud sub appleid privatekey algorithm keyid refactoring the oauth callback behaviour to support this doesn t seem like it would be too hard originally posted by iaincollins in
| 0
|
521,530
| 15,110,562,852
|
IssuesEvent
|
2021-02-08 19:23:57
|
LycheeOrg/Lychee
|
https://api.github.com/repos/LycheeOrg/Lychee
|
closed
|
migration from v3 to v4 failed
|
High Priority bug
|
### Detailed description of the problem [REQUIRED]
*Something is wrong with the database migration.
I tried hints from the project main github page, but something went wrong.
All permissions were good. No errors. .env file has been set successfully.
What happend after migration, is a problem with DB.
I am using Mysql, and what I found is
1. After I created a user on the first page, (Lychee suggested), a Users table in my DB had one entry, but the ID was 0 !!! Not 1... But the primary key is there !
2. Album tables in DB are clear, no entries there.
3. Photos in DB are incorrect as well.*
### Steps to reproduce the issue
**Steps to reproduce the behavior:**
Just follow steps. I did nothing more.
One thing I found, in the logs table, that the prefix was probably incorrectly written in the "env" file.
But Ive just done what was suggested during installation in the env file - to put the old prefix (but I had only the default prefix: "lychee_" in my DB), so I wrote it down there, as the file suggested. And if You look now on my screenshots, You will understand what am I talking about.
**Screenshots**







### Output of the diagnostics [REQUIRED]
```
*Diagnostics
-----------
Warning: Dropbox import not working. dropbox_key is empty.
Warning: You may experience problems when uploading a photo of large size. Take a look in the FAQ for details.
Warning: You may experience problems when uploading a photos of large size. Take a look in the FAQ for details.
Warning: You may experience problems when uploading a large amount of photos. Take a look in the FAQ for details.
System Information
------------------
Lychee Version (git): master (a040cbc) - Data not in Cache
DB Version: 4.2.1
composer install: --no-dev
APP_ENV: production
APP_DEBUG: true
System: Linux
PHP Version: 7.4
Max uploaded file size: 20971520
Max post size: 20971520
MySQL Version: 5.7.33
Imagick: 1
Imagick Active: 1
Imagick Version: 1687
GD Version: 2.2.5
Config Information
------------------
version: 040201
check_for_updates: 0
sorting_Photos_col: takestamp
sorting_Photos_order: ASC
sorting_Albums_col: max_takestamp
sorting_Albums_order: ASC
imagick: 1
skip_duplicates: 0
small_max_width: 0
small_max_height: 360
medium_max_width: 1920
medium_max_height: 1080
lang: en
layout: 1
image_overlay: 1
image_overlay_type: desc
default_license: none
compression_quality: 90
full_photo: 1
delete_imported: 0
Mod_Frame: 1
Mod_Frame_refresh: 30
thumb_2x: 1
small_2x: 1
medium_2x: 1
landing_page_enable: 0
landing_owner: John Smith
landing_title: John Smith
landing_subtitle: Cats, Dogs & Humans Photography
landing_facebook: https://www.facebook.com/JohnSmith
landing_flickr: https://www.flickr.com/JohnSmith
landing_twitter: https://www.twitter.com/JohnSmith
landing_instagram: https://instagram.com/JohnSmith
landing_youtube: https://www.youtube.com/JohnSmith
landing_background: dist/cat.jpg
site_title: Lychee v4
site_copyright_enable: 1
site_copyright_begin: 2019
site_copyright_end: 2019
additional_footer_text:
display_social_in_gallery: 0
public_search: 0
SL_enable: 0
SL_for_admin: 0
public_recent: 0
recent_age: 1
public_starred: 0
downloadable: 0
photos_wraparound: 1
map_display: 0
zip64: 1
map_display_public: 0
map_provider: Wikimedia
force_32bit_ids: 0
map_include_subalbums: 0
update_check_every_days: 3
has_exiftool: 0
share_button_visible: 0
import_via_symlink: 0
has_ffmpeg: 0
location_decoding: 0
location_decoding_timeout: 30
location_show: 1
location_show_public: 0
rss_enable: 0
rss_recent_days: 7
rss_max_items: 100
prefer_available_xmp_metadata: 0
editor_enabled: 1
lossless_optimization: 0
swipe_tolerance_x: 150
swipe_tolerance_y: 250
local_takestamp_video_formats: .avi|.mov
log_max_num_line: 1000
unlock_password_photos_with_url_param: 0
nsfw_visible: 1
nsfw_blur: 0
nsfw_warning: 0
nsfw_warning_admin: 0
map_display_direction: 1
album_subtitle_type: oldstyle
```
### Browser and system
OS: Debian 10
Using Firefox: 78.7.0esr (64 bity)
One more thing - migration of tables seem to be OK - look:
```
`**************************************
* Application In Production! *
**************************************
Do you really wish to run this command? (yes/no) [no]:
>
Migration table created successfully.
Migrating: 2014_10_12_000000_create_users_table
Migrated: 2014_10_12_000000_create_users_table (11.87ms)
Migrating: 2018_08_03_110935_create_albums_table
Migrated: 2018_08_03_110935_create_albums_table (27.78ms)
Migrating: 2018_08_03_110936_create_photos_table
Migrated: 2018_08_03_110936_create_photos_table (31.05ms)
Migrating: 2018_08_03_110942_create_configs_table
Migrated: 2018_08_03_110942_create_configs_table (10.05ms)
Migrating: 2018_08_03_111324_create_logs_table
Migrated: 2018_08_03_111324_create_logs_table (7.98ms)
Migrating: 2018_08_10_134924_move_settings
Migrated: 2018_08_10_134924_move_settings (11.47ms)
Migrating: 2018_08_15_102039_move_albums
Migrated: 2018_08_15_102039_move_albums (9.82ms)
Migrating: 2018_08_15_103716_move_photos
Migrated: 2018_08_15_103716_move_photos (3.86ms)
Migrating: 2018_10_30_135411_sharing
Migrated: 2018_10_30_135411_sharing (46.33ms)
Migrating: 2019_02_21_114356_create_pages_table
Migrated: 2019_02_21_114356_create_pages_table (7.66ms)
Migrating: 2019_02_21_114408_create_page_contents_table
Migrated: 2019_02_21_114408_create_page_contents_table (19.70ms)
Migrating: 2019_06_21_180451_create_sym_links_table
Migrated: 2019_06_21_180451_create_sym_links_table (6.60ms)
Migrating: 2019_09_28_171753_config_fix
Migrated: 2019_09_28_171753_config_fix (301.59ms)
Migrating: 2019_09_28_190822_photos_fix
Migrated: 2019_09_28_190822_photos_fix (40.05ms)
Migrating: 2019_10_01_add_livephoto_cols
Migrated: 2019_10_01_add_livephoto_cols (39.94ms)
Migrating: 2019_10_02_1400_config_map_display_public
Migrated: 2019_10_02_1400_config_map_display_public (1.01ms)
Migrating: 2019_10_03_214750_frame_refresh_in_sec
Migrated: 2019_10_03_214750_frame_refresh_in_sec (1.69ms)
Migrating: 2019_10_06_1400_config_map_providers
Migrated: 2019_10_06_1400_config_map_providers (1.39ms)
Migrating: 2019_10_06_152017_add_force_32bit_ids
Migrated: 2019_10_06_152017_add_force_32bit_ids (0.81ms)
Migrating: 2019_10_07_0900_config_map_include_sub_albums
Migrated: 2019_10_07_0900_config_map_include_sub_albums (0.69ms)
Migrating: 2019_10_09_233402_config_map_mod
Migrated: 2019_10_09_233402_config_map_mod (1.00ms)
Migrating: 2019_10_11_093442_config_check_update_every
Migrated: 2019_10_11_093442_config_check_update_every (0.80ms)
Migrating: 2019_12_02_2100_config_exiftool
Migrated: 2019_12_02_2100_config_exiftool (0.64ms)
Migrating: 2019_12_15_0700_add_share_button_visible_option
Migrated: 2019_12_15_0700_add_share_button_visible_option (19.02ms)
Migrating: 2019_12_15_1000_config_check_update_every_cat_fix
Migrated: 2019_12_15_1000_config_check_update_every_cat_fix (3.84ms)
Migrating: 2019_12_25_0600_config_exiftool_ternary
Migrated: 2019_12_25_0600_config_exiftool_ternary (7.32ms)
Migrating: 2020_01_018_2300_config_import_via_symlink
Migrated: 2020_01_018_2300_config_import_via_symlink (1.19ms)
Migrating: 2020_01_04_1200_config_has_ffmpeg
Migrated: 2020_01_04_1200_config_has_ffmpeg (6.19ms)
Migrating: 2020_01_26_1200_config_public_sorting
Migrated: 2020_01_26_1200_config_public_sorting (3.40ms)
Migrating: 2020_01_28_133201_composer_update
Migrated: 2020_01_28_133201_composer_update (1.20ms)
Migrating: 2020_02_14_0600_location_decoding
Migrated: 2020_02_14_0600_location_decoding (19.69ms)
Migrating: 2020_03_11_124417_increase_length_photo_type
Migrated: 2020_03_11_124417_increase_length_photo_type (6.35ms)
Migrating: 2020_03_17_200000_unhide_configs
Migrated: 2020_03_17_200000_unhide_configs (1.81ms)
Migrating: 2020_04_19_122905_bump_version
Migrated: 2020_04_19_122905_bump_version (1.23ms)
Migrating: 2020_04_22_155712_bump_version040002
Migrated: 2020_04_22_155712_bump_version040002 (1.14ms)
Migrating: 2020_04_29_000250_bump_version040003
Migrated: 2020_04_29_000250_bump_version040003 (1.46ms)
Migrating: 2020_05_12_114228_rss
Migrated: 2020_05_12_114228_rss (1.01ms)
Migrating: 2020_05_12_161427_bump_version040005
Migrated: 2020_05_12_161427_bump_version040005 (1.19ms)
Migrating: 2020_05_19_174233_config_prefer_available_xmp_metadata
Migrated: 2020_05_19_174233_config_prefer_available_xmp_metadata (0.86ms)
Migrating: 2020_05_26_135052_bump_version040006
Migrated: 2020_05_26_135052_bump_version040006 (1.58ms)
Migrating: 2020_06_04_104605_config_editor_enabled
Migrated: 2020_06_04_104605_config_editor_enabled (0.74ms)
Migrating: 2020_07_11_104605_config_lossless_optimization
Migrated: 2020_07_11_104605_config_lossless_optimization (1.24ms)
Migrating: 2020_07_11_184605_update_licences
Migrated: 2020_07_11_184605_update_licences (11.82ms)
Migrating: 2020_07_26_085322_config_swipe_tolerance
Migrated: 2020_07_26_085322_config_swipe_tolerance (1.39ms)
Migrating: 2020_07_29_132731_config_local_takestamp
Migrated: 2020_07_29_132731_config_local_takestamp (1.32ms)
Migrating: 2020_08_21_123622_add_smart_tag_album_cols
Migrated: 2020_08_21_123622_add_smart_tag_album_cols (25.71ms)
Migrating: 2020_10_09_130043_bump_version040007
Migrated: 2020_10_09_130043_bump_version040007 (1.62ms)
Migrating: 2020_10_15_104504_add_log_max_num_line
Migrated: 2020_10_15_104504_add_log_max_num_line (0.76ms)
Migrating: 2020_10_15_161346_sort_image_per_album
Migrated: 2020_10_15_161346_sort_image_per_album (25.15ms)
Migrating: 2020_11_12_183345_config_password_url_param_for_smart_album
Migrated: 2020_11_12_183345_config_password_url_param_for_smart_album (1.42ms)
Migrating: 2020_11_19_231553_bump_version040008
Migrated: 2020_11_19_231553_bump_version040008 (1.70ms)
Migrating: 2020_12_12_203153_migrate_admin_user
Migrated: 2020_12_12_203153_migrate_admin_user (6.28ms)
Migrating: 2020_12_12_203831_create_web_authn_tables
Migrated: 2020_12_12_203831_create_web_authn_tables (21.90ms)
Migrating: 2020_12_18_162100_bump_version040009
Migrated: 2020_12_18_162100_bump_version040009 (0.97ms)
Migrating: 2020_12_18_162155_add_nsfw_album
Migrated: 2020_12_18_162155_add_nsfw_album (21.27ms)
Migrating: 2020_12_18_163800_bump_version040010
Migrated: 2020_12_18_163800_bump_version040010 (0.88ms)
Migrating: 2020_12_24_022307_bump_version040100
Migrated: 2020_12_24_022307_bump_version040100 (0.87ms)
Migrating: 2020_12_26_153220_nested_set_for_albums
Migrated: 2020_12_26_153220_nested_set_for_albums (35.60ms)
Migrating: 2021_01_09_163715_remove_max_min_takestamps
Migrated: 2021_01_09_163715_remove_max_min_takestamps (26.06ms)
Migrating: 2021_01_12_122546_bump_version040200
Migrated: 2021_01_12_122546_bump_version040200 (1.04ms)
Migrating: 2021_01_18_103729_add_album_cover
Migrated: 2021_01_18_103729_add_album_cover (13.39ms)
Migrating: 2021_01_20_113912_bump_version040201
Migrated: 2021_01_20_113912_bump_version040201 (0.98ms)
Migrating: 2021_01_24_231904_fix-rotation
Migrated: 2021_01_24_231904_fix-rotation (2.32ms)
Migrating: 2021_01_27_085903_config_map_display_direction
Migrated: 2021_01_27_085903_config_map_display_direction (1.33ms)
Migrating: 2021_01_30_111736_display_takedate
Migrated: 2021_01_30_111736_display_takedate (0.89ms)
Migrating: 2064_12_25_0000_generate_installed_log
Migrated: 2064_12_25_0000_generate_installed_log (1.96ms)
```
I think the problem lies in album tables.
Not everything is being moved...
|
1.0
|
migration from v3 to v4 failed - ### Detailed description of the problem [REQUIRED]
*Something is wrong with the database migration.
I tried hints from the project main github page, but something went wrong.
All permissions were good. No errors. .env file has been set successfully.
What happend after migration, is a problem with DB.
I am using Mysql, and what I found is
1. After I created a user on the first page, (Lychee suggested), a Users table in my DB had one entry, but the ID was 0 !!! Not 1... But the primary key is there !
2. Album tables in DB are clear, no entries there.
3. Photos in DB are incorrect as well.*
### Steps to reproduce the issue
**Steps to reproduce the behavior:**
Just follow steps. I did nothing more.
One thing I found, in the logs table, that the prefix was probably incorrectly written in the "env" file.
But Ive just done what was suggested during installation in the env file - to put the old prefix (but I had only the default prefix: "lychee_" in my DB), so I wrote it down there, as the file suggested. And if You look now on my screenshots, You will understand what am I talking about.
**Screenshots**







### Output of the diagnostics [REQUIRED]
```
*Diagnostics
-----------
Warning: Dropbox import not working. dropbox_key is empty.
Warning: You may experience problems when uploading a photo of large size. Take a look in the FAQ for details.
Warning: You may experience problems when uploading a photos of large size. Take a look in the FAQ for details.
Warning: You may experience problems when uploading a large amount of photos. Take a look in the FAQ for details.
System Information
------------------
Lychee Version (git): master (a040cbc) - Data not in Cache
DB Version: 4.2.1
composer install: --no-dev
APP_ENV: production
APP_DEBUG: true
System: Linux
PHP Version: 7.4
Max uploaded file size: 20971520
Max post size: 20971520
MySQL Version: 5.7.33
Imagick: 1
Imagick Active: 1
Imagick Version: 1687
GD Version: 2.2.5
Config Information
------------------
version: 040201
check_for_updates: 0
sorting_Photos_col: takestamp
sorting_Photos_order: ASC
sorting_Albums_col: max_takestamp
sorting_Albums_order: ASC
imagick: 1
skip_duplicates: 0
small_max_width: 0
small_max_height: 360
medium_max_width: 1920
medium_max_height: 1080
lang: en
layout: 1
image_overlay: 1
image_overlay_type: desc
default_license: none
compression_quality: 90
full_photo: 1
delete_imported: 0
Mod_Frame: 1
Mod_Frame_refresh: 30
thumb_2x: 1
small_2x: 1
medium_2x: 1
landing_page_enable: 0
landing_owner: John Smith
landing_title: John Smith
landing_subtitle: Cats, Dogs & Humans Photography
landing_facebook: https://www.facebook.com/JohnSmith
landing_flickr: https://www.flickr.com/JohnSmith
landing_twitter: https://www.twitter.com/JohnSmith
landing_instagram: https://instagram.com/JohnSmith
landing_youtube: https://www.youtube.com/JohnSmith
landing_background: dist/cat.jpg
site_title: Lychee v4
site_copyright_enable: 1
site_copyright_begin: 2019
site_copyright_end: 2019
additional_footer_text:
display_social_in_gallery: 0
public_search: 0
SL_enable: 0
SL_for_admin: 0
public_recent: 0
recent_age: 1
public_starred: 0
downloadable: 0
photos_wraparound: 1
map_display: 0
zip64: 1
map_display_public: 0
map_provider: Wikimedia
force_32bit_ids: 0
map_include_subalbums: 0
update_check_every_days: 3
has_exiftool: 0
share_button_visible: 0
import_via_symlink: 0
has_ffmpeg: 0
location_decoding: 0
location_decoding_timeout: 30
location_show: 1
location_show_public: 0
rss_enable: 0
rss_recent_days: 7
rss_max_items: 100
prefer_available_xmp_metadata: 0
editor_enabled: 1
lossless_optimization: 0
swipe_tolerance_x: 150
swipe_tolerance_y: 250
local_takestamp_video_formats: .avi|.mov
log_max_num_line: 1000
unlock_password_photos_with_url_param: 0
nsfw_visible: 1
nsfw_blur: 0
nsfw_warning: 0
nsfw_warning_admin: 0
map_display_direction: 1
album_subtitle_type: oldstyle
```
### Browser and system
OS: Debian 10
Using Firefox: 78.7.0esr (64 bity)
One more thing - migration of tables seem to be OK - look:
```
`**************************************
* Application In Production! *
**************************************
Do you really wish to run this command? (yes/no) [no]:
>
Migration table created successfully.
Migrating: 2014_10_12_000000_create_users_table
Migrated: 2014_10_12_000000_create_users_table (11.87ms)
Migrating: 2018_08_03_110935_create_albums_table
Migrated: 2018_08_03_110935_create_albums_table (27.78ms)
Migrating: 2018_08_03_110936_create_photos_table
Migrated: 2018_08_03_110936_create_photos_table (31.05ms)
Migrating: 2018_08_03_110942_create_configs_table
Migrated: 2018_08_03_110942_create_configs_table (10.05ms)
Migrating: 2018_08_03_111324_create_logs_table
Migrated: 2018_08_03_111324_create_logs_table (7.98ms)
Migrating: 2018_08_10_134924_move_settings
Migrated: 2018_08_10_134924_move_settings (11.47ms)
Migrating: 2018_08_15_102039_move_albums
Migrated: 2018_08_15_102039_move_albums (9.82ms)
Migrating: 2018_08_15_103716_move_photos
Migrated: 2018_08_15_103716_move_photos (3.86ms)
Migrating: 2018_10_30_135411_sharing
Migrated: 2018_10_30_135411_sharing (46.33ms)
Migrating: 2019_02_21_114356_create_pages_table
Migrated: 2019_02_21_114356_create_pages_table (7.66ms)
Migrating: 2019_02_21_114408_create_page_contents_table
Migrated: 2019_02_21_114408_create_page_contents_table (19.70ms)
Migrating: 2019_06_21_180451_create_sym_links_table
Migrated: 2019_06_21_180451_create_sym_links_table (6.60ms)
Migrating: 2019_09_28_171753_config_fix
Migrated: 2019_09_28_171753_config_fix (301.59ms)
Migrating: 2019_09_28_190822_photos_fix
Migrated: 2019_09_28_190822_photos_fix (40.05ms)
Migrating: 2019_10_01_add_livephoto_cols
Migrated: 2019_10_01_add_livephoto_cols (39.94ms)
Migrating: 2019_10_02_1400_config_map_display_public
Migrated: 2019_10_02_1400_config_map_display_public (1.01ms)
Migrating: 2019_10_03_214750_frame_refresh_in_sec
Migrated: 2019_10_03_214750_frame_refresh_in_sec (1.69ms)
Migrating: 2019_10_06_1400_config_map_providers
Migrated: 2019_10_06_1400_config_map_providers (1.39ms)
Migrating: 2019_10_06_152017_add_force_32bit_ids
Migrated: 2019_10_06_152017_add_force_32bit_ids (0.81ms)
Migrating: 2019_10_07_0900_config_map_include_sub_albums
Migrated: 2019_10_07_0900_config_map_include_sub_albums (0.69ms)
Migrating: 2019_10_09_233402_config_map_mod
Migrated: 2019_10_09_233402_config_map_mod (1.00ms)
Migrating: 2019_10_11_093442_config_check_update_every
Migrated: 2019_10_11_093442_config_check_update_every (0.80ms)
Migrating: 2019_12_02_2100_config_exiftool
Migrated: 2019_12_02_2100_config_exiftool (0.64ms)
Migrating: 2019_12_15_0700_add_share_button_visible_option
Migrated: 2019_12_15_0700_add_share_button_visible_option (19.02ms)
Migrating: 2019_12_15_1000_config_check_update_every_cat_fix
Migrated: 2019_12_15_1000_config_check_update_every_cat_fix (3.84ms)
Migrating: 2019_12_25_0600_config_exiftool_ternary
Migrated: 2019_12_25_0600_config_exiftool_ternary (7.32ms)
Migrating: 2020_01_018_2300_config_import_via_symlink
Migrated: 2020_01_018_2300_config_import_via_symlink (1.19ms)
Migrating: 2020_01_04_1200_config_has_ffmpeg
Migrated: 2020_01_04_1200_config_has_ffmpeg (6.19ms)
Migrating: 2020_01_26_1200_config_public_sorting
Migrated: 2020_01_26_1200_config_public_sorting (3.40ms)
Migrating: 2020_01_28_133201_composer_update
Migrated: 2020_01_28_133201_composer_update (1.20ms)
Migrating: 2020_02_14_0600_location_decoding
Migrated: 2020_02_14_0600_location_decoding (19.69ms)
Migrating: 2020_03_11_124417_increase_length_photo_type
Migrated: 2020_03_11_124417_increase_length_photo_type (6.35ms)
Migrating: 2020_03_17_200000_unhide_configs
Migrated: 2020_03_17_200000_unhide_configs (1.81ms)
Migrating: 2020_04_19_122905_bump_version
Migrated: 2020_04_19_122905_bump_version (1.23ms)
Migrating: 2020_04_22_155712_bump_version040002
Migrated: 2020_04_22_155712_bump_version040002 (1.14ms)
Migrating: 2020_04_29_000250_bump_version040003
Migrated: 2020_04_29_000250_bump_version040003 (1.46ms)
Migrating: 2020_05_12_114228_rss
Migrated: 2020_05_12_114228_rss (1.01ms)
Migrating: 2020_05_12_161427_bump_version040005
Migrated: 2020_05_12_161427_bump_version040005 (1.19ms)
Migrating: 2020_05_19_174233_config_prefer_available_xmp_metadata
Migrated: 2020_05_19_174233_config_prefer_available_xmp_metadata (0.86ms)
Migrating: 2020_05_26_135052_bump_version040006
Migrated: 2020_05_26_135052_bump_version040006 (1.58ms)
Migrating: 2020_06_04_104605_config_editor_enabled
Migrated: 2020_06_04_104605_config_editor_enabled (0.74ms)
Migrating: 2020_07_11_104605_config_lossless_optimization
Migrated: 2020_07_11_104605_config_lossless_optimization (1.24ms)
Migrating: 2020_07_11_184605_update_licences
Migrated: 2020_07_11_184605_update_licences (11.82ms)
Migrating: 2020_07_26_085322_config_swipe_tolerance
Migrated: 2020_07_26_085322_config_swipe_tolerance (1.39ms)
Migrating: 2020_07_29_132731_config_local_takestamp
Migrated: 2020_07_29_132731_config_local_takestamp (1.32ms)
Migrating: 2020_08_21_123622_add_smart_tag_album_cols
Migrated: 2020_08_21_123622_add_smart_tag_album_cols (25.71ms)
Migrating: 2020_10_09_130043_bump_version040007
Migrated: 2020_10_09_130043_bump_version040007 (1.62ms)
Migrating: 2020_10_15_104504_add_log_max_num_line
Migrated: 2020_10_15_104504_add_log_max_num_line (0.76ms)
Migrating: 2020_10_15_161346_sort_image_per_album
Migrated: 2020_10_15_161346_sort_image_per_album (25.15ms)
Migrating: 2020_11_12_183345_config_password_url_param_for_smart_album
Migrated: 2020_11_12_183345_config_password_url_param_for_smart_album (1.42ms)
Migrating: 2020_11_19_231553_bump_version040008
Migrated: 2020_11_19_231553_bump_version040008 (1.70ms)
Migrating: 2020_12_12_203153_migrate_admin_user
Migrated: 2020_12_12_203153_migrate_admin_user (6.28ms)
Migrating: 2020_12_12_203831_create_web_authn_tables
Migrated: 2020_12_12_203831_create_web_authn_tables (21.90ms)
Migrating: 2020_12_18_162100_bump_version040009
Migrated: 2020_12_18_162100_bump_version040009 (0.97ms)
Migrating: 2020_12_18_162155_add_nsfw_album
Migrated: 2020_12_18_162155_add_nsfw_album (21.27ms)
Migrating: 2020_12_18_163800_bump_version040010
Migrated: 2020_12_18_163800_bump_version040010 (0.88ms)
Migrating: 2020_12_24_022307_bump_version040100
Migrated: 2020_12_24_022307_bump_version040100 (0.87ms)
Migrating: 2020_12_26_153220_nested_set_for_albums
Migrated: 2020_12_26_153220_nested_set_for_albums (35.60ms)
Migrating: 2021_01_09_163715_remove_max_min_takestamps
Migrated: 2021_01_09_163715_remove_max_min_takestamps (26.06ms)
Migrating: 2021_01_12_122546_bump_version040200
Migrated: 2021_01_12_122546_bump_version040200 (1.04ms)
Migrating: 2021_01_18_103729_add_album_cover
Migrated: 2021_01_18_103729_add_album_cover (13.39ms)
Migrating: 2021_01_20_113912_bump_version040201
Migrated: 2021_01_20_113912_bump_version040201 (0.98ms)
Migrating: 2021_01_24_231904_fix-rotation
Migrated: 2021_01_24_231904_fix-rotation (2.32ms)
Migrating: 2021_01_27_085903_config_map_display_direction
Migrated: 2021_01_27_085903_config_map_display_direction (1.33ms)
Migrating: 2021_01_30_111736_display_takedate
Migrated: 2021_01_30_111736_display_takedate (0.89ms)
Migrating: 2064_12_25_0000_generate_installed_log
Migrated: 2064_12_25_0000_generate_installed_log (1.96ms)
```
I think the problem lies in album tables.
Not everything is being moved...
|
non_test
|
migration from to failed detailed description of the problem something is wrong with the database migration i tried hints from the project main github page but something went wrong all permissions were good no errors env file has been set successfully what happend after migration is a problem with db i am using mysql and what i found is after i created a user on the first page lychee suggested a users table in my db had one entry but the id was not but the primary key is there album tables in db are clear no entries there photos in db are incorrect as well steps to reproduce the issue steps to reproduce the behavior just follow steps i did nothing more one thing i found in the logs table that the prefix was probably incorrectly written in the env file but ive just done what was suggested during installation in the env file to put the old prefix but i had only the default prefix lychee in my db so i wrote it down there as the file suggested and if you look now on my screenshots you will understand what am i talking about screenshots output of the diagnostics diagnostics warning dropbox import not working dropbox key is empty warning you may experience problems when uploading a photo of large size take a look in the faq for details warning you may experience problems when uploading a photos of large size take a look in the faq for details warning you may experience problems when uploading a large amount of photos take a look in the faq for details system information lychee version git master data not in cache db version composer install no dev app env production app debug true system linux php version max uploaded file size max post size mysql version imagick imagick active imagick version gd version config information version check for updates sorting photos col takestamp sorting photos order asc sorting albums col max takestamp sorting albums order asc imagick skip duplicates small max width small max height medium max width medium max height lang en layout image overlay image overlay type desc default license none compression quality full photo delete imported mod frame mod frame refresh thumb small medium landing page enable landing owner john smith landing title john smith landing subtitle cats dogs humans photography landing facebook landing flickr landing twitter landing instagram landing youtube landing background dist cat jpg site title lychee site copyright enable site copyright begin site copyright end additional footer text display social in gallery public search sl enable sl for admin public recent recent age public starred downloadable photos wraparound map display map display public map provider wikimedia force ids map include subalbums update check every days has exiftool share button visible import via symlink has ffmpeg location decoding location decoding timeout location show location show public rss enable rss recent days rss max items prefer available xmp metadata editor enabled lossless optimization swipe tolerance x swipe tolerance y local takestamp video formats avi mov log max num line unlock password photos with url param nsfw visible nsfw blur nsfw warning nsfw warning admin map display direction album subtitle type oldstyle browser and system os debian using firefox bity one more thing migration of tables seem to be ok look application in production do you really wish to run this command yes no migration table created successfully migrating create users table migrated create users table migrating create albums table migrated create albums table migrating create photos table migrated create photos table migrating create configs table migrated create configs table migrating create logs table migrated create logs table migrating move settings migrated move settings migrating move albums migrated move albums migrating move photos migrated move photos migrating sharing migrated sharing migrating create pages table migrated create pages table migrating create page contents table migrated create page contents table migrating create sym links table migrated create sym links table migrating config fix migrated config fix migrating photos fix migrated photos fix migrating add livephoto cols migrated add livephoto cols migrating config map display public migrated config map display public migrating frame refresh in sec migrated frame refresh in sec migrating config map providers migrated config map providers migrating add force ids migrated add force ids migrating config map include sub albums migrated config map include sub albums migrating config map mod migrated config map mod migrating config check update every migrated config check update every migrating config exiftool migrated config exiftool migrating add share button visible option migrated add share button visible option migrating config check update every cat fix migrated config check update every cat fix migrating config exiftool ternary migrated config exiftool ternary migrating config import via symlink migrated config import via symlink migrating config has ffmpeg migrated config has ffmpeg migrating config public sorting migrated config public sorting migrating composer update migrated composer update migrating location decoding migrated location decoding migrating increase length photo type migrated increase length photo type migrating unhide configs migrated unhide configs migrating bump version migrated bump version migrating bump migrated bump migrating bump migrated bump migrating rss migrated rss migrating bump migrated bump migrating config prefer available xmp metadata migrated config prefer available xmp metadata migrating bump migrated bump migrating config editor enabled migrated config editor enabled migrating config lossless optimization migrated config lossless optimization migrating update licences migrated update licences migrating config swipe tolerance migrated config swipe tolerance migrating config local takestamp migrated config local takestamp migrating add smart tag album cols migrated add smart tag album cols migrating bump migrated bump migrating add log max num line migrated add log max num line migrating sort image per album migrated sort image per album migrating config password url param for smart album migrated config password url param for smart album migrating bump migrated bump migrating migrate admin user migrated migrate admin user migrating create web authn tables migrated create web authn tables migrating bump migrated bump migrating add nsfw album migrated add nsfw album migrating bump migrated bump migrating bump migrated bump migrating nested set for albums migrated nested set for albums migrating remove max min takestamps migrated remove max min takestamps migrating bump migrated bump migrating add album cover migrated add album cover migrating bump migrated bump migrating fix rotation migrated fix rotation migrating config map display direction migrated config map display direction migrating display takedate migrated display takedate migrating generate installed log migrated generate installed log i think the problem lies in album tables not everything is being moved
| 0
|
32,454
| 4,772,289,344
|
IssuesEvent
|
2016-10-26 20:25:16
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
: TestServerQuery failed under stress
|
Robot test-failure
|
SHA: https://github.com/cockroachdb/cockroach/commits/4d203e83dbc49ab1bd9283c3e15caa06e5288bb8
Stress build found a failed test:
```
I161024 06:32:18.670002 318 storage/engine/rocksdb.go:349 opening in memory rocksdb instance
I161024 06:32:18.684425 318 gossip/gossip.go:236 [n?] initial resolvers: []
W161024 06:32:18.684512 318 gossip/gossip.go:1063 [n?] no resolvers found; use --join to specify a connected node
W161024 06:32:18.690296 318 server/status/runtime.go:116 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I161024 06:32:18.692870 318 server/node.go:422 [n?] store [n0,s0] not bootstrapped
I161024 06:32:18.700093 156 storage/replica_proposal.go:292 [s1,r1/1:/M{in-ax}] new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 410358h32m27.698566326s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0s [physicalTime=2016-10-24 06:32:18.699977265 +0000 UTC]
I161024 06:32:18.702536 318 util/stop/stopper.go:353 stop has been called, stopping or quiescing all running tasks
I161024 06:32:18.703040 318 server/node.go:351 [n?] **** cluster {986d665c-612f-4dea-b335-b40f9135b2cd} has been created
I161024 06:32:18.703139 318 server/node.go:352 [n?] **** add additional nodes by specifying --join=127.0.0.1:41507
I161024 06:32:18.704074 318 gossip/gossip.go:271 [n?] NodeID set to 1
I161024 06:32:18.708736 318 server/node.go:435 [n?] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:0}
I161024 06:32:18.708897 318 server/node.go:320 [n?] node ID 1 initialized
I161024 06:32:18.709036 318 gossip/gossip.go:286 [n?] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41507" > attrs:<> locality:<>
I161024 06:32:18.709702 318 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I161024 06:32:18.710012 318 server/node.go:565 [n?] connecting to gossip network to verify cluster ID...
I161024 06:32:18.710133 318 server/node.go:586 [n?] node connected via gossip and verified as part of cluster {"986d665c-612f-4dea-b335-b40f9135b2cd"}
I161024 06:32:18.710331 440 storage/split_queue.go:103 [n?,split,s1,r1/1:/M{in-ax}] splitting at keys [/Table/11/0 /Table/12/0 /Table/13/0 /Table/14/0]
I161024 06:32:18.710493 318 server/node.go:370 [n?] node=1: started with [[]=] engine(s) and attributes []
I161024 06:32:18.710749 360 sql/event_log.go:94 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41507} Attrs: Locality:} ClusterID:{UUID:986d665c-612f-4dea-b335-b40f9135b2cd} StartedAt:1477290738710157233}
I161024 06:32:18.711655 318 server/server.go:571 [n1] starting https server at 127.0.0.1:51885
I161024 06:32:18.711759 318 server/server.go:572 [n1] starting grpc/postgres server at 127.0.0.1:41507
I161024 06:32:18.711844 318 server/server.go:573 [n1] advertising CockroachDB node at 127.0.0.1:41507
I161024 06:32:18.737203 360 sql/event_log.go:94 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41507} Attrs: Locality:} ClusterID:{UUID:986d665c-612f-4dea-b335-b40f9135b2cd} StartedAt:1477290738710157233}
I161024 06:32:18.773235 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /Table/11 [r2]
I161024 06:32:18.823711 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/12 [r3]
I161024 06:32:18.841218 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/13 [r4]
I161024 06:32:18.866969 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/14 [r5]
W161024 06:32:18.665371 318 util/hlc/hlc.go:167 backward time jump detected (-0.270467 seconds)
I161024 06:32:18.667318 318 util/stop/stopper.go:353 stop has been called, stopping or quiescing all running tasks
I161024 06:32:18.678414 324 kv/transport_race.go:71 transport race promotion: ran 10 iterations on up to 131 requests
server_test.go:101: storage/store.go:2101: rejecting command with timestamp in the future: 1477290738935746931 (270.132915ms ahead)
ERROR: exit status 1
2511 runs completed, 1 failures, over 10m41s
Makefile:128: recipe for target 'stress' failed
make: *** [stress] Error 1
```
|
1.0
|
: TestServerQuery failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/4d203e83dbc49ab1bd9283c3e15caa06e5288bb8
Stress build found a failed test:
```
I161024 06:32:18.670002 318 storage/engine/rocksdb.go:349 opening in memory rocksdb instance
I161024 06:32:18.684425 318 gossip/gossip.go:236 [n?] initial resolvers: []
W161024 06:32:18.684512 318 gossip/gossip.go:1063 [n?] no resolvers found; use --join to specify a connected node
W161024 06:32:18.690296 318 server/status/runtime.go:116 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I161024 06:32:18.692870 318 server/node.go:422 [n?] store [n0,s0] not bootstrapped
I161024 06:32:18.700093 156 storage/replica_proposal.go:292 [s1,r1/1:/M{in-ax}] new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 410358h32m27.698566326s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0s [physicalTime=2016-10-24 06:32:18.699977265 +0000 UTC]
I161024 06:32:18.702536 318 util/stop/stopper.go:353 stop has been called, stopping or quiescing all running tasks
I161024 06:32:18.703040 318 server/node.go:351 [n?] **** cluster {986d665c-612f-4dea-b335-b40f9135b2cd} has been created
I161024 06:32:18.703139 318 server/node.go:352 [n?] **** add additional nodes by specifying --join=127.0.0.1:41507
I161024 06:32:18.704074 318 gossip/gossip.go:271 [n?] NodeID set to 1
I161024 06:32:18.708736 318 server/node.go:435 [n?] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:0}
I161024 06:32:18.708897 318 server/node.go:320 [n?] node ID 1 initialized
I161024 06:32:18.709036 318 gossip/gossip.go:286 [n?] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41507" > attrs:<> locality:<>
I161024 06:32:18.709702 318 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I161024 06:32:18.710012 318 server/node.go:565 [n?] connecting to gossip network to verify cluster ID...
I161024 06:32:18.710133 318 server/node.go:586 [n?] node connected via gossip and verified as part of cluster {"986d665c-612f-4dea-b335-b40f9135b2cd"}
I161024 06:32:18.710331 440 storage/split_queue.go:103 [n?,split,s1,r1/1:/M{in-ax}] splitting at keys [/Table/11/0 /Table/12/0 /Table/13/0 /Table/14/0]
I161024 06:32:18.710493 318 server/node.go:370 [n?] node=1: started with [[]=] engine(s) and attributes []
I161024 06:32:18.710749 360 sql/event_log.go:94 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41507} Attrs: Locality:} ClusterID:{UUID:986d665c-612f-4dea-b335-b40f9135b2cd} StartedAt:1477290738710157233}
I161024 06:32:18.711655 318 server/server.go:571 [n1] starting https server at 127.0.0.1:51885
I161024 06:32:18.711759 318 server/server.go:572 [n1] starting grpc/postgres server at 127.0.0.1:41507
I161024 06:32:18.711844 318 server/server.go:573 [n1] advertising CockroachDB node at 127.0.0.1:41507
I161024 06:32:18.737203 360 sql/event_log.go:94 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41507} Attrs: Locality:} ClusterID:{UUID:986d665c-612f-4dea-b335-b40f9135b2cd} StartedAt:1477290738710157233}
I161024 06:32:18.773235 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /Table/11 [r2]
I161024 06:32:18.823711 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/12 [r3]
I161024 06:32:18.841218 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/13 [r4]
I161024 06:32:18.866969 440 storage/replica_command.go:2354 [n1,split,s1,r1/1:/{Min-Table/11}] initiating a split of this range at key /Table/14 [r5]
W161024 06:32:18.665371 318 util/hlc/hlc.go:167 backward time jump detected (-0.270467 seconds)
I161024 06:32:18.667318 318 util/stop/stopper.go:353 stop has been called, stopping or quiescing all running tasks
I161024 06:32:18.678414 324 kv/transport_race.go:71 transport race promotion: ran 10 iterations on up to 131 requests
server_test.go:101: storage/store.go:2101: rejecting command with timestamp in the future: 1477290738935746931 (270.132915ms ahead)
ERROR: exit status 1
2511 runs completed, 1 failures, over 10m41s
Makefile:128: recipe for target 'stress' failed
make: *** [stress] Error 1
```
|
test
|
testserverquery failed under stress sha stress build found a failed test storage engine rocksdb go opening in memory rocksdb instance gossip gossip go initial resolvers gossip gossip go no resolvers found use join to specify a connected node server status runtime go could not parse build timestamp parsing time as cannot parse as server node go store not bootstrapped storage replica proposal go new range lease replica utc following replica utc util stop stopper go stop has been called stopping or quiescing all running tasks server node go cluster has been created server node go add additional nodes by specifying join gossip gossip go nodeid set to server node go initialized store capacity available rangecount server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster storage split queue go splitting at keys server node go node started with engine s and attributes sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid uuid startedat server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid uuid startedat storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table util hlc hlc go backward time jump detected seconds util stop stopper go stop has been called stopping or quiescing all running tasks kv transport race go transport race promotion ran iterations on up to requests server test go storage store go rejecting command with timestamp in the future ahead error exit status runs completed failures over makefile recipe for target stress failed make error
| 1
|
54,626
| 7,902,273,521
|
IssuesEvent
|
2018-07-01 01:29:47
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
Document documentation
|
documentation
|
The documentation gets generated and rebuilt on every commit and release
"Latest"/HEAD is at
https://storage.googleapis.com/skaffold/builds/latest/docs/index.html
We should add this to the README
|
1.0
|
Document documentation - The documentation gets generated and rebuilt on every commit and release
"Latest"/HEAD is at
https://storage.googleapis.com/skaffold/builds/latest/docs/index.html
We should add this to the README
|
non_test
|
document documentation the documentation gets generated and rebuilt on every commit and release latest head is at we should add this to the readme
| 0
|
27,275
| 21,530,132,650
|
IssuesEvent
|
2022-04-28 23:17:01
|
aodn/aodn-portal
|
https://api.github.com/repos/aodn/aodn-portal
|
closed
|
Get Feature Info is not displaying Depth or lat/long at click point
|
T1 - IMOS T2 - O&M - Continuous Improvement T3 Information Infrastructure
|
This issue tracker is only for AODN Portal issues.
### Steps to reproduce the issue
1. Add any collection to the portal
2.Click any point on step 2
### Expected behaviour
In the get feature window can see the depth at the click point and the lat long at which the point is
### Actual behaviour
Just says loading
### What version of the aodn portal are you using?
4.42.85 Systest - it is in prod - does this come from a Geoserver currently at this point in time 28/4/22 1532 geoserver -123 is showing as down in nagios but static is up...
|
1.0
|
Get Feature Info is not displaying Depth or lat/long at click point - This issue tracker is only for AODN Portal issues.
### Steps to reproduce the issue
1. Add any collection to the portal
2.Click any point on step 2
### Expected behaviour
In the get feature window can see the depth at the click point and the lat long at which the point is
### Actual behaviour
Just says loading
### What version of the aodn portal are you using?
4.42.85 Systest - it is in prod - does this come from a Geoserver currently at this point in time 28/4/22 1532 geoserver -123 is showing as down in nagios but static is up...
|
non_test
|
get feature info is not displaying depth or lat long at click point this issue tracker is only for aodn portal issues steps to reproduce the issue add any collection to the portal click any point on step expected behaviour in the get feature window can see the depth at the click point and the lat long at which the point is actual behaviour just says loading what version of the aodn portal are you using systest it is in prod does this come from a geoserver currently at this point in time geoserver is showing as down in nagios but static is up
| 0
|
146,212
| 19,393,966,044
|
IssuesEvent
|
2021-12-18 01:43:52
|
SmartBear/readyapi4j
|
https://api.github.com/repos/SmartBear/readyapi4j
|
opened
|
CVE-2021-4104 (High) detected in log4j-1.2.14.jar
|
security vulnerability
|
## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.14.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: readyapi4j/modules/cucumber/modules/cucumber4oas/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar</p>
<p>
Dependency Hierarchy:
- readyapi4j-maven-plugin-1.0.0-SNAPSHOT.jar (Root Library)
- readyapi4j-facade-1.0.0-SNAPSHOT.jar
- readyapi4j-local-1.0.0-SNAPSHOT.jar
- soapui-testserver-api-5.5.0.jar
- soapui-5.5.0.jar
- :x: **log4j-1.2.14.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Java logging library Apache Log4j in version 1.x. JMSAppender in Log4j 1.x is vulnerable to deserialization of untrusted data. This allows a remote attacker to execute code on the server if the deployed application is configured to use JMSAppender and to the attacker's JMS Broker.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.14","packageFilePaths":["/modules/cucumber/modules/cucumber4oas/pom.xml","/modules/samples/groovy/pom.xml","/modules/local/pom.xml","/modules/maven-plugin/pom.xml","/modules/cucumber/modules/runner/pom.xml","/modules/cucumber/modules/samples/pom.xml","/modules/cucumber/modules/core/pom.xml","/modules/cucumber/modules/stepdefs/pom.xml","/modules/samples/java/pom.xml","/modules/facade/pom.xml","/modules/maven-plugin-tester/pom.xml","/modules/cucumber/modules/studio-runner/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear.readyapi:readyapi4j-maven-plugin:1.0.0-SNAPSHOT;com.smartbear.readyapi:readyapi4j-facade:1.0.0-SNAPSHOT;com.smartbear.readyapi:readyapi4j-local:1.0.0-SNAPSHOT;com.smartbear.soapui:soapui-testserver-api:5.5.0;com.smartbear.soapui:soapui:5.5.0;log4j:log4j:1.2.14","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"A flaw was found in the Java logging library Apache Log4j in version 1.x. JMSAppender in Log4j 1.x is vulnerable to deserialization of untrusted data. This allows a remote attacker to execute code on the server if the deployed application is configured to use JMSAppender and to the attacker\u0027s JMS Broker.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-4104 (High) detected in log4j-1.2.14.jar - ## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.14.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: readyapi4j/modules/cucumber/modules/cucumber4oas/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.14/log4j-1.2.14.jar</p>
<p>
Dependency Hierarchy:
- readyapi4j-maven-plugin-1.0.0-SNAPSHOT.jar (Root Library)
- readyapi4j-facade-1.0.0-SNAPSHOT.jar
- readyapi4j-local-1.0.0-SNAPSHOT.jar
- soapui-testserver-api-5.5.0.jar
- soapui-5.5.0.jar
- :x: **log4j-1.2.14.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Java logging library Apache Log4j in version 1.x. JMSAppender in Log4j 1.x is vulnerable to deserialization of untrusted data. This allows a remote attacker to execute code on the server if the deployed application is configured to use JMSAppender and to the attacker's JMS Broker.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.14","packageFilePaths":["/modules/cucumber/modules/cucumber4oas/pom.xml","/modules/samples/groovy/pom.xml","/modules/local/pom.xml","/modules/maven-plugin/pom.xml","/modules/cucumber/modules/runner/pom.xml","/modules/cucumber/modules/samples/pom.xml","/modules/cucumber/modules/core/pom.xml","/modules/cucumber/modules/stepdefs/pom.xml","/modules/samples/java/pom.xml","/modules/facade/pom.xml","/modules/maven-plugin-tester/pom.xml","/modules/cucumber/modules/studio-runner/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear.readyapi:readyapi4j-maven-plugin:1.0.0-SNAPSHOT;com.smartbear.readyapi:readyapi4j-facade:1.0.0-SNAPSHOT;com.smartbear.readyapi:readyapi4j-local:1.0.0-SNAPSHOT;com.smartbear.soapui:soapui-testserver-api:5.5.0;com.smartbear.soapui:soapui:5.5.0;log4j:log4j:1.2.14","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"A flaw was found in the Java logging library Apache Log4j in version 1.x. JMSAppender in Log4j 1.x is vulnerable to deserialization of untrusted data. This allows a remote attacker to execute code on the server if the deployed application is configured to use JMSAppender and to the attacker\u0027s JMS Broker.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high detected in jar cve high severity vulnerability vulnerable library jar library home page a href path to dependency file modules cucumber modules pom xml path to vulnerable library home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar dependency hierarchy maven plugin snapshot jar root library facade snapshot jar local snapshot jar soapui testserver api jar soapui jar x jar vulnerable library found in base branch master vulnerability details a flaw was found in the java logging library apache in version x jmsappender in x is vulnerable to deserialization of untrusted data this allows a remote attacker to execute code on the server if the deployed application is configured to use jmsappender and to the attacker s jms broker publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear readyapi maven plugin snapshot com smartbear readyapi facade snapshot com smartbear readyapi local snapshot com smartbear soapui soapui testserver api com smartbear soapui soapui isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was found in the java logging library apache in version x jmsappender in x is vulnerable to deserialization of untrusted data this allows a remote attacker to execute code on the server if the deployed application is configured to use jmsappender and to the attacker jms broker vulnerabilityurl
| 0
|
213,503
| 16,525,362,714
|
IssuesEvent
|
2021-05-26 19:20:42
|
puppetlabs/pdkgo
|
https://api.github.com/repos/puppetlabs/pdkgo
|
opened
|
Template authoring tips
|
documentation
|
Add to initial doc a set of authoring tips:
- use pct new puppet-content-template to get started
- install go-template vscode extension for syntax highlighting
|
1.0
|
Template authoring tips - Add to initial doc a set of authoring tips:
- use pct new puppet-content-template to get started
- install go-template vscode extension for syntax highlighting
|
non_test
|
template authoring tips add to initial doc a set of authoring tips use pct new puppet content template to get started install go template vscode extension for syntax highlighting
| 0
|
50,046
| 6,050,870,976
|
IssuesEvent
|
2017-06-12 22:07:10
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed tests on master: Example-ORMs/TestSQLAlchemy, Example-ORMs/TestSQLAlchemy/FirstRun, Example-ORMs/TestSQLAlchemy/SecondRun, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order
|
Robot test-failure
|
The following tests appear to have failed:
[#270537](https://teamcity.cockroachdb.com/viewLog.html?buildId=270537):
```
--- FAIL: Example-ORMs/TestSQLAlchemy (183.940s)
------- Stderr: -------
2017/06/12 22:01:12 process 13640 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver461633119 --listening-url-file=/tmp/cockroach-testserver461633119/listen-url
--- FAIL: Example-ORMs/TestSQLAlchemy/FirstRun (180.570s)
main_test.go:160: Get http://localhost:6543/ping: dial tcp [::1]:6543: getsockopt: connection refused
------- Stderr: -------
2017/06/12 22:01:12 signal: killed
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun (1.800s)
null
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s)
null
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.040s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.040s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.050s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed tests on master: Example-ORMs/TestSQLAlchemy, Example-ORMs/TestSQLAlchemy/FirstRun, Example-ORMs/TestSQLAlchemy/SecondRun, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers, Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order - The following tests appear to have failed:
[#270537](https://teamcity.cockroachdb.com/viewLog.html?buildId=270537):
```
--- FAIL: Example-ORMs/TestSQLAlchemy (183.940s)
------- Stderr: -------
2017/06/12 22:01:12 process 13640 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver461633119 --listening-url-file=/tmp/cockroach-testserver461633119/listen-url
--- FAIL: Example-ORMs/TestSQLAlchemy/FirstRun (180.570s)
main_test.go:160: Get http://localhost:6543/ping: dial tcp [::1]:6543: getsockopt: connection refused
------- Stderr: -------
2017/06/12 22:01:12 signal: killed
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun (1.800s)
null
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s)
null
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.040s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.040s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
--- FAIL: Example-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.050s)
test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found []
```
Please assign, take a look and update the issue accordingly.
|
test
|
teamcity failed tests on master example orms testsqlalchemy example orms testsqlalchemy firstrun example orms testsqlalchemy secondrun example orms testsqlalchemy secondrun retrievefromapiafterrestart example orms testsqlalchemy secondrun retrievefromapiafterrestart products example orms testsqlalchemy secondrun retrievefromapiafterrestart customers example orms testsqlalchemy secondrun retrievefromapiafterrestart order the following tests appear to have failed fail example orms testsqlalchemy stderr process started cockroach start logtostderr insecure host localhost port http port store tmp cockroach listening url file tmp cockroach listen url fail example orms testsqlalchemy firstrun main test go get dial tcp getsockopt connection refused stderr signal killed fail example orms testsqlalchemy secondrun null fail example orms testsqlalchemy secondrun retrievefromapiafterrestart null fail example orms testsqlalchemy secondrun retrievefromapiafterrestart products test driver go expecting products from api after creation to be found fail example orms testsqlalchemy secondrun retrievefromapiafterrestart customers test driver go expecting products from api after creation to be found fail example orms testsqlalchemy secondrun retrievefromapiafterrestart order test driver go expecting products from api after creation to be found please assign take a look and update the issue accordingly
| 1
|
20,042
| 14,945,029,222
|
IssuesEvent
|
2021-01-26 02:58:17
|
w3c/coga
|
https://api.github.com/repos/w3c/coga
|
closed
|
Editorial suggestions in Introduction from Wordsmithing Meeting
|
content-usable discuss
|
1. Consider changing "Making websites and applications that are **friendly** for people with cognitive impairments affects every part of design and development." to "Making websites and applications that are **usable** for people with cognitive impairments affects every part of design and development." Also discuss "to" vs "for". (NOTE: This is also in the Design Guide Introduction)
2. Consider whether "Often the issues that affect people with cognitive and learning disabilities include design, context, structure, language, usability, and other factors that are difficult to include in general guidelines." should be presented in list form.
3. Consider removing the bold from: Some design patterns create barriers for people with disabilities. The patterns presented in this document have been designed to avoid such barriers for people with cognitive and learning disabilities. While this guidance may improve usability for all, these patterns are essential for some people **with cognitive and learning impairments** to be able to use content independently. (NOTE: This is also in the Design Guide Introduction)
4. Consider changing "The Objectives and Patterns build on the:" to "The Objectives and Patterns build on prior work" or "The Objective and Patterns build on prior research by the COGA Task Force."
5. Consider switching the paragraph "Some design patterns create..." with "The objectives and patterns..."
|
True
|
Editorial suggestions in Introduction from Wordsmithing Meeting - 1. Consider changing "Making websites and applications that are **friendly** for people with cognitive impairments affects every part of design and development." to "Making websites and applications that are **usable** for people with cognitive impairments affects every part of design and development." Also discuss "to" vs "for". (NOTE: This is also in the Design Guide Introduction)
2. Consider whether "Often the issues that affect people with cognitive and learning disabilities include design, context, structure, language, usability, and other factors that are difficult to include in general guidelines." should be presented in list form.
3. Consider removing the bold from: Some design patterns create barriers for people with disabilities. The patterns presented in this document have been designed to avoid such barriers for people with cognitive and learning disabilities. While this guidance may improve usability for all, these patterns are essential for some people **with cognitive and learning impairments** to be able to use content independently. (NOTE: This is also in the Design Guide Introduction)
4. Consider changing "The Objectives and Patterns build on the:" to "The Objectives and Patterns build on prior work" or "The Objective and Patterns build on prior research by the COGA Task Force."
5. Consider switching the paragraph "Some design patterns create..." with "The objectives and patterns..."
|
non_test
|
editorial suggestions in introduction from wordsmithing meeting consider changing making websites and applications that are friendly for people with cognitive impairments affects every part of design and development to making websites and applications that are usable for people with cognitive impairments affects every part of design and development also discuss to vs for note this is also in the design guide introduction consider whether often the issues that affect people with cognitive and learning disabilities include design context structure language usability and other factors that are difficult to include in general guidelines should be presented in list form consider removing the bold from some design patterns create barriers for people with disabilities the patterns presented in this document have been designed to avoid such barriers for people with cognitive and learning disabilities while this guidance may improve usability for all these patterns are essential for some people with cognitive and learning impairments to be able to use content independently note this is also in the design guide introduction consider changing the objectives and patterns build on the to the objectives and patterns build on prior work or the objective and patterns build on prior research by the coga task force consider switching the paragraph some design patterns create with the objectives and patterns
| 0
|
87,365
| 10,544,335,751
|
IssuesEvent
|
2019-10-02 16:42:24
|
hexatomic/hexatomic
|
https://api.github.com/repos/hexatomic/hexatomic
|
opened
|
Document how Hexatomic is modularized
|
documentation
|
E.g., when to create a new Eclipse plugin (more general software engineering, SOC, etc.), when to create a new Eclipse feature, e.g., every editor/view gets its own feature (apart from the one in #40).
|
1.0
|
Document how Hexatomic is modularized - E.g., when to create a new Eclipse plugin (more general software engineering, SOC, etc.), when to create a new Eclipse feature, e.g., every editor/view gets its own feature (apart from the one in #40).
|
non_test
|
document how hexatomic is modularized e g when to create a new eclipse plugin more general software engineering soc etc when to create a new eclipse feature e g every editor view gets its own feature apart from the one in
| 0
|
810,300
| 30,235,346,600
|
IssuesEvent
|
2023-07-06 09:49:41
|
horizon-efrei/HorizonBot
|
https://api.github.com/repos/horizon-efrei/HorizonBot
|
closed
|
Expliquer en DM les raisons de refus sur un menu de réaction-rôle
|
priority: low type: feature difficulty: easy status: approved
|
<!-- Décrivez votre idée en rentrant le plus possible dans les détails. -->
Permettre aux utilisateurs de changer de rôles en dm pu de leur indiquer ce qui ne va pas (blocage)
|
1.0
|
Expliquer en DM les raisons de refus sur un menu de réaction-rôle - <!-- Décrivez votre idée en rentrant le plus possible dans les détails. -->
Permettre aux utilisateurs de changer de rôles en dm pu de leur indiquer ce qui ne va pas (blocage)
|
non_test
|
expliquer en dm les raisons de refus sur un menu de réaction rôle permettre aux utilisateurs de changer de rôles en dm pu de leur indiquer ce qui ne va pas blocage
| 0
|
19,217
| 10,337,472,295
|
IssuesEvent
|
2019-09-03 14:55:40
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
CSP breaks OIDC authentication flow
|
Feature:CSP Feature:Security/Authentication Team:Security
|
As noted by @azasypkin in https://github.com/elastic/kibana/pull/43553#discussion_r319444103, the CSP change from using `nonce` to `'self'` breaks the OIDC authentication flow which requires an inline script. I have not yet verified this manually.
I'm opening this issue to discuss the blocker status of this for the 7.4 release as well as how we should go about fixing this. It seems to me this will be a blocker.
For 7.4, I propose that we revert #43553 from the 7.4 branch while working on a solution in master and 7.x.
In terms of the solution, it's not clear to me how a nonce is strictly safer than simply allowing inline scripts since any malicious script can read the nonce off the page. That said, we may gain some protection from any malicious dependency that is just targeting any page that does not use a nonce policy.
|
True
|
CSP breaks OIDC authentication flow - As noted by @azasypkin in https://github.com/elastic/kibana/pull/43553#discussion_r319444103, the CSP change from using `nonce` to `'self'` breaks the OIDC authentication flow which requires an inline script. I have not yet verified this manually.
I'm opening this issue to discuss the blocker status of this for the 7.4 release as well as how we should go about fixing this. It seems to me this will be a blocker.
For 7.4, I propose that we revert #43553 from the 7.4 branch while working on a solution in master and 7.x.
In terms of the solution, it's not clear to me how a nonce is strictly safer than simply allowing inline scripts since any malicious script can read the nonce off the page. That said, we may gain some protection from any malicious dependency that is just targeting any page that does not use a nonce policy.
|
non_test
|
csp breaks oidc authentication flow as noted by azasypkin in the csp change from using nonce to self breaks the oidc authentication flow which requires an inline script i have not yet verified this manually i m opening this issue to discuss the blocker status of this for the release as well as how we should go about fixing this it seems to me this will be a blocker for i propose that we revert from the branch while working on a solution in master and x in terms of the solution it s not clear to me how a nonce is strictly safer than simply allowing inline scripts since any malicious script can read the nonce off the page that said we may gain some protection from any malicious dependency that is just targeting any page that does not use a nonce policy
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.