Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
23,169 | 15,876,515,460 | IssuesEvent | 2021-04-09 08:28:14 | ManimCommunity/manim | https://api.github.com/repos/ManimCommunity/manim | closed | Nondeterministic build failures in `Tex('The horse does not eat cucumber salad.')` | infrastructure | ## Description of bug / unexpected behavior
<!-- Add a clear and concise description of the problem you encountered. -->
As I've been figuring out the test coverage toolset for Manim, I've done quite a lot of automated builds on a copied (not forked) repository (https://github.com/MrMallIronmaker/manim-cov/). For some reason, occasionally one of the Ubuntu builds fails while I've been playing around with coverage settings. Whether it's 3.7, 3.8, or 3.9 is apparently random.
All of them fail in the Tex doctest.
## Expected behavior
It should pass the test.
## How to reproduce the issue
<!-- Provide a piece of code illustrating the undesired behavior. -->
<details><summary>Code for reproducing the problem</summary>
In theory the problematic code is here:
```py
Tex('The horse does not eat cucumber salad.')
```
However, I've not been able to replicate this issue
</details>
## Additional media files
<!-- Paste in the files manim produced on rendering the code above. Note that GitHub doesn't allow posting videos, so you may need to convert it to a GIF or use the `-i` rendering option. -->
<details><summary>Build output</summary>
```
Run poetry run pytest --cov-append --doctest-modules manim
poetry run pytest --cov-append --doctest-modules manim
shell: /usr/bin/bash -e {0}
env:
POETRY_VIRTUALENVS_CREATE: false
pythonLocation: /opt/hostedtoolcache/Python/3.8.8/x64
LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.8/x64/lib
Skipping virtualenv creation, as specified in config file.
============================= test session starts ==============================
platform linux -- Python 3.8.8, pytest-6.2.2, py-1.10.0, pluggy-0.13.1
rootdir: /home/runner/work/manim-cov/manim-cov, configfile: pyproject.toml
plugins: cov-2.11.1
collected 30 items
manim/_config/__init__.py . [ 3%]
manim/_config/utils.py ..
## Logs
<details><summary>Terminal output</summary>
<!-- Add "-v DEBUG" when calling manim to generate more detailed logs -->
```
PASTE HERE OR PROVIDE LINK TO https://pastebin.com/ OR SIMILAR
```
<!-- Insert screenshots here (only when absolutely necessary, we prefer copy/pasted output!) -->
</details>
## System specifications
<details><summary>System Details</summary>
- OS (with version, e.g Windows 10 v2004 or macOS 10.15 (Catalina)):
- RAM:
- Python version (`python/py/python3 --version`):
- Installed modules (provide output from `pip list`):
```
PASTE HERE
```
</details>
<details><summary>LaTeX details</summary>
+ LaTeX distribution (e.g. TeX Live 2020):
+ Installed LaTeX packages:
<!-- output of `tlmgr list --only-installed` for TeX Live or a screenshot of the Packages page for MikTeX -->
</details>
<details><summary>FFMPEG</summary>
Output of `ffmpeg -version`:
```
PASTE HERE
```
</details>
## Additional comments
<!-- Add further context that you think might be relevant for this issue here. -->
[ 10%]
manim/animation/animation.py . [ 13%]
manim/mobject/geometry.py ......... [ 43%]
manim/mobject/mobject.py ...... [ 63%]
manim/mobject/svg/tex_mobject.py ..F [ 73%]
manim/mobject/svg/text_mobject.py ... [ 83%]
manim/mobject/types/vectorized_mobject.py .... [ 96%]
manim/utils/color.py . [100%]
=================================== FAILURES ===================================
_________________ [doctest] manim.mobject.svg.tex_mobject.Tex __________________
491 A string compiled with LaTeX in normal mode.
492
493 Tests
494 -----
495
496 Check whether writing a LaTeX string works::
497
498 >>> Tex('The horse does not eat cucumber salad.')
UNEXPECTED EXCEPTION: IndexError('list index out of range')
Traceback (most recent call last):
File "/opt/hostedtoolcache/Python/3.8.8/x64/lib/python3.8/doctest.py", line 1336, in __run
exec(compile(example.source, filename, "single",
File "<doctest manim.mobject.svg.tex_mobject.Tex[0]>", line 1, in <module>
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 506, in __init__
MathTex.__init__(
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 392, in __init__
self.break_up_by_substrings()
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 434, in break_up_by_substrings
sub_tex_mob.move_to(self.submobjects[last_submob_index], RIGHT)
IndexError: list index out of range
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py:498: UnexpectedException
----------------------------- Captured stdout call -----------------------------
INFO Writing "The horse does not tex_file_writing.py:81
eat cucumber salad." to medi
a/Tex/3ecc83aec1683253.tex
------------------------------ Captured log call -------------------------------
INFO manim:tex_file_writing.py:81 Writing "The horse does not eat cucumber salad." to media/Tex/3ecc83aec1683253.tex
=============================== warnings summary ===============================
manim/mobject/mobject.py::manim.mobject.mobject.Mobject.set
<doctest manim.mobject.mobject.Mobject.set[1]>:1: DeprecationWarning: This method is not guaranteed to stay around. Please prefer setting the attribute normally or with Mobject.set().
manim/mobject/mobject.py::manim.mobject.mobject.Mobject.set
<doctest manim.mobject.mobject.Mobject.set[2]>:1: DeprecationWarning: This method is not guaranteed to stay around. Please prefer getting the attribute normally.
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-84, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-104, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-101, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-111, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-114, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-115, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-100, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-110, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-116, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-97, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-99, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-117, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-109, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-98, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-108, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-46, which is not recognized
warnings.warn(warning_text)
-- Docs: https://docs.pytest.org/en/stable/warnings.html
----------- coverage: platform linux, python 3.8.8-final-0 -----------
Coverage XML written to file coverage.xml
=========================== short test summary info ============================
FAILED manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
================== 1 failed, 29 passed, 18 warnings in 6.56s ===================
Error: Process completed with exit code 1.
```
</details> | 1.0 | Nondeterministic build failures in `Tex('The horse does not eat cucumber salad.')` - ## Description of bug / unexpected behavior
<!-- Add a clear and concise description of the problem you encountered. -->
As I've been figuring out the test coverage toolset for Manim, I've done quite a lot of automated builds on a copied (not forked) repository (https://github.com/MrMallIronmaker/manim-cov/). For some reason, occasionally one of the Ubuntu builds fails while I've been playing around with coverage settings. Whether it's 3.7, 3.8, or 3.9 is apparently random.
All of them fail in the Tex doctest.
## Expected behavior
It should pass the test.
## How to reproduce the issue
<!-- Provide a piece of code illustrating the undesired behavior. -->
<details><summary>Code for reproducing the problem</summary>
In theory the problematic code is here:
```py
Tex('The horse does not eat cucumber salad.')
```
However, I've not been able to replicate this issue
</details>
## Additional media files
<!-- Paste in the files manim produced on rendering the code above. Note that GitHub doesn't allow posting videos, so you may need to convert it to a GIF or use the `-i` rendering option. -->
<details><summary>Build output</summary>
```
Run poetry run pytest --cov-append --doctest-modules manim
poetry run pytest --cov-append --doctest-modules manim
shell: /usr/bin/bash -e {0}
env:
POETRY_VIRTUALENVS_CREATE: false
pythonLocation: /opt/hostedtoolcache/Python/3.8.8/x64
LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.8/x64/lib
Skipping virtualenv creation, as specified in config file.
============================= test session starts ==============================
platform linux -- Python 3.8.8, pytest-6.2.2, py-1.10.0, pluggy-0.13.1
rootdir: /home/runner/work/manim-cov/manim-cov, configfile: pyproject.toml
plugins: cov-2.11.1
collected 30 items
manim/_config/__init__.py . [ 3%]
manim/_config/utils.py ..
## Logs
<details><summary>Terminal output</summary>
<!-- Add "-v DEBUG" when calling manim to generate more detailed logs -->
```
PASTE HERE OR PROVIDE LINK TO https://pastebin.com/ OR SIMILAR
```
<!-- Insert screenshots here (only when absolutely necessary, we prefer copy/pasted output!) -->
</details>
## System specifications
<details><summary>System Details</summary>
- OS (with version, e.g Windows 10 v2004 or macOS 10.15 (Catalina)):
- RAM:
- Python version (`python/py/python3 --version`):
- Installed modules (provide output from `pip list`):
```
PASTE HERE
```
</details>
<details><summary>LaTeX details</summary>
+ LaTeX distribution (e.g. TeX Live 2020):
+ Installed LaTeX packages:
<!-- output of `tlmgr list --only-installed` for TeX Live or a screenshot of the Packages page for MikTeX -->
</details>
<details><summary>FFMPEG</summary>
Output of `ffmpeg -version`:
```
PASTE HERE
```
</details>
## Additional comments
<!-- Add further context that you think might be relevant for this issue here. -->
[ 10%]
manim/animation/animation.py . [ 13%]
manim/mobject/geometry.py ......... [ 43%]
manim/mobject/mobject.py ...... [ 63%]
manim/mobject/svg/tex_mobject.py ..F [ 73%]
manim/mobject/svg/text_mobject.py ... [ 83%]
manim/mobject/types/vectorized_mobject.py .... [ 96%]
manim/utils/color.py . [100%]
=================================== FAILURES ===================================
_________________ [doctest] manim.mobject.svg.tex_mobject.Tex __________________
491 A string compiled with LaTeX in normal mode.
492
493 Tests
494 -----
495
496 Check whether writing a LaTeX string works::
497
498 >>> Tex('The horse does not eat cucumber salad.')
UNEXPECTED EXCEPTION: IndexError('list index out of range')
Traceback (most recent call last):
File "/opt/hostedtoolcache/Python/3.8.8/x64/lib/python3.8/doctest.py", line 1336, in __run
exec(compile(example.source, filename, "single",
File "<doctest manim.mobject.svg.tex_mobject.Tex[0]>", line 1, in <module>
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 506, in __init__
MathTex.__init__(
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 392, in __init__
self.break_up_by_substrings()
File "/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py", line 434, in break_up_by_substrings
sub_tex_mob.move_to(self.submobjects[last_submob_index], RIGHT)
IndexError: list index out of range
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/tex_mobject.py:498: UnexpectedException
----------------------------- Captured stdout call -----------------------------
INFO Writing "The horse does not tex_file_writing.py:81
eat cucumber salad." to medi
a/Tex/3ecc83aec1683253.tex
------------------------------ Captured log call -------------------------------
INFO manim:tex_file_writing.py:81 Writing "The horse does not eat cucumber salad." to media/Tex/3ecc83aec1683253.tex
=============================== warnings summary ===============================
manim/mobject/mobject.py::manim.mobject.mobject.Mobject.set
<doctest manim.mobject.mobject.Mobject.set[1]>:1: DeprecationWarning: This method is not guaranteed to stay around. Please prefer setting the attribute normally or with Mobject.set().
manim/mobject/mobject.py::manim.mobject.mobject.Mobject.set
<doctest manim.mobject.mobject.Mobject.set[2]>:1: DeprecationWarning: This method is not guaranteed to stay around. Please prefer getting the attribute normally.
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-84, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-104, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-101, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-111, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-114, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-115, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-100, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-110, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-116, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-97, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-99, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-117, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-109, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-98, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-108, which is not recognized
warnings.warn(warning_text)
manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
/home/runner/work/manim-cov/manim-cov/manim/mobject/svg/svg_mobject.py:261: UserWarning: media/Tex/3ecc83aec1683253.svg contains a reference to id #g0-46, which is not recognized
warnings.warn(warning_text)
-- Docs: https://docs.pytest.org/en/stable/warnings.html
----------- coverage: platform linux, python 3.8.8-final-0 -----------
Coverage XML written to file coverage.xml
=========================== short test summary info ============================
FAILED manim/mobject/svg/tex_mobject.py::manim.mobject.svg.tex_mobject.Tex
================== 1 failed, 29 passed, 18 warnings in 6.56s ===================
Error: Process completed with exit code 1.
```
</details> | infrastructure | nondeterministic build failures in tex the horse does not eat cucumber salad description of bug unexpected behavior as i ve been figuring out the test coverage toolset for manim i ve done quite a lot of automated builds on a copied not forked repository for some reason occasionally one of the ubuntu builds fails while i ve been playing around with coverage settings whether it s or is apparently random all of them fail in the tex doctest expected behavior it should pass the test how to reproduce the issue code for reproducing the problem in theory the problematic code is here py tex the horse does not eat cucumber salad however i ve not been able to replicate this issue additional media files build output run poetry run pytest cov append doctest modules manim poetry run pytest cov append doctest modules manim shell usr bin bash e env poetry virtualenvs create false pythonlocation opt hostedtoolcache python ld library path opt hostedtoolcache python lib skipping virtualenv creation as specified in config file test session starts platform linux python pytest py pluggy rootdir home runner work manim cov manim cov configfile pyproject toml plugins cov collected items manim config init py manim config utils py logs terminal output paste here or provide link to or similar system specifications system details os with version e g windows or macos catalina ram python version python py version installed modules provide output from pip list paste here latex details latex distribution e g tex live installed latex packages ffmpeg output of ffmpeg version paste here additional comments manim animation animation py manim mobject geometry py manim mobject mobject py manim mobject svg tex mobject py f manim mobject svg text mobject py manim mobject types vectorized mobject py manim utils color py failures manim mobject svg tex mobject tex a string compiled with latex in normal mode tests check whether writing a latex string works tex the horse does not eat cucumber salad unexpected exception indexerror list index out of range traceback most recent call last file opt hostedtoolcache python lib doctest py line in run exec compile example source filename single file line in file home runner work manim cov manim cov manim mobject svg tex mobject py line in init mathtex init file home runner work manim cov manim cov manim mobject svg tex mobject py line in init self break up by substrings file home runner work manim cov manim cov manim mobject svg tex mobject py line in break up by substrings sub tex mob move to self submobjects right indexerror list index out of range home runner work manim cov manim cov manim mobject svg tex mobject py unexpectedexception captured stdout call info writing the horse does not tex file writing py eat cucumber salad to medi a tex tex captured log call info manim tex file writing py writing the horse does not eat cucumber salad to media tex tex warnings summary manim mobject mobject py manim mobject mobject mobject set deprecationwarning this method is not guaranteed to stay around please prefer setting the attribute normally or with mobject set manim mobject mobject py manim mobject mobject mobject set deprecationwarning this method is not guaranteed to stay around please prefer getting the attribute normally manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text manim mobject svg tex mobject py manim mobject svg tex mobject tex home runner work manim cov manim cov manim mobject svg svg mobject py userwarning media tex svg contains a reference to id which is not recognized warnings warn warning text docs coverage platform linux python final coverage xml written to file coverage xml short test summary info failed manim mobject svg tex mobject py manim mobject svg tex mobject tex failed passed warnings in error process completed with exit code | 1 |
25,136 | 18,157,222,918 | IssuesEvent | 2021-09-27 04:21:59 | happy-travel/agent-app-project | https://api.github.com/repos/happy-travel/agent-app-project | closed | Check all connector updaters to have stdout logger and sentry connected | backend infrastructure static-data updater npt-testable | Sentry projects management is on @evm-andrey side.
In this task it is needed to check the code | 1.0 | Check all connector updaters to have stdout logger and sentry connected - Sentry projects management is on @evm-andrey side.
In this task it is needed to check the code | infrastructure | check all connector updaters to have stdout logger and sentry connected sentry projects management is on evm andrey side in this task it is needed to check the code | 1 |
26,305 | 19,978,409,779 | IssuesEvent | 2022-01-29 13:40:16 | angular/material.angular.io | https://api.github.com/repos/angular/material.angular.io | closed | Update Stackblitz package.json to use Angular ^13.0.0 | P1: urgent infrastructure | <!--------
๐
Use the Angular Components repository (https://github.com/angular/components/issues/new/choose)
to report issues.
The Angular team can't provide general troubleshooting help. This is especially true when the
problem is specific to your app and cannot be reproduced in a StackBlitz demo.
However, the extended community of users may be able to provide help via the following channels:
- StackOverflow: https://stackoverflow.com/questions/tagged/angular-material2
- Gitter: https://gitter.im/angular/material2
- Google Groups: https://groups.google.com/forum/#!forum/angular-material2
-------->
Currently the [package.json](https://github.com/angular/material.angular.io/blob/master/src/assets/stack-blitz/package.json) for Stackblitz demos uses Angular version `~13.0.0-next.0` and it results in buggy projects like this: https://github.com/angular/components/issues/24173. | 1.0 | Update Stackblitz package.json to use Angular ^13.0.0 - <!--------
๐
Use the Angular Components repository (https://github.com/angular/components/issues/new/choose)
to report issues.
The Angular team can't provide general troubleshooting help. This is especially true when the
problem is specific to your app and cannot be reproduced in a StackBlitz demo.
However, the extended community of users may be able to provide help via the following channels:
- StackOverflow: https://stackoverflow.com/questions/tagged/angular-material2
- Gitter: https://gitter.im/angular/material2
- Google Groups: https://groups.google.com/forum/#!forum/angular-material2
-------->
Currently the [package.json](https://github.com/angular/material.angular.io/blob/master/src/assets/stack-blitz/package.json) for Stackblitz demos uses Angular version `~13.0.0-next.0` and it results in buggy projects like this: https://github.com/angular/components/issues/24173. | infrastructure | update stackblitz package json to use angular ๐ use the angular components repository to report issues the angular team can t provide general troubleshooting help this is especially true when the problem is specific to your app and cannot be reproduced in a stackblitz demo however the extended community of users may be able to provide help via the following channels stackoverflow gitter google groups currently the for stackblitz demos uses angular version next and it results in buggy projects like this | 1 |
25,175 | 18,239,297,706 | IssuesEvent | 2021-10-01 10:53:42 | tskit-dev/tskit | https://api.github.com/repos/tskit-dev/tskit | closed | warnings building tskit with Ming64 on Windows | C API Infrastructure and tools | @rdinnager reports the following warnings when building tskit (actually tskit's code inside SLiM, but it's the same code) on Windows using the Ming64 compiler:
```
F:/SLiM/treerec/tskit/core.c: In function 'get_random_bytes':
F:/SLiM/treerec/tskit/core.c:47:29: warning: initialization of 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
47 | HCRYPTPROV hCryptProv = NULL;
| ^~~~
F:/SLiM/treerec/tskit/core.c:57:20: warning: assignment to 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
57 | hCryptProv = NULL;
| ^
F:/SLiM/treerec/tskit/core.c:60:16: warning: assignment to 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
60 | hCryptProv = NULL;
| ^
F:/SLiM/treerec/tskit/core.c:63:20: warning: comparison between pointer and integer
63 | if (hCryptProv != NULL) {
| ^~
```
According to the Microsoft docs (https://docs.microsoft.com/en-us/windows/win32/seccrypto/hcryptprov) `HCRYPTPROV` is supposed to be a pointer to `unsigned long`:
`typedef ULONG_PTR HCRYPTPROV;`
It looks like whatever headers are coming from Russell's toolchain define it as `long long unsigned int` instead; I have no idea why. I would suggest that maybe tskit ought to change those four assignments above, of `NULL`, to be assignments of `(HCRYPTPROV)NULL` so that it builds without warnings regardless of what type `HCRYPTPROV` is defined to be. Should be a trivial fix. Thanks!
This was originally discussed in https://github.com/MesserLab/SLiM/issues/66 but there's no need to read through that, it just says the same things I say above. :-> | 1.0 | warnings building tskit with Ming64 on Windows - @rdinnager reports the following warnings when building tskit (actually tskit's code inside SLiM, but it's the same code) on Windows using the Ming64 compiler:
```
F:/SLiM/treerec/tskit/core.c: In function 'get_random_bytes':
F:/SLiM/treerec/tskit/core.c:47:29: warning: initialization of 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
47 | HCRYPTPROV hCryptProv = NULL;
| ^~~~
F:/SLiM/treerec/tskit/core.c:57:20: warning: assignment to 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
57 | hCryptProv = NULL;
| ^
F:/SLiM/treerec/tskit/core.c:60:16: warning: assignment to 'HCRYPTPROV' {aka 'long long unsigned int'} from 'void *' makes integer from pointer without a cast [-Wint-conversion]
60 | hCryptProv = NULL;
| ^
F:/SLiM/treerec/tskit/core.c:63:20: warning: comparison between pointer and integer
63 | if (hCryptProv != NULL) {
| ^~
```
According to the Microsoft docs (https://docs.microsoft.com/en-us/windows/win32/seccrypto/hcryptprov) `HCRYPTPROV` is supposed to be a pointer to `unsigned long`:
`typedef ULONG_PTR HCRYPTPROV;`
It looks like whatever headers are coming from Russell's toolchain define it as `long long unsigned int` instead; I have no idea why. I would suggest that maybe tskit ought to change those four assignments above, of `NULL`, to be assignments of `(HCRYPTPROV)NULL` so that it builds without warnings regardless of what type `HCRYPTPROV` is defined to be. Should be a trivial fix. Thanks!
This was originally discussed in https://github.com/MesserLab/SLiM/issues/66 but there's no need to read through that, it just says the same things I say above. :-> | infrastructure | warnings building tskit with on windows rdinnager reports the following warnings when building tskit actually tskit s code inside slim but it s the same code on windows using the compiler f slim treerec tskit core c in function get random bytes f slim treerec tskit core c warning initialization of hcryptprov aka long long unsigned int from void makes integer from pointer without a cast hcryptprov hcryptprov null f slim treerec tskit core c warning assignment to hcryptprov aka long long unsigned int from void makes integer from pointer without a cast hcryptprov null f slim treerec tskit core c warning assignment to hcryptprov aka long long unsigned int from void makes integer from pointer without a cast hcryptprov null f slim treerec tskit core c warning comparison between pointer and integer if hcryptprov null according to the microsoft docs hcryptprov is supposed to be a pointer to unsigned long typedef ulong ptr hcryptprov it looks like whatever headers are coming from russell s toolchain define it as long long unsigned int instead i have no idea why i would suggest that maybe tskit ought to change those four assignments above of null to be assignments of hcryptprov null so that it builds without warnings regardless of what type hcryptprov is defined to be should be a trivial fix thanks this was originally discussed in but there s no need to read through that it just says the same things i say above | 1 |
339,665 | 10,257,558,875 | IssuesEvent | 2019-08-21 20:25:39 | Apexal/late | https://api.github.com/repos/Apexal/late | closed | Calendar Blocks | Enhancement Front End Priority: Low | **Describe the bug**
Calendar blocks are too small to be viewed.
**To Reproduce** *optional*
Steps to reproduce the behavior:
Login and view the calendar on your phone
**Expected behavior**

**Screenshots** *optional*

**Device**
iPhone XR | 1.0 | Calendar Blocks - **Describe the bug**
Calendar blocks are too small to be viewed.
**To Reproduce** *optional*
Steps to reproduce the behavior:
Login and view the calendar on your phone
**Expected behavior**

**Screenshots** *optional*

**Device**
iPhone XR | non_infrastructure | calendar blocks describe the bug calendar blocks are too small to be viewed to reproduce optional steps to reproduce the behavior login and view the calendar on your phone expected behavior screenshots optional device iphone xr | 0 |
7,804 | 7,104,241,126 | IssuesEvent | 2018-01-16 09:18:21 | eclipse/vorto | https://api.github.com/repos/eclipse/vorto | closed | Automated Web Site Deployment | Infrastructure Website | When a PR for web site change is accepted and merged into the codebase, the Vorto CI (Hudson) builds the website with Jekyll and pushes it to Eclipse web site git repository, so that these changes are immediately visible on www.eclipse.org/vorto
| 1.0 | Automated Web Site Deployment - When a PR for web site change is accepted and merged into the codebase, the Vorto CI (Hudson) builds the website with Jekyll and pushes it to Eclipse web site git repository, so that these changes are immediately visible on www.eclipse.org/vorto
| infrastructure | automated web site deployment when a pr for web site change is accepted and merged into the codebase the vorto ci hudson builds the website with jekyll and pushes it to eclipse web site git repository so that these changes are immediately visible on | 1 |
4,427 | 5,068,325,493 | IssuesEvent | 2016-12-24 15:11:48 | Hackerfleet/meta | https://api.github.com/repos/Hackerfleet/meta | opened | Set up c3 RPi | events infrastructure | * [ ] Attach where it cannot be de-attached
* [ ] Make sure it works
* [ ] Tighten security as required | 1.0 | Set up c3 RPi - * [ ] Attach where it cannot be de-attached
* [ ] Make sure it works
* [ ] Tighten security as required | infrastructure | set up rpi attach where it cannot be de attached make sure it works tighten security as required | 1 |
31,021 | 25,259,630,120 | IssuesEvent | 2022-11-15 21:25:54 | dotnet/razor | https://api.github.com/repos/dotnet/razor | closed | Build source code produce: "The variable '$vsMajorVersion' cannot be retrieved because it has not been set." | area-infrastructure | **Describe the bug:**
Cannot build the source code using the command line on **Windows** Docker image, but the same error occurs also on my machine
**Version used:**
PowerShell Core 7.2
**To reproduce:**
1. Fresh Windows servercore machine
2. Install Git 2.38.1
3. Install NodeJs 19
4. Install Python39
5. Globally install yarn -> `npm I -g yarn`
6. clone repo
7. launch build.cmd
**Expected behavior:**
Build Successfully
**Actual behavior:**
```
Using xcopy-msbuild version of 17.2.1 since VS version 16.3 provided in global.json is not compatible
The variable '$vsMajorVersion' cannot be retrieved because it has not been set.
```
**Additional context:**
I have tried both Linux(ubuntu) and Windows(servercore:ltsc2022)
Linux: build OK
Windows: Error above
I added the Dockerfiles used
## Linux Dockerfile
```
FROM ubuntu
# Add required packages
RUN apt update && \
install curl \
-y
# Add nodejs source
RUN curl -s https://deb.nodesource.com/setup_16.x | bash
# Install nodejs 16 dotnet6, python3 is already installed
RUN apt install nodejs dotnet6 wget git -y
# Run the machine and clone/build repo
```
## Win Dockerfile
```
FROM mcr.microsoft.com/windows/servercore:ltsc2022
# Download software
WORKDIR c:/temp
# -L means allow redirect, -O same name specified on URL
RUN curl -L -O https://github.com/PowerShell/PowerShell/releases/download/v7.2.6/PowerShell-7.2.6-win-x64.zip
RUN curl -L -O https://www.7-zip.org/a/7z2201-x64.exe
RUN curl -L -O https://github.com/git-for-windows/git/releases/download/v2.38.1.windows.1/PortableGit-2.38.1-64-bit.7z.exe
# Extract all
RUN 7z2201-x64.exe /S /D=c:/apps/7zip
RUN c:/apps/7zip/7z.exe x c:/temp/PowerShell-7.2.6-win-x64.zip -oc:/apps/PowerShell
RUN c:/apps/7zip/7z.exe x c:/temp/PortableGit-2.38.1-64-bit.7z.exe -oc:/apps/git
RUN setx path "%path%;C:\apps\PowerShell;c:\apps\git\bin;c:\apps\7zip"
RUN del c:\temp /q
RUN net user user01 /ADD
USER user01
RUN setx path "%path%;C:\apps\PowerShell;c:\apps\git\bin;c:\apps\7zip"
RUN pwsh -command Set-ExecutionPolicy RemoteSigned -Scope CurrentUser
RUN ["pwsh", "-command", "irm get.scoop.sh | iex"]
RUN scoop bucket add versions
RUN scoop bucket add extras
RUN scoop install nodejs
RUN scoop install python39
RUN npm i -g yarn
# Run the machine and clone/build repo
```
| 1.0 | Build source code produce: "The variable '$vsMajorVersion' cannot be retrieved because it has not been set." - **Describe the bug:**
Cannot build the source code using the command line on **Windows** Docker image, but the same error occurs also on my machine
**Version used:**
PowerShell Core 7.2
**To reproduce:**
1. Fresh Windows servercore machine
2. Install Git 2.38.1
3. Install NodeJs 19
4. Install Python39
5. Globally install yarn -> `npm I -g yarn`
6. clone repo
7. launch build.cmd
**Expected behavior:**
Build Successfully
**Actual behavior:**
```
Using xcopy-msbuild version of 17.2.1 since VS version 16.3 provided in global.json is not compatible
The variable '$vsMajorVersion' cannot be retrieved because it has not been set.
```
**Additional context:**
I have tried both Linux(ubuntu) and Windows(servercore:ltsc2022)
Linux: build OK
Windows: Error above
I added the Dockerfiles used
## Linux Dockerfile
```
FROM ubuntu
# Add required packages
RUN apt update && \
install curl \
-y
# Add nodejs source
RUN curl -s https://deb.nodesource.com/setup_16.x | bash
# Install nodejs 16 dotnet6, python3 is already installed
RUN apt install nodejs dotnet6 wget git -y
# Run the machine and clone/build repo
```
## Win Dockerfile
```
FROM mcr.microsoft.com/windows/servercore:ltsc2022
# Download software
WORKDIR c:/temp
# -L means allow redirect, -O same name specified on URL
RUN curl -L -O https://github.com/PowerShell/PowerShell/releases/download/v7.2.6/PowerShell-7.2.6-win-x64.zip
RUN curl -L -O https://www.7-zip.org/a/7z2201-x64.exe
RUN curl -L -O https://github.com/git-for-windows/git/releases/download/v2.38.1.windows.1/PortableGit-2.38.1-64-bit.7z.exe
# Extract all
RUN 7z2201-x64.exe /S /D=c:/apps/7zip
RUN c:/apps/7zip/7z.exe x c:/temp/PowerShell-7.2.6-win-x64.zip -oc:/apps/PowerShell
RUN c:/apps/7zip/7z.exe x c:/temp/PortableGit-2.38.1-64-bit.7z.exe -oc:/apps/git
RUN setx path "%path%;C:\apps\PowerShell;c:\apps\git\bin;c:\apps\7zip"
RUN del c:\temp /q
RUN net user user01 /ADD
USER user01
RUN setx path "%path%;C:\apps\PowerShell;c:\apps\git\bin;c:\apps\7zip"
RUN pwsh -command Set-ExecutionPolicy RemoteSigned -Scope CurrentUser
RUN ["pwsh", "-command", "irm get.scoop.sh | iex"]
RUN scoop bucket add versions
RUN scoop bucket add extras
RUN scoop install nodejs
RUN scoop install python39
RUN npm i -g yarn
# Run the machine and clone/build repo
```
| infrastructure | build source code produce the variable vsmajorversion cannot be retrieved because it has not been set describe the bug cannot build the source code using the command line on windows docker image but the same error occurs also on my machine version used powershell core to reproduce fresh windows servercore machine install git install nodejs install globally install yarn npm i g yarn clone repo launch build cmd expected behavior build successfully actual behavior using xcopy msbuild version of since vs version provided in global json is not compatible the variable vsmajorversion cannot be retrieved because it has not been set additional context i have tried both linux ubuntu and windows servercore linux build ok windows error above i added the dockerfiles used linux dockerfile from ubuntu add required packages run apt update install curl y add nodejs source run curl s bash install nodejs is already installed run apt install nodejs wget git y run the machine and clone build repo win dockerfile from mcr microsoft com windows servercore download software workdir c temp l means allow redirect o same name specified on url run curl l o run curl l o run curl l o extract all run exe s d c apps run c apps exe x c temp powershell win zip oc apps powershell run c apps exe x c temp portablegit bit exe oc apps git run setx path path c apps powershell c apps git bin c apps run del c temp q run net user add user run setx path path c apps powershell c apps git bin c apps run pwsh command set executionpolicy remotesigned scope currentuser run run scoop bucket add versions run scoop bucket add extras run scoop install nodejs run scoop install run npm i g yarn run the machine and clone build repo | 1 |
26,300 | 19,974,804,101 | IssuesEvent | 2022-01-29 00:34:07 | acm-toce/documentation | https://api.github.com/repos/acm-toce/documentation | closed | Revise author guidelines | infrastructure | Some suggested revisions:
* Add a set of contribution types (like the broad set often discussed in HCI) to help authors understand how their submissions will be evaluated.
* Add a section on reporting standards, link to the CSEdResearch page on how to report studies.
* Clarify anonymization practices (e.g., anonymize self-citations w/ third person)
* Clarify whether dissertations are prior publications (e.g., an author wrote: "I have a paper recently published under the ACM Transactions for Computing Education for which you oversee. The paper is part of my PhD thesis. So, I wonder with the copy right contract that I had signed, would I be able to use the content word-by-word in my thesis along with an acknowledge and citation? I want to make sure I have the approval before including the article as one of my chapters" -> https://www.acm.org/publications/policies/copyright-policy#permanent%20rights
* Explain how to handle overlapping prior publications (30% rule, link to ACM)
* Explain to authors how to interpret the reviews they receive (e.g., AE's have discretion, point to AE and reviewer guidelines) | 1.0 | Revise author guidelines - Some suggested revisions:
* Add a set of contribution types (like the broad set often discussed in HCI) to help authors understand how their submissions will be evaluated.
* Add a section on reporting standards, link to the CSEdResearch page on how to report studies.
* Clarify anonymization practices (e.g., anonymize self-citations w/ third person)
* Clarify whether dissertations are prior publications (e.g., an author wrote: "I have a paper recently published under the ACM Transactions for Computing Education for which you oversee. The paper is part of my PhD thesis. So, I wonder with the copy right contract that I had signed, would I be able to use the content word-by-word in my thesis along with an acknowledge and citation? I want to make sure I have the approval before including the article as one of my chapters" -> https://www.acm.org/publications/policies/copyright-policy#permanent%20rights
* Explain how to handle overlapping prior publications (30% rule, link to ACM)
* Explain to authors how to interpret the reviews they receive (e.g., AE's have discretion, point to AE and reviewer guidelines) | infrastructure | revise author guidelines some suggested revisions add a set of contribution types like the broad set often discussed in hci to help authors understand how their submissions will be evaluated add a section on reporting standards link to the csedresearch page on how to report studies clarify anonymization practices e g anonymize self citations w third person clarify whether dissertations are prior publications e g an author wrote i have a paper recently published under the acm transactions for computing education for which you oversee the paper is part of my phd thesis so i wonder with the copy right contract that i had signed would i be able to use the content word by word in my thesis along with an acknowledge and citation i want to make sure i have the approval before including the article as one of my chapters explain how to handle overlapping prior publications rule link to acm explain to authors how to interpret the reviews they receive e g ae s have discretion point to ae and reviewer guidelines | 1 |
11,303 | 9,087,175,093 | IssuesEvent | 2019-02-18 13:03:33 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | REST API cant filter datetimes by deleted field | category:models-and-data-infrastructure status:stale type:bug ๐ | <!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Please ensure that what you are reporting is specific to this project.
- Try to add as much detail as possible. Be specific!
- Make sure you read the README.md for the project regarding posting issues.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before reporting bugs (unless you are reporting an issue disovered in a branch).
- Disable all plugins and switch to a default theme to ensure its not a plugin/theme conflict issue.
- To report a security issue, please visit this page: https://eventespresso.com/report-a-security-vulnerability/
-->
## Issue Overview
<!-- Describe what this issue is about. -->
Queries like this always return an empty set: `/datetimes?where[DTT_deleted]=false` returns an empty set (even when there are definetely non-trashed datetimes) and so does `/datetimes?where[DTT_deleted]=true`.
## Bug report or feature request?
* [x] Bug
* [ ] Feature
* [ ] Neither
## Environment Data:
Version of EE: <!-- Can be a branch name or the version of EE the issue happened in. -->
4.9.71
Version of WordPress:
4.9.7
PHP Version: <!-- if known, add your php version here -->
7.1
Browser used: <!-- also include your browser version if possible -->
Firefox/postman
## Steps to Reproduce (for bugs)
<!-- If possible provide any links to a live example, or an unambiguous set of steps to reproduce this bug -->
<!-- Feel free to include code to reproduce if relevant. -->
1. Create an event with two datetimes, and trash one.
2. Send a request to `wp-json/ee/v4.8.36/datetimes`, you should see the one untrashed one. Then add `?where[DTT_deleted]=false` and somehow you'll get an empty set. Then change it to `?where[DTT_deleted]=true` and it's still empty. Weirdness.
## Expected Behaviour
<!-- If you're describing a bug, tell us what should happen -->
<!-- If you're describing a feature/enhancement, explain the difference from current behaviour -->
`/wp-json/ee/v4.8.36/datetimes?where[DTT_deleted]=false` should return undeleted datetimes. And instead using `where[DTT_deleted]=true` should... probably tell you you're not allowed to see deleted datetimes, as they're not visible on the frontend; or maybe show them. It's up for debate, but probably an error would be better.
## Current Behaviour
<!-- If describing a bug, what is the current behaviour and how does it differ from expected behaviour? -->
<!-- If describing a feature, describe what the current behaviour is in the part of the application that you want your feature suggestion to improve on. -->
See steps to reproduce.
## Related Information:
<!--
- If you were directed to create an issue here by the EE support team, you can include the link to your original EE support forum thread.
- You can also include any other links you think may be useful (related issues and/or Pull Requests)
- Any screenshots or screencasts that help illustrate what you are describing is always useful.
-->
| 1.0 | REST API cant filter datetimes by deleted field - <!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Please ensure that what you are reporting is specific to this project.
- Try to add as much detail as possible. Be specific!
- Make sure you read the README.md for the project regarding posting issues.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before reporting bugs (unless you are reporting an issue disovered in a branch).
- Disable all plugins and switch to a default theme to ensure its not a plugin/theme conflict issue.
- To report a security issue, please visit this page: https://eventespresso.com/report-a-security-vulnerability/
-->
## Issue Overview
<!-- Describe what this issue is about. -->
Queries like this always return an empty set: `/datetimes?where[DTT_deleted]=false` returns an empty set (even when there are definetely non-trashed datetimes) and so does `/datetimes?where[DTT_deleted]=true`.
## Bug report or feature request?
* [x] Bug
* [ ] Feature
* [ ] Neither
## Environment Data:
Version of EE: <!-- Can be a branch name or the version of EE the issue happened in. -->
4.9.71
Version of WordPress:
4.9.7
PHP Version: <!-- if known, add your php version here -->
7.1
Browser used: <!-- also include your browser version if possible -->
Firefox/postman
## Steps to Reproduce (for bugs)
<!-- If possible provide any links to a live example, or an unambiguous set of steps to reproduce this bug -->
<!-- Feel free to include code to reproduce if relevant. -->
1. Create an event with two datetimes, and trash one.
2. Send a request to `wp-json/ee/v4.8.36/datetimes`, you should see the one untrashed one. Then add `?where[DTT_deleted]=false` and somehow you'll get an empty set. Then change it to `?where[DTT_deleted]=true` and it's still empty. Weirdness.
## Expected Behaviour
<!-- If you're describing a bug, tell us what should happen -->
<!-- If you're describing a feature/enhancement, explain the difference from current behaviour -->
`/wp-json/ee/v4.8.36/datetimes?where[DTT_deleted]=false` should return undeleted datetimes. And instead using `where[DTT_deleted]=true` should... probably tell you you're not allowed to see deleted datetimes, as they're not visible on the frontend; or maybe show them. It's up for debate, but probably an error would be better.
## Current Behaviour
<!-- If describing a bug, what is the current behaviour and how does it differ from expected behaviour? -->
<!-- If describing a feature, describe what the current behaviour is in the part of the application that you want your feature suggestion to improve on. -->
See steps to reproduce.
## Related Information:
<!--
- If you were directed to create an issue here by the EE support team, you can include the link to your original EE support forum thread.
- You can also include any other links you think may be useful (related issues and/or Pull Requests)
- Any screenshots or screencasts that help illustrate what you are describing is always useful.
-->
| infrastructure | rest api cant filter datetimes by deleted field before posting your issue these comments won t show up when you submit the issue please ensure that what you are reporting is specific to this project try to add as much detail as possible be specific make sure you read the readme md for the project regarding posting issues search this repository for issues and pull requests and whether it has been fixed or reported already ensure you are using the latest code before reporting bugs unless you are reporting an issue disovered in a branch disable all plugins and switch to a default theme to ensure its not a plugin theme conflict issue to report a security issue please visit this page issue overview queries like this always return an empty set datetimes where false returns an empty set even when there are definetely non trashed datetimes and so does datetimes where true bug report or feature request bug feature neither environment data version of ee version of wordpress php version browser used firefox postman steps to reproduce for bugs create an event with two datetimes and trash one send a request to wp json ee datetimes you should see the one untrashed one then add where false and somehow you ll get an empty set then change it to where true and it s still empty weirdness expected behaviour wp json ee datetimes where false should return undeleted datetimes and instead using where true should probably tell you you re not allowed to see deleted datetimes as they re not visible on the frontend or maybe show them it s up for debate but probably an error would be better current behaviour see steps to reproduce related information if you were directed to create an issue here by the ee support team you can include the link to your original ee support forum thread you can also include any other links you think may be useful related issues and or pull requests any screenshots or screencasts that help illustrate what you are describing is always useful | 1 |
96,683 | 12,150,381,169 | IssuesEvent | 2020-04-24 17:52:27 | tesshucom/jpsonic | https://api.github.com/repos/tesshucom/jpsonic | opened | Add an option to specify the genre handled as audiobook | in : search status: pending-design-work type: enhancement | There is often song data that is not music, such as the first intro of 10 songs in the album.
These are treated as AUDIOBOOK by specifying the genre and excluded from the shuffle.
There are some notes.
- Should it be included in the genre master
- Whether to increase the MediaType (although it is difficult to verify).
| 1.0 | Add an option to specify the genre handled as audiobook - There is often song data that is not music, such as the first intro of 10 songs in the album.
These are treated as AUDIOBOOK by specifying the genre and excluded from the shuffle.
There are some notes.
- Should it be included in the genre master
- Whether to increase the MediaType (although it is difficult to verify).
| non_infrastructure | add an option to specify the genre handled as audiobook there is often song data that is not music such as the first intro of songs in the album these are treated as audiobook by specifying the genre and excluded from the shuffle there are some notes should it be included in the genre master whether to increase the mediatype although it is difficult to verify | 0 |
433,257 | 30,320,507,290 | IssuesEvent | 2023-07-10 18:51:17 | gravitational/teleport | https://api.github.com/repos/gravitational/teleport | opened | Install instructions for rhel-like distros fail | documentation | ## Applies To
https://goteleport.com/docs/installation/#linux
## Details
The instructions in the `Amazon Linux 2023/RHEL 8+ (dnf)` tab in the Enterprise section don't work on Rhel.
```
[root@rheltest /]# source os-release && dnf config-manager --add-repo "$(rpm --eval https://yum.releases.teleport.dev/${ID}/${VERSION_ID}/Teleport/%{_arch}/stable/v13/teleport.repo)"
Adding repo from: https://yum.releases.teleport.dev/rhel/8.8/Teleport/x86_64/stable/v13/teleport.repo
Status code: 404 for https://yum.releases.teleport.dev/rhel/8.8/Teleport/x86_64/stable/v13/teleport.repo (IP: 18.238.4.22)
Error: Configuration of repo failed
```
Additionally, I would expect that our instructions would work for centos stream, rockylinux 7/8, and almalinux 7/8.
As far as I can tell, the valid values for `${ID}` and `${VERSION_ID}` are as follows:
* centos stream 8 -> centos 8
* rocky linux 8 -> rhel 8
* alma linux 8 -> rhel 8
* rhel 8.x -> rhel 8
rocky, alma, and rhel seem to set `VERSION_ID` to 8.6 or 8.8 or similar, which doesn't work.
The docs advise that it may be necessary to use `ID_LIKE` instead of `ID`, but rocky and alma set that to `rhel centos fedora`, which would not be valid.
It might be feasible to write a command that tries to map these values to the necessary values based on what's in the repository. Another approach would be to list the valid values in our repo and let the end-user pick which values are appropriate.
Here is a table of various rhel-like distros and what their ID, ID_LIKE, and VERSION_ID values are:
| distro | id | id_like | version_id |
|--------------------------|------------|-----------------------|------------|
| rockylinux 8 | rocky | rhel centos fedora | 8.8 |
| rockylinux 9 | rocky | rhel centos fedora | 9.2 |
| almalinux 8 | almalinux | rhel centos fedora | 8.8 |
| almalinux 9 | almalinux | rhel centos fedora | 9.2 |
| oraclelinux 7 | ol | fedora | 7.9 |
| oraclelinux 8 | ol | fedora | 8.8 |
| oraclelinux 9 | ol | fedora | 9.2 |
| rhel7 | rhel | fedora | 7.9 |
| rhel8 | rhel | fedora | 8.8 |
| rhel9 | rhel | fedora | 9.2 |
| centos stream 8 | centos | rhel fedora | 8 |
| centos stream 9 | centos | rhel fedora | 9 |
| centos 7 | centos | rhel fedora | 7 |
| centos 8 (EOL/Canceled) | centos | rhel fedora | 8 |
## How will we know this is resolved?
Instructions on rhel and clones will work without modification.
## Related Issues
| 1.0 | Install instructions for rhel-like distros fail - ## Applies To
https://goteleport.com/docs/installation/#linux
## Details
The instructions in the `Amazon Linux 2023/RHEL 8+ (dnf)` tab in the Enterprise section don't work on Rhel.
```
[root@rheltest /]# source os-release && dnf config-manager --add-repo "$(rpm --eval https://yum.releases.teleport.dev/${ID}/${VERSION_ID}/Teleport/%{_arch}/stable/v13/teleport.repo)"
Adding repo from: https://yum.releases.teleport.dev/rhel/8.8/Teleport/x86_64/stable/v13/teleport.repo
Status code: 404 for https://yum.releases.teleport.dev/rhel/8.8/Teleport/x86_64/stable/v13/teleport.repo (IP: 18.238.4.22)
Error: Configuration of repo failed
```
Additionally, I would expect that our instructions would work for centos stream, rockylinux 7/8, and almalinux 7/8.
As far as I can tell, the valid values for `${ID}` and `${VERSION_ID}` are as follows:
* centos stream 8 -> centos 8
* rocky linux 8 -> rhel 8
* alma linux 8 -> rhel 8
* rhel 8.x -> rhel 8
rocky, alma, and rhel seem to set `VERSION_ID` to 8.6 or 8.8 or similar, which doesn't work.
The docs advise that it may be necessary to use `ID_LIKE` instead of `ID`, but rocky and alma set that to `rhel centos fedora`, which would not be valid.
It might be feasible to write a command that tries to map these values to the necessary values based on what's in the repository. Another approach would be to list the valid values in our repo and let the end-user pick which values are appropriate.
Here is a table of various rhel-like distros and what their ID, ID_LIKE, and VERSION_ID values are:
| distro | id | id_like | version_id |
|--------------------------|------------|-----------------------|------------|
| rockylinux 8 | rocky | rhel centos fedora | 8.8 |
| rockylinux 9 | rocky | rhel centos fedora | 9.2 |
| almalinux 8 | almalinux | rhel centos fedora | 8.8 |
| almalinux 9 | almalinux | rhel centos fedora | 9.2 |
| oraclelinux 7 | ol | fedora | 7.9 |
| oraclelinux 8 | ol | fedora | 8.8 |
| oraclelinux 9 | ol | fedora | 9.2 |
| rhel7 | rhel | fedora | 7.9 |
| rhel8 | rhel | fedora | 8.8 |
| rhel9 | rhel | fedora | 9.2 |
| centos stream 8 | centos | rhel fedora | 8 |
| centos stream 9 | centos | rhel fedora | 9 |
| centos 7 | centos | rhel fedora | 7 |
| centos 8 (EOL/Canceled) | centos | rhel fedora | 8 |
## How will we know this is resolved?
Instructions on rhel and clones will work without modification.
## Related Issues
| non_infrastructure | install instructions for rhel like distros fail applies to details the instructions in the amazon linux rhel dnf tab in the enterprise section don t work on rhel source os release dnf config manager add repo rpm eval adding repo from status code for ip error configuration of repo failed additionally i would expect that our instructions would work for centos stream rockylinux and almalinux as far as i can tell the valid values for id and version id are as follows centos stream centos rocky linux rhel alma linux rhel rhel x rhel rocky alma and rhel seem to set version id to or or similar which doesn t work the docs advise that it may be necessary to use id like instead of id but rocky and alma set that to rhel centos fedora which would not be valid it might be feasible to write a command that tries to map these values to the necessary values based on what s in the repository another approach would be to list the valid values in our repo and let the end user pick which values are appropriate here is a table of various rhel like distros and what their id id like and version id values are distro id id like version id rockylinux rocky rhel centos fedora rockylinux rocky rhel centos fedora almalinux almalinux rhel centos fedora almalinux almalinux rhel centos fedora oraclelinux ol fedora oraclelinux ol fedora oraclelinux ol fedora rhel fedora rhel fedora rhel fedora centos stream centos rhel fedora centos stream centos rhel fedora centos centos rhel fedora centos eol canceled centos rhel fedora how will we know this is resolved instructions on rhel and clones will work without modification related issues | 0 |
16,640 | 12,086,248,457 | IssuesEvent | 2020-04-18 08:58:23 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | CI unable to install dotnet-ef | area-infrastructure | Win7 and Win8 agents are failing in ProjectTemplates.Tests trying to install dotnet-ef.
```
Running 'C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1'
'C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1' completed with exit code '1'
Exception in InstallAspNetAppIfNeeded: System.InvalidOperationException: Command C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1 returned exit code 1
at RunTests.ProcessUtil.RunAsync(String filename, String arguments, String workingDirectory, Boolean throwOnError, IDictionary`2 environmentVariables, Action`1 outputDataReceived, Action`1 errorDataReceived, Action`1 onStart, CancellationToken cancellationToken) in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\ProcessUtil.cs:line 144
at RunTests.ProcessUtil.RunAsync(String filename, String arguments, String workingDirectory, Boolean throwOnError, IDictionary`2 environmentVariables, Action`1 outputDataReceived, Action`1 errorDataReceived, Action`1 onStart, CancellationToken cancellationToken) in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\ProcessUtil.cs:line 144
at RunTests.TestRunner.InstallAspNetAppIfNeededAsync() in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\TestRunner.cs:line 136
```
@HaoK is working on adding more logs here.
https://dev.azure.com/dnceng/internal/_build/results?buildId=603871&view=ms.vss-test-web.build-test-results-tab
https://dev.azure.com/dnceng/internal/_build/results?buildId=604194&view=ms.vss-test-web.build-test-results-tab&runId=18998130&resultId=120299&paneView=debug
https://dev.azure.com/dnceng/internal/_build/results?buildId=604848&view=ms.vss-test-web.build-test-results-tab
https://dev.azure.com/dnceng/internal/_build/results?buildId=605130&view=ms.vss-test-web.build-test-results-tab&runId=19026724&resultId=120297&paneView=attachments
https://dev.azure.com/dnceng/public/_build/results?buildId=604460&view=ms.vss-test-web.build-test-results-tab
This one is a bit different:
`Tool 'dotnet-ef' is already installed.`
https://dev.azure.com/dnceng/public/_build/results?buildId=605452&view=ms.vss-test-web.build-test-results-tab&runId=19034852&resultId=120299&paneView=attachments | 1.0 | CI unable to install dotnet-ef - Win7 and Win8 agents are failing in ProjectTemplates.Tests trying to install dotnet-ef.
```
Running 'C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1'
'C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1' completed with exit code '1'
Exception in InstallAspNetAppIfNeeded: System.InvalidOperationException: Command C:\h\w\A3FF092E\p\sdk\x64/dotnet tool install dotnet-ef --global --version 5.0.0-preview.4.20215.1 returned exit code 1
at RunTests.ProcessUtil.RunAsync(String filename, String arguments, String workingDirectory, Boolean throwOnError, IDictionary`2 environmentVariables, Action`1 outputDataReceived, Action`1 errorDataReceived, Action`1 onStart, CancellationToken cancellationToken) in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\ProcessUtil.cs:line 144
at RunTests.ProcessUtil.RunAsync(String filename, String arguments, String workingDirectory, Boolean throwOnError, IDictionary`2 environmentVariables, Action`1 outputDataReceived, Action`1 errorDataReceived, Action`1 onStart, CancellationToken cancellationToken) in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\ProcessUtil.cs:line 144
at RunTests.TestRunner.InstallAspNetAppIfNeededAsync() in C:\h\w\A3FF092E\w\B32B0A1A\e\RunTests\TestRunner.cs:line 136
```
@HaoK is working on adding more logs here.
https://dev.azure.com/dnceng/internal/_build/results?buildId=603871&view=ms.vss-test-web.build-test-results-tab
https://dev.azure.com/dnceng/internal/_build/results?buildId=604194&view=ms.vss-test-web.build-test-results-tab&runId=18998130&resultId=120299&paneView=debug
https://dev.azure.com/dnceng/internal/_build/results?buildId=604848&view=ms.vss-test-web.build-test-results-tab
https://dev.azure.com/dnceng/internal/_build/results?buildId=605130&view=ms.vss-test-web.build-test-results-tab&runId=19026724&resultId=120297&paneView=attachments
https://dev.azure.com/dnceng/public/_build/results?buildId=604460&view=ms.vss-test-web.build-test-results-tab
This one is a bit different:
`Tool 'dotnet-ef' is already installed.`
https://dev.azure.com/dnceng/public/_build/results?buildId=605452&view=ms.vss-test-web.build-test-results-tab&runId=19034852&resultId=120299&paneView=attachments | infrastructure | ci unable to install dotnet ef and agents are failing in projecttemplates tests trying to install dotnet ef running c h w p sdk dotnet tool install dotnet ef global version preview c h w p sdk dotnet tool install dotnet ef global version preview completed with exit code exception in installaspnetappifneeded system invalidoperationexception command c h w p sdk dotnet tool install dotnet ef global version preview returned exit code at runtests processutil runasync string filename string arguments string workingdirectory boolean throwonerror idictionary environmentvariables action outputdatareceived action errordatareceived action onstart cancellationtoken cancellationtoken in c h w w e runtests processutil cs line at runtests processutil runasync string filename string arguments string workingdirectory boolean throwonerror idictionary environmentvariables action outputdatareceived action errordatareceived action onstart cancellationtoken cancellationtoken in c h w w e runtests processutil cs line at runtests testrunner installaspnetappifneededasync in c h w w e runtests testrunner cs line haok is working on adding more logs here this one is a bit different tool dotnet ef is already installed | 1 |
597,994 | 18,233,684,465 | IssuesEvent | 2021-10-01 02:27:07 | AY2122S1-CS2103T-W12-3/tp | https://api.github.com/repos/AY2122S1-CS2103T-W12-3/tp | closed | Update UG for v1.1 | type.Task priority.High | Ensure that UG details are correct and move it into the repo
- [x] Table of contents
- [x] Quick start
- [x] Features
- [x] Add command
- [x] Edit command
- [x] Delete command
- [x] List command
- [x] View command
- [x] Save data
- [x] Edit data directly
- [x] Exit command
- [x] Archive data
- [x] Command summary
- [x] Param constraints
- [x] Glossary
- [x] FAQ | 1.0 | Update UG for v1.1 - Ensure that UG details are correct and move it into the repo
- [x] Table of contents
- [x] Quick start
- [x] Features
- [x] Add command
- [x] Edit command
- [x] Delete command
- [x] List command
- [x] View command
- [x] Save data
- [x] Edit data directly
- [x] Exit command
- [x] Archive data
- [x] Command summary
- [x] Param constraints
- [x] Glossary
- [x] FAQ | non_infrastructure | update ug for ensure that ug details are correct and move it into the repo table of contents quick start features add command edit command delete command list command view command save data edit data directly exit command archive data command summary param constraints glossary faq | 0 |
51,115 | 10,587,531,314 | IssuesEvent | 2019-10-08 22:26:18 | TauCetiStation/TauCetiClassic | https://api.github.com/repos/TauCetiStation/TauCetiClassic | closed | [Proposal]ะะฐ ัะตัะฒะตัะต ะดะตะฑะธะปั ะฝะฐ ัะพััะผะต | ÿ Admin Awaiting Author Bad Issue Description Balance Bug Code Improvements Contentious Could Not Reproduce DO NOT MERGE Duplicate Issue Experimental Exploit FEATURE FREEZE Feature Fix Global Problems Good First Issue HONK Help Wanted I ded MAP FREEZE Maintainability Improvements Map Edit Map Issue Map PR With No Screenshot Mapmerge/Mapedit Fail Merge Conflict Merge Ready Not a Bug Performance Port Request Proposal Refactor Resprite Revert / Removal Sprite Needs Sprites Stalled PR Test Feedback Test Merge Candidate Tools Tweak Work In Progress | #### ะะพะดัะพะฑะฝะพะต ะพะฟะธัะฐะฝะธะต ะฟัะพะฑะปะตะผั
ะะ ัะตัะฒะตัะต ะฟะตะดะฐะปัั ะณะพะฒะฝะพะตะดั
#### ะงัะพ ะดะพะปะถะฝะพ ะฑัะปะพ ะฟัะพะธะทะพะนัะธ
ะะฐ ัะตัะฒะตัะต ะฝะต ะดะพะปะถะฝั ะฟะตะดะฐะปะธัั ะณะพะฒะฝะพะตะดั
#### ะงัะพ ะฟัะพะธะทะพัะปะพ ะฝะฐ ัะฐะผะพะผ ะดะตะปะต
ะ ะฐะดะผะธะฝะธัััะฐัะธะธ ะพะดะฝะธ ะณะพะฒะฝะพะตะดั
#### ะะฐะบ ะฟะพะฒัะพัะธัั
ะะพััะฐะฒะธัั ะฝะฐ ััะพะฝ ัััะฟะพะบัะฐัะธั
#### ะะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั:
ะะฝะต ะฟะพั
ัะน, ั ะฟะพะปะบะพะฒะฝะธะบ
https://pastebin.com/fqkgx3mG
https://i.imgur.com/CKH6SEn.png | 1.0 | [Proposal]ะะฐ ัะตัะฒะตัะต ะดะตะฑะธะปั ะฝะฐ ัะพััะผะต - #### ะะพะดัะพะฑะฝะพะต ะพะฟะธัะฐะฝะธะต ะฟัะพะฑะปะตะผั
ะะ ัะตัะฒะตัะต ะฟะตะดะฐะปัั ะณะพะฒะฝะพะตะดั
#### ะงัะพ ะดะพะปะถะฝะพ ะฑัะปะพ ะฟัะพะธะทะพะนัะธ
ะะฐ ัะตัะฒะตัะต ะฝะต ะดะพะปะถะฝั ะฟะตะดะฐะปะธัั ะณะพะฒะฝะพะตะดั
#### ะงัะพ ะฟัะพะธะทะพัะปะพ ะฝะฐ ัะฐะผะพะผ ะดะตะปะต
ะ ะฐะดะผะธะฝะธัััะฐัะธะธ ะพะดะฝะธ ะณะพะฒะฝะพะตะดั
#### ะะฐะบ ะฟะพะฒัะพัะธัั
ะะพััะฐะฒะธัั ะฝะฐ ััะพะฝ ัััะฟะพะบัะฐัะธั
#### ะะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั:
ะะฝะต ะฟะพั
ัะน, ั ะฟะพะปะบะพะฒะฝะธะบ
https://pastebin.com/fqkgx3mG
https://i.imgur.com/CKH6SEn.png | non_infrastructure | ะฝะฐ ัะตัะฒะตัะต ะดะตะฑะธะปั ะฝะฐ ัะพััะผะต ะฟะพะดัะพะฑะฝะพะต ะพะฟะธัะฐะฝะธะต ะฟัะพะฑะปะตะผั ะฝะฐ ัะตัะฒะตัะต ะฟะตะดะฐะปัั ะณะพะฒะฝะพะตะดั ััะพ ะดะพะปะถะฝะพ ะฑัะปะพ ะฟัะพะธะทะพะนัะธ ะฝะฐ ัะตัะฒะตัะต ะฝะต ะดะพะปะถะฝั ะฟะตะดะฐะปะธัั ะณะพะฒะฝะพะตะดั ััะพ ะฟัะพะธะทะพัะปะพ ะฝะฐ ัะฐะผะพะผ ะดะตะปะต ะฒ ะฐะดะผะธะฝะธัััะฐัะธะธ ะพะดะฝะธ ะณะพะฒะฝะพะตะดั ะบะฐะบ ะฟะพะฒัะพัะธัั ะฟะพััะฐะฒะธัั ะฝะฐ ััะพะฝ ัััะฟะพะบัะฐัะธั ะดะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั ะผะฝะต ะฟะพั
ัะน ั ะฟะพะปะบะพะฒะฝะธะบ | 0 |
388,100 | 26,751,459,519 | IssuesEvent | 2023-01-30 19:58:57 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | Documentation: Impossible to use search with other languages. | comp-documentation | Two issues, kinda.
1. When you switch a language, it redirects you to the main page (in the old documentation you stayed in the same article).
2. Search always returns results with links to EN site it does not matter which language is current.
Select Russian. Search for quantileExact, you will get EN link
`https://clickhouse.com/docs/en/sql-reference/aggregate-functions/reference/quantileexact#quantileexact`
then switch language to Russian you will get `https://clickhouse.com/docs/ru`.
Desired link is `https://clickhouse.com/docs/ru/sql-reference/aggregate-functions/reference/quantileexact#quantileexact` | 1.0 | Documentation: Impossible to use search with other languages. - Two issues, kinda.
1. When you switch a language, it redirects you to the main page (in the old documentation you stayed in the same article).
2. Search always returns results with links to EN site it does not matter which language is current.
Select Russian. Search for quantileExact, you will get EN link
`https://clickhouse.com/docs/en/sql-reference/aggregate-functions/reference/quantileexact#quantileexact`
then switch language to Russian you will get `https://clickhouse.com/docs/ru`.
Desired link is `https://clickhouse.com/docs/ru/sql-reference/aggregate-functions/reference/quantileexact#quantileexact` | non_infrastructure | documentation impossible to use search with other languages two issues kinda when you switch a language it redirects you to the main page in the old documentation you stayed in the same article search always returns results with links to en site it does not matter which language is current select russian search for quantileexact you will get en link then switch language to russian you will get desired link is | 0 |
32,333 | 26,633,993,203 | IssuesEvent | 2023-01-24 20:13:19 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Links in yml file to https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema lead to a 404 error | area-infrastructure | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
Links to,
[https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema](https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema)
are found in several yaml files, but they lead to a 404 error when followed.
This behaviour is the same if learn.microsoft.com instead of docs.microsoft.com
### Expected Behavior
Upon following the links in the files, a web page containing information should open.
### Steps To Reproduce
Just follow the links to,
[https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema](https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema)
in the yml files in aspnetcore/.azure/pipelines
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_ | 1.0 | Links in yml file to https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema lead to a 404 error - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
Links to,
[https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema](https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema)
are found in several yaml files, but they lead to a 404 error when followed.
This behaviour is the same if learn.microsoft.com instead of docs.microsoft.com
### Expected Behavior
Upon following the links in the files, a web page containing information should open.
### Steps To Reproduce
Just follow the links to,
[https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema](https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema)
in the yml files in aspnetcore/.azure/pipelines
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_ | infrastructure | links in yml file to lead to a error is there an existing issue for this i have searched the existing issues describe the bug links to are found in several yaml files but they lead to a error when followed this behaviour is the same if learn microsoft com instead of docs microsoft com expected behavior upon following the links in the files a web page containing information should open steps to reproduce just follow the links to in the yml files in aspnetcore azure pipelines exceptions if any no response net version no response anything else no response | 1 |
25,638 | 18,944,893,402 | IssuesEvent | 2021-11-18 09:07:52 | ExeClim/Isca | https://api.github.com/repos/ExeClim/Isca | opened | This module file was not generated by any release of this compiler. | infrastructure | Dear Sir,
we ran into problems while compiling the test-experiment for the first time. We tried reinstalling the model several times, but compilation stops due to the following error: โThis module file was not generated by any release of this compilerโ. Although we have checked, and netcdf in our server is compiled with this compiler.
I am attaching the failed compilation log file, as well as link to the file emps-gv with our configuration:
[held_suarez_test_case.log](https://github.com/ExeClim/Isca/files/7561233/held_suarez_test_case.log)
https://drive.google.com/file/d/1d5I3F0BdwMu3O3CXtiUseWaDUwt60HHw/view?usp=sharing
To install the model, we used the instructions from the site and did the following:
First of all, we make sure the conda is completely removed! That is, there is no mention in -bash_profile, .bashrc and there is no .conda folder and .condarc file in the home directory. Then we install newest version of it:
```
wget https://repo.anaconda.com/archive/Anaconda3-2021.05-Linux-x86_64.sh
chmod +x Anaconda3-2021.05-Linux-x86_64.sh
# Install into /storage/workstation/zyulyaeva/anaconda3
conda config --set auto_activate_base false
```
Then we install ISCA:
```
conda activate base
git clone https://github.com/ExeClim/Isca
cd Isca
conda env create -f ci/environment-py3.9.yml # ัััะฐะฝะฐะฒะปะธะฒะฐะตะผ ะพะบััะถะตะฝะธะต
conda activate isca_env โ ะฐะบัะธะฒะธััะตะผ ะพะบััะถะตะฝะธะต
cd src/extra/python/
pip install -e .
```
We write the paths to the folders in .bashrc, exit the server, log in again and compile test experiment:
```
conda activate isca_env
cd $GFDL_BASE/exp/test_cases/held_suarez
python held_suarez_test_case.py
```
And it stops with the aforementioned error.
Maybe you can tell us what are we doing wrong? | 1.0 | This module file was not generated by any release of this compiler. - Dear Sir,
we ran into problems while compiling the test-experiment for the first time. We tried reinstalling the model several times, but compilation stops due to the following error: โThis module file was not generated by any release of this compilerโ. Although we have checked, and netcdf in our server is compiled with this compiler.
I am attaching the failed compilation log file, as well as link to the file emps-gv with our configuration:
[held_suarez_test_case.log](https://github.com/ExeClim/Isca/files/7561233/held_suarez_test_case.log)
https://drive.google.com/file/d/1d5I3F0BdwMu3O3CXtiUseWaDUwt60HHw/view?usp=sharing
To install the model, we used the instructions from the site and did the following:
First of all, we make sure the conda is completely removed! That is, there is no mention in -bash_profile, .bashrc and there is no .conda folder and .condarc file in the home directory. Then we install newest version of it:
```
wget https://repo.anaconda.com/archive/Anaconda3-2021.05-Linux-x86_64.sh
chmod +x Anaconda3-2021.05-Linux-x86_64.sh
# Install into /storage/workstation/zyulyaeva/anaconda3
conda config --set auto_activate_base false
```
Then we install ISCA:
```
conda activate base
git clone https://github.com/ExeClim/Isca
cd Isca
conda env create -f ci/environment-py3.9.yml # ัััะฐะฝะฐะฒะปะธะฒะฐะตะผ ะพะบััะถะตะฝะธะต
conda activate isca_env โ ะฐะบัะธะฒะธััะตะผ ะพะบััะถะตะฝะธะต
cd src/extra/python/
pip install -e .
```
We write the paths to the folders in .bashrc, exit the server, log in again and compile test experiment:
```
conda activate isca_env
cd $GFDL_BASE/exp/test_cases/held_suarez
python held_suarez_test_case.py
```
And it stops with the aforementioned error.
Maybe you can tell us what are we doing wrong? | infrastructure | this module file was not generated by any release of this compiler dear sir we ran into problems while compiling the test experiment for the first time we tried reinstalling the model several times but compilation stops due to the following error โthis module file was not generated by any release of this compilerโ although we have checked and netcdf in our server is compiled with this compiler i am attaching the failed compilation log file as well as link to the file emps gv with our configuration to install the model we used the instructions from the site and did the following first of all we make sure the conda is completely removed that is there is no mention in bash profile bashrc and there is no conda folder and condarc file in the home directory then we install newest version of it wget chmod x linux sh install into storage workstation zyulyaeva conda config set auto activate base false then we install isca conda activate base git clone cd isca conda env create f ci environment yml ัััะฐะฝะฐะฒะปะธะฒะฐะตะผ ะพะบััะถะตะฝะธะต conda activate isca env โ ะฐะบัะธะฒะธััะตะผ ะพะบััะถะตะฝะธะต cd src extra python pip install e we write the paths to the folders in bashrc exit the server log in again and compile test experiment conda activate isca env cd gfdl base exp test cases held suarez python held suarez test case py and it stops with the aforementioned error maybe you can tell us what are we doing wrong | 1 |
526,578 | 15,296,116,835 | IssuesEvent | 2021-02-24 06:16:01 | AY2021S2-CS2103T-W15-4/tp | https://api.github.com/repos/AY2021S2-CS2103T-W15-4/tp | opened | Update README page to reflect Focuris' goals | priority.High type.Enhancement | ## Details
- Update UI Mockup
- Update contents to make sure they reflect our project and not AB3
- Update link of Github status badge
- Acknowledge original source of code | 1.0 | Update README page to reflect Focuris' goals - ## Details
- Update UI Mockup
- Update contents to make sure they reflect our project and not AB3
- Update link of Github status badge
- Acknowledge original source of code | non_infrastructure | update readme page to reflect focuris goals details update ui mockup update contents to make sure they reflect our project and not update link of github status badge acknowledge original source of code | 0 |
12,009 | 9,548,100,197 | IssuesEvent | 2019-05-02 03:13:03 | servo/servo | https://api.github.com/repos/servo/servo | opened | Linux nightly s3 upload is broken | A-infrastructure I-bustage | ```
aws s3 sync --size-only --acl public-read ./etc/ci/performance/output s3://servo-perf
in dir /home/servo/buildbot/slave/linux-nightly/build (timeout 1200 secs)
watching logfiles {}
argv: ['aws', 's3', 'sync', '--size-only', '--acl', 'public-read', './etc/ci/performance/output', 's3://servo-perf']
using PTY: False
Traceback (most recent call last):
File "/usr/local/bin/aws", line 19, in <module>
import awscli.clidriver
File "/usr/local/lib/python2.7/dist-packages/awscli/clidriver.py", line 28, in <module>
from awscli.formatter import get_formatter
File "/usr/local/lib/python2.7/dist-packages/awscli/formatter.py", line 20, in <module>
from awscli.table import MultiTable, Styler, ColorizedStyler
File "/usr/local/lib/python2.7/dist-packages/awscli/table.py", line 17, in <module>
import colorama
ImportError: No module named colorama
``` | 1.0 | Linux nightly s3 upload is broken - ```
aws s3 sync --size-only --acl public-read ./etc/ci/performance/output s3://servo-perf
in dir /home/servo/buildbot/slave/linux-nightly/build (timeout 1200 secs)
watching logfiles {}
argv: ['aws', 's3', 'sync', '--size-only', '--acl', 'public-read', './etc/ci/performance/output', 's3://servo-perf']
using PTY: False
Traceback (most recent call last):
File "/usr/local/bin/aws", line 19, in <module>
import awscli.clidriver
File "/usr/local/lib/python2.7/dist-packages/awscli/clidriver.py", line 28, in <module>
from awscli.formatter import get_formatter
File "/usr/local/lib/python2.7/dist-packages/awscli/formatter.py", line 20, in <module>
from awscli.table import MultiTable, Styler, ColorizedStyler
File "/usr/local/lib/python2.7/dist-packages/awscli/table.py", line 17, in <module>
import colorama
ImportError: No module named colorama
``` | infrastructure | linux nightly upload is broken aws sync size only acl public read etc ci performance output servo perf in dir home servo buildbot slave linux nightly build timeout secs watching logfiles argv using pty false traceback most recent call last file usr local bin aws line in import awscli clidriver file usr local lib dist packages awscli clidriver py line in from awscli formatter import get formatter file usr local lib dist packages awscli formatter py line in from awscli table import multitable styler colorizedstyler file usr local lib dist packages awscli table py line in import colorama importerror no module named colorama | 1 |
348,225 | 31,474,701,852 | IssuesEvent | 2023-08-30 09:52:51 | IfcOpenShell/IfcOpenShell | https://api.github.com/repos/IfcOpenShell/IfcOpenShell | closed | If the value of relations is uppercase, an error occurs. | IfcTester | I checked IDS in IfcOpenShell, and the result is different depending on whether I write IfcRelAggregates or IFCRELAGGREGATES in the PartOf facet.
If I write IfcRelAggregates in the relations, IfcTester works.
However, if I write IFCRELAGGREGATES in relations, IfcTester gives an error.
I checked the contents of the XSD and it says that I have to write IFCRELAGGREGATES for relations. (i.e. all uppercase).
Why does it give me an error if I write it in capitals?
<ids:partOf maxOccurs="unbounded" relation="IFCRELAGGREGATES"> | 1.0 | If the value of relations is uppercase, an error occurs. - I checked IDS in IfcOpenShell, and the result is different depending on whether I write IfcRelAggregates or IFCRELAGGREGATES in the PartOf facet.
If I write IfcRelAggregates in the relations, IfcTester works.
However, if I write IFCRELAGGREGATES in relations, IfcTester gives an error.
I checked the contents of the XSD and it says that I have to write IFCRELAGGREGATES for relations. (i.e. all uppercase).
Why does it give me an error if I write it in capitals?
<ids:partOf maxOccurs="unbounded" relation="IFCRELAGGREGATES"> | non_infrastructure | if the value of relations is uppercase an error occurs i checked ids in ifcopenshell and the result is different depending on whether i write ifcrelaggregates or ifcrelaggregates in the partof facet if i write ifcrelaggregates in the relations ifctester works however if i write ifcrelaggregates in relations ifctester gives an error i checked the contents of the xsd and it says that i have to write ifcrelaggregates for relations i e all uppercase why does it give me an error if i write it in capitals | 0 |
54,272 | 23,222,575,332 | IssuesEvent | 2022-08-02 19:46:09 | openstreetmap/operations | https://api.github.com/repos/openstreetmap/operations | closed | Don't reserve ext4 space on non-root tile server storage | service:tiles | /store on the tile servers with separate boot and storage volumes has reserved space for the superuser. This makes sense on boot drives, but is using space that could be used for the tile store. It would be best to remove the reserved space.
Pyrene only has one volume, but 130 GiB reserved is more than it needs to be. | 1.0 | Don't reserve ext4 space on non-root tile server storage - /store on the tile servers with separate boot and storage volumes has reserved space for the superuser. This makes sense on boot drives, but is using space that could be used for the tile store. It would be best to remove the reserved space.
Pyrene only has one volume, but 130 GiB reserved is more than it needs to be. | non_infrastructure | don t reserve space on non root tile server storage store on the tile servers with separate boot and storage volumes has reserved space for the superuser this makes sense on boot drives but is using space that could be used for the tile store it would be best to remove the reserved space pyrene only has one volume but gib reserved is more than it needs to be | 0 |
15,563 | 8,958,756,293 | IssuesEvent | 2019-01-27 16:52:03 | graphprotocol/graph-node | https://api.github.com/repos/graphprotocol/graph-node | opened | Soft-reset the module between handler calls | area/runtime performance | After #702 we re-instantiate the WASM runtime between each call, to free memory and ensure no state is carried over. This works well, but it can cost us a couple milliseconds on each call. A soft reset of the module, which means resetting memory contents and size, and also global state, would likely be more performant. This was attempted in [this branch](https://github.com/graphprotocol/graph-node/blob/leo/reset-memory/runtime/wasm/src/module/mod.rs) with no success. Parity needs to do similar things with their runtime, so we may be able to cooperate with them when revisiting this in the future. | True | Soft-reset the module between handler calls - After #702 we re-instantiate the WASM runtime between each call, to free memory and ensure no state is carried over. This works well, but it can cost us a couple milliseconds on each call. A soft reset of the module, which means resetting memory contents and size, and also global state, would likely be more performant. This was attempted in [this branch](https://github.com/graphprotocol/graph-node/blob/leo/reset-memory/runtime/wasm/src/module/mod.rs) with no success. Parity needs to do similar things with their runtime, so we may be able to cooperate with them when revisiting this in the future. | non_infrastructure | soft reset the module between handler calls after we re instantiate the wasm runtime between each call to free memory and ensure no state is carried over this works well but it can cost us a couple milliseconds on each call a soft reset of the module which means resetting memory contents and size and also global state would likely be more performant this was attempted in with no success parity needs to do similar things with their runtime so we may be able to cooperate with them when revisiting this in the future | 0 |
9,844 | 8,201,545,553 | IssuesEvent | 2018-09-01 19:01:12 | triplea-game/triplea | https://api.github.com/repos/triplea-game/triplea | opened | Use pure systemd instead of sysvinit "wrapper" | Infrastructure | Migrated from: https://github.com/triplea-game/infrastructure/issues/40
> Just a small note:
We currently use the service triplea-bot@<number> start|stop|restart|status etc. command to start and stop bots, however this is the "legacy" SysVinit syntax that is kept for backwards compatibility.
We should prefer the systemd syntax systemctl start|stop|restart|reload|status triplea-bot@<number> to be more consistent with all the other systemd related commands.
When changing this we should tell all the people having access (like prastle) about this | 1.0 | Use pure systemd instead of sysvinit "wrapper" - Migrated from: https://github.com/triplea-game/infrastructure/issues/40
> Just a small note:
We currently use the service triplea-bot@<number> start|stop|restart|status etc. command to start and stop bots, however this is the "legacy" SysVinit syntax that is kept for backwards compatibility.
We should prefer the systemd syntax systemctl start|stop|restart|reload|status triplea-bot@<number> to be more consistent with all the other systemd related commands.
When changing this we should tell all the people having access (like prastle) about this | infrastructure | use pure systemd instead of sysvinit wrapper migrated from just a small note we currently use the service triplea bot start stop restart status etc command to start and stop bots however this is the legacy sysvinit syntax that is kept for backwards compatibility we should prefer the systemd syntax systemctl start stop restart reload status triplea bot to be more consistent with all the other systemd related commands when changing this we should tell all the people having access like prastle about this | 1 |
2,554 | 3,755,641,133 | IssuesEvent | 2016-03-12 20:08:33 | catapult-project/catapult | https://api.github.com/repos/catapult-project/catapult | opened | Use dependency manager for getting SDK for perf dashboard. | Dependency Manager Infrastructure Perf Dashboard | Currently, to run the perf dashboard tests or deploy or start a local instance requires having the App Engine SDK directory in PYTHONPATH.
This adds a barrier for people to contribute, and makes it so that the SDK is required to be set up a certain way before the tests can be run. It also requires set-up before any pylint checks that try importing dependency libraries can be run (such as import-error; which we want to enable, see #1888).
This is related to #1407 (migrate to Google Cloud SDK), but neither task is necessarily blocking the other. | 1.0 | Use dependency manager for getting SDK for perf dashboard. - Currently, to run the perf dashboard tests or deploy or start a local instance requires having the App Engine SDK directory in PYTHONPATH.
This adds a barrier for people to contribute, and makes it so that the SDK is required to be set up a certain way before the tests can be run. It also requires set-up before any pylint checks that try importing dependency libraries can be run (such as import-error; which we want to enable, see #1888).
This is related to #1407 (migrate to Google Cloud SDK), but neither task is necessarily blocking the other. | infrastructure | use dependency manager for getting sdk for perf dashboard currently to run the perf dashboard tests or deploy or start a local instance requires having the app engine sdk directory in pythonpath this adds a barrier for people to contribute and makes it so that the sdk is required to be set up a certain way before the tests can be run it also requires set up before any pylint checks that try importing dependency libraries can be run such as import error which we want to enable see this is related to migrate to google cloud sdk but neither task is necessarily blocking the other | 1 |
54,091 | 13,894,689,801 | IssuesEvent | 2020-10-19 14:58:30 | jgeraigery/thingworx-gitbackup-extension | https://api.github.com/repos/jgeraigery/thingworx-gitbackup-extension | opened | CVE-2019-20330 (High) detected in jackson-databind-2.9.8.jar | security vulnerability | ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-20330 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library thingworx gitbackup extension twx lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before lacks certain net sf ehcache blocking publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before lacks certain net sf ehcache blocking vulnerabilityurl | 0 |
30,865 | 25,126,748,844 | IssuesEvent | 2022-11-09 12:20:19 | GoogleForCreators/web-stories-wp | https://api.github.com/repos/GoogleForCreators/web-stories-wp | closed | TypeScript: Use with Jest | P2 Type: Infrastructure Pod: WP | Let's use [ts-jest](https://github.com/kulshekhar/ts-jest) for improved TypeScript support in Jest so that we can also write unit tests with TypeScript โwith proper type checking and not just transpilation.
See https://kulshekhar.github.io/ts-jest/docs/babel7-or-ts/
See https://kulshekhar.github.io/ts-jest/docs/27.0/guides/test-helpers/
* [ ] Update ESLint config
* [ ] Add `preset: 'ts-jest'` to Jest config | 1.0 | TypeScript: Use with Jest - Let's use [ts-jest](https://github.com/kulshekhar/ts-jest) for improved TypeScript support in Jest so that we can also write unit tests with TypeScript โwith proper type checking and not just transpilation.
See https://kulshekhar.github.io/ts-jest/docs/babel7-or-ts/
See https://kulshekhar.github.io/ts-jest/docs/27.0/guides/test-helpers/
* [ ] Update ESLint config
* [ ] Add `preset: 'ts-jest'` to Jest config | infrastructure | typescript use with jest let s use for improved typescript support in jest so that we can also write unit tests with typescript โwith proper type checking and not just transpilation see see update eslint config add preset ts jest to jest config | 1 |
18,292 | 12,884,271,964 | IssuesEvent | 2020-07-13 02:22:57 | FullScreenShenanigans/EightBittr | https://api.github.com/repos/FullScreenShenanigans/EightBittr | opened | Enable a lint rule against "unnecessary" (inferable) type annotations | status: blocked type: infrastructure | Blocked on https://github.com/typescript-eslint/typescript-eslint/issues/295 for actually creating the rule. I used to be a big proponent of enabling [TSLint's `typedef`](https://palantir.github.io/tslint/rules/typedef) many years ago, back when TypeScript was less powerful at inferring types. I have since converted to the opposite side - that they're practically almost always useless, except for large object definitions that benefit from improved intellisense. | 1.0 | Enable a lint rule against "unnecessary" (inferable) type annotations - Blocked on https://github.com/typescript-eslint/typescript-eslint/issues/295 for actually creating the rule. I used to be a big proponent of enabling [TSLint's `typedef`](https://palantir.github.io/tslint/rules/typedef) many years ago, back when TypeScript was less powerful at inferring types. I have since converted to the opposite side - that they're practically almost always useless, except for large object definitions that benefit from improved intellisense. | infrastructure | enable a lint rule against unnecessary inferable type annotations blocked on for actually creating the rule i used to be a big proponent of enabling many years ago back when typescript was less powerful at inferring types i have since converted to the opposite side that they re practically almost always useless except for large object definitions that benefit from improved intellisense | 1 |
33,842 | 27,873,549,613 | IssuesEvent | 2023-03-21 14:49:01 | camunda/camunda-modeler | https://api.github.com/repos/camunda/camunda-modeler | opened | Throttle events sent by Sentry for plugin-triggered errors | infrastructure backlog spring cleaning | ### What should we do?
For some Sentry events, we can see from the call stack that they originated in a plugin, e.g.: `~/username/client-bundle.js`. Let's limit the number of such events sent to Sentry.
Cf. https://camunda-modeling.sentry.io/issues/1673760145/?project=5223041&query=is%3Aunresolved&referrer=issue-stream
### Why should we do it?
We don't want to flood Sentry with errors which we cannot really fix. | 1.0 | Throttle events sent by Sentry for plugin-triggered errors - ### What should we do?
For some Sentry events, we can see from the call stack that they originated in a plugin, e.g.: `~/username/client-bundle.js`. Let's limit the number of such events sent to Sentry.
Cf. https://camunda-modeling.sentry.io/issues/1673760145/?project=5223041&query=is%3Aunresolved&referrer=issue-stream
### Why should we do it?
We don't want to flood Sentry with errors which we cannot really fix. | infrastructure | throttle events sent by sentry for plugin triggered errors what should we do for some sentry events we can see from the call stack that they originated in a plugin e g username client bundle js let s limit the number of such events sent to sentry cf why should we do it we don t want to flood sentry with errors which we cannot really fix | 1 |
5,892 | 6,025,876,346 | IssuesEvent | 2017-06-08 09:47:50 | vitalyster/SharpXMPP | https://api.github.com/repos/vitalyster/SharpXMPP | closed | Upgrade dotnet build system | kind:infrastructure kind:refactor | We need to migrate from that nasty `project.json` back to orthodox `csproj`. | 1.0 | Upgrade dotnet build system - We need to migrate from that nasty `project.json` back to orthodox `csproj`. | infrastructure | upgrade dotnet build system we need to migrate from that nasty project json back to orthodox csproj | 1 |
11,848 | 9,465,837,563 | IssuesEvent | 2019-04-18 01:33:48 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | [eng] Optimus in Prod! | Q1-2019 infrastructure needs refinestimation optimus | After Optimus is completed and approved, we will want to start running it in prod.
AC:
- mintegration and integration tests test optimus pipeline (do we want the integration test to always test all pipelines?)
- Subscriptions turned on
- Cut a v1.0 and release Lira | 1.0 | [eng] Optimus in Prod! - After Optimus is completed and approved, we will want to start running it in prod.
AC:
- mintegration and integration tests test optimus pipeline (do we want the integration test to always test all pipelines?)
- Subscriptions turned on
- Cut a v1.0 and release Lira | infrastructure | optimus in prod after optimus is completed and approved we will want to start running it in prod ac mintegration and integration tests test optimus pipeline do we want the integration test to always test all pipelines subscriptions turned on cut a and release lira | 1 |
12,865 | 9,984,379,182 | IssuesEvent | 2019-07-10 14:24:42 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | opened | Verify a SpringBoot application can use Liberty datasources | in:SpringBoot team:OSGi Infrastructure | Develop a test case containing a SpringBoot application that demonstrates the use of a Liberty datasource. | 1.0 | Verify a SpringBoot application can use Liberty datasources - Develop a test case containing a SpringBoot application that demonstrates the use of a Liberty datasource. | infrastructure | verify a springboot application can use liberty datasources develop a test case containing a springboot application that demonstrates the use of a liberty datasource | 1 |
6,495 | 6,484,621,999 | IssuesEvent | 2017-08-19 01:38:17 | TimothyLuke/GnomeSequencer-Enhanced | https://api.github.com/repos/TimothyLuke/GnomeSequencer-Enhanced | closed | Refactor to move GUI back to Sub Mod | Infrastructure | In GSE1 the GUI was a seperate mod. This got muddled with interdependencies. To improve performance change the GUI to be a seperate mod. Also things like LDB can be a seperate mod of its own. Also explore loading the GUI on Demand
- [x] LDB Module
- [x] GUI Module
- [x] Move Selectable Panel to its own repository and make it a separate library | 1.0 | Refactor to move GUI back to Sub Mod - In GSE1 the GUI was a seperate mod. This got muddled with interdependencies. To improve performance change the GUI to be a seperate mod. Also things like LDB can be a seperate mod of its own. Also explore loading the GUI on Demand
- [x] LDB Module
- [x] GUI Module
- [x] Move Selectable Panel to its own repository and make it a separate library | infrastructure | refactor to move gui back to sub mod in the gui was a seperate mod this got muddled with interdependencies to improve performance change the gui to be a seperate mod also things like ldb can be a seperate mod of its own also explore loading the gui on demand ldb module gui module move selectable panel to its own repository and make it a separate library | 1 |
34,434 | 29,929,266,344 | IssuesEvent | 2023-06-22 08:19:31 | recallwei/dolphin-admin-web | https://api.github.com/repos/recallwei/dolphin-admin-web | opened | Feat: Theme | Type: Feature Infrastructure | - [ ] Basic dolphin theme.
- [ ] Support theme switch.
- [ ] Dark mode.
- [ ] Auto-detect device preference.
- [ ] Remember the theme which the user picked. | 1.0 | Feat: Theme - - [ ] Basic dolphin theme.
- [ ] Support theme switch.
- [ ] Dark mode.
- [ ] Auto-detect device preference.
- [ ] Remember the theme which the user picked. | infrastructure | feat theme basic dolphin theme support theme switch dark mode auto detect device preference remember the theme which the user picked | 1 |
527,185 | 15,325,207,657 | IssuesEvent | 2021-02-26 00:46:57 | sButtons/sbuttons | https://api.github.com/repos/sButtons/sbuttons | opened | [BUTTON IDEA]: | Priority: Low button-idea | **Dropdown buttons**:
**Example of using a dropdown menu with a button:*:
**The dropdown animation happen downward side**:
**Yes**:
| 1.0 | [BUTTON IDEA]: - **Dropdown buttons**:
**Example of using a dropdown menu with a button:*:
**The dropdown animation happen downward side**:
**Yes**:
| non_infrastructure | dropdown buttons example of using a dropdown menu with a button the dropdown animation happen downward side yes | 0 |
28,872 | 23,543,909,414 | IssuesEvent | 2022-08-20 20:45:08 | skyportal/skyportal | https://api.github.com/repos/skyportal/skyportal | closed | Run the CI daily to catch potential external-stuff-related issues early | infrastructure | ...such as broken dependencies, for example. Doesn't have to be the complete suite, maybe just a subset thereof.
Add
```
schedule:
- cron: '* 12 * * *'
```
or something to the GHA workflow defs? | 1.0 | Run the CI daily to catch potential external-stuff-related issues early - ...such as broken dependencies, for example. Doesn't have to be the complete suite, maybe just a subset thereof.
Add
```
schedule:
- cron: '* 12 * * *'
```
or something to the GHA workflow defs? | infrastructure | run the ci daily to catch potential external stuff related issues early such as broken dependencies for example doesn t have to be the complete suite maybe just a subset thereof add schedule cron or something to the gha workflow defs | 1 |
17,822 | 12,604,858,450 | IssuesEvent | 2020-06-11 15:35:13 | Budibase/budibase | https://api.github.com/repos/Budibase/budibase | closed | Serverless Deployment POC | backend frontend infrastructure | We should understand the work involved to get Budibase up and running in a serverless environment for beta.
This POC would include:
- Creation of custom subdomain in route53
- Upload of assets to s3
- use of aws-serverless-express to wrap the Koa app | 1.0 | Serverless Deployment POC - We should understand the work involved to get Budibase up and running in a serverless environment for beta.
This POC would include:
- Creation of custom subdomain in route53
- Upload of assets to s3
- use of aws-serverless-express to wrap the Koa app | infrastructure | serverless deployment poc we should understand the work involved to get budibase up and running in a serverless environment for beta this poc would include creation of custom subdomain in upload of assets to use of aws serverless express to wrap the koa app | 1 |
106,945 | 16,736,719,414 | IssuesEvent | 2021-06-11 03:14:03 | atlslscsrv-app/upgraded-fiesta | https://api.github.com/repos/atlslscsrv-app/upgraded-fiesta | closed | WS-2019-0170 (Medium) detected in serve-9.6.0.tgz | no-issue-activity security vulnerability | ## WS-2019-0170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serve-9.6.0.tgz</b></p></summary>
<p>Static file serving and directory listing</p>
<p>Library home page: <a href="https://registry.npmjs.org/serve/-/serve-9.6.0.tgz">https://registry.npmjs.org/serve/-/serve-9.6.0.tgz</a></p>
<p>Path to dependency file: upgraded-fiesta/package.json</p>
<p>Path to vulnerable library: upgraded-fiesta/node_modules/serve/package.json</p>
<p>
Dependency Hierarchy:
- :x: **serve-9.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/atlslscsrv-app/upgraded-fiesta/commits/ec68291ce25200691b7a48431969b7c8021b35e2">ec68291ce25200691b7a48431969b7c8021b35e2</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
serve before 1.0.2 vulnerable to cross-site scripting when not ancdoing output, allowing attackers to execute arbitrary JavaScript in the victim's browser if user-supplied input is rendered.
<p>Publish Date: 2018-09-24
<p>URL: <a href=https://github.com/zeit/serve-handler/pull/64>WS-2019-0170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/971">https://www.npmjs.com/advisories/971</a></p>
<p>Release Date: 2019-07-15</p>
<p>Fix Resolution: 10.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0170 (Medium) detected in serve-9.6.0.tgz - ## WS-2019-0170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serve-9.6.0.tgz</b></p></summary>
<p>Static file serving and directory listing</p>
<p>Library home page: <a href="https://registry.npmjs.org/serve/-/serve-9.6.0.tgz">https://registry.npmjs.org/serve/-/serve-9.6.0.tgz</a></p>
<p>Path to dependency file: upgraded-fiesta/package.json</p>
<p>Path to vulnerable library: upgraded-fiesta/node_modules/serve/package.json</p>
<p>
Dependency Hierarchy:
- :x: **serve-9.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/atlslscsrv-app/upgraded-fiesta/commits/ec68291ce25200691b7a48431969b7c8021b35e2">ec68291ce25200691b7a48431969b7c8021b35e2</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
serve before 1.0.2 vulnerable to cross-site scripting when not ancdoing output, allowing attackers to execute arbitrary JavaScript in the victim's browser if user-supplied input is rendered.
<p>Publish Date: 2018-09-24
<p>URL: <a href=https://github.com/zeit/serve-handler/pull/64>WS-2019-0170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/971">https://www.npmjs.com/advisories/971</a></p>
<p>Release Date: 2019-07-15</p>
<p>Fix Resolution: 10.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | ws medium detected in serve tgz ws medium severity vulnerability vulnerable library serve tgz static file serving and directory listing library home page a href path to dependency file upgraded fiesta package json path to vulnerable library upgraded fiesta node modules serve package json dependency hierarchy x serve tgz vulnerable library found in head commit a href found in base branch master vulnerability details serve before vulnerable to cross site scripting when not ancdoing output allowing attackers to execute arbitrary javascript in the victim s browser if user supplied input is rendered publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
87,975 | 15,791,394,970 | IssuesEvent | 2021-04-02 04:23:45 | elastic/integrations | https://api.github.com/repos/elastic/integrations | closed | [Windows] Custom Event Channel support | 7.13 candidate Integration:Windows Team:Security-External Integrations enhancement | The Windows/System integrations currently support Application, System, Security, Sysmon and PowerShell events. To complete our winlogbeat migration, we need to ensure users can ingest events from channels outside of our supported channels. Example: Applications and Services Logs \Microsoft\Windows\Applocker\EXE and DLL.
Based on [this design doc](https://docs.google.com/document/d/1U7ZnMYGj5K3wOiYZdph8vrG40ZWOdXNqGwpFf24ZY4Y/edit), our initial plan involved adding Custom Windows Events to the existing Windows integration, but given that multi-input is not currently a priority, this may not be feasible. This workaround may be an option: https://github.com/elastic/kibana/pull/82783
If multi-input is a non-runner, the preference is to provide a Custom Windows Event integration whereby a user adds a single event channel per custom integration.
**Acceptance Criteria**
- Users can ingest Windows Events from custom channels (Windows Integration vs Custom Windows Integration TBD)
- Users can specify the ingest node pipeline to write to. | True | [Windows] Custom Event Channel support - The Windows/System integrations currently support Application, System, Security, Sysmon and PowerShell events. To complete our winlogbeat migration, we need to ensure users can ingest events from channels outside of our supported channels. Example: Applications and Services Logs \Microsoft\Windows\Applocker\EXE and DLL.
Based on [this design doc](https://docs.google.com/document/d/1U7ZnMYGj5K3wOiYZdph8vrG40ZWOdXNqGwpFf24ZY4Y/edit), our initial plan involved adding Custom Windows Events to the existing Windows integration, but given that multi-input is not currently a priority, this may not be feasible. This workaround may be an option: https://github.com/elastic/kibana/pull/82783
If multi-input is a non-runner, the preference is to provide a Custom Windows Event integration whereby a user adds a single event channel per custom integration.
**Acceptance Criteria**
- Users can ingest Windows Events from custom channels (Windows Integration vs Custom Windows Integration TBD)
- Users can specify the ingest node pipeline to write to. | non_infrastructure | custom event channel support the windows system integrations currently support application system security sysmon and powershell events to complete our winlogbeat migration we need to ensure users can ingest events from channels outside of our supported channels example applications and services logs microsoft windows applocker exe and dll based on our initial plan involved adding custom windows events to the existing windows integration but given that multi input is not currently a priority this may not be feasible this workaround may be an option if multi input is a non runner the preference is to provide a custom windows event integration whereby a user adds a single event channel per custom integration acceptance criteria users can ingest windows events from custom channels windows integration vs custom windows integration tbd users can specify the ingest node pipeline to write to | 0 |
199,426 | 15,038,485,275 | IssuesEvent | 2021-02-02 17:29:58 | Scholar-6/brillder | https://api.github.com/repos/Scholar-6/brillder | closed | Tessa Ticket: Line breaks for highlighting | Adjustments Betatester Request | - [x] make text for word and line highlighting smaller in phone preview (same as reg text)
- [x] prevent hyphens ('-') from causing line breaks in phone preview
<img width="1361" alt="Screenshot 2021-01-29 at 13 12 43" src="https://user-images.githubusercontent.com/59654112/106274022-e765e180-6233-11eb-885e-5e5509ad4af7.png">
<img width="400" alt="Screenshot 2021-01-29 at 13 16 17" src="https://user-images.githubusercontent.com/59654112/106274194-2e53d700-6234-11eb-9cd4-3456fa38fcc2.png">
Line breaks should be as follows
Fillet of a fenny snake,
In the cauldron boil and bake;
Eye of newt, and toe of frog,
Wool of bat, and tongue of dog,
Adderโs fork, and blind-wormโs sting,
Lizardโs leg, and howletโs wing,
For a charm of powerful trouble,
Like a hell-broth boil and bubble.
<img width="532" alt="Screenshot 2021-01-29 at 13 28 33" src="https://user-images.githubusercontent.com/59654112/106275336-edf55880-6235-11eb-8f73-1757f8418aba.png">
- [x] place highlighter closer to text and remove outer circle, make default highlighter now dark blue
- [x] when I hover on default highlighter, circle appears so hover makes it looks like default currently does - dark blue outer circle, grey/no fill inside
- [x] In popup when when clicked, pink selector should be dark blue
<img width="208" alt="Screenshot 2021-01-29 at 13 34 15" src="https://user-images.githubusercontent.com/59654112/106275775-b044ff80-6236-11eb-9094-62fb5b2a2cc3.png">
- [x] Add a circle to the right of highlighter with 'br' inside
- [x] Default is off - colours: f5f6f7 circle, acacac text
- [x] background of circle goes white on hover
- [x] if 'yes' clicked for line break, circle goes green, text white
- [x] cool make 'br' text bigger, so same as text on left (click the highlighter.....) | 1.0 | Tessa Ticket: Line breaks for highlighting - - [x] make text for word and line highlighting smaller in phone preview (same as reg text)
- [x] prevent hyphens ('-') from causing line breaks in phone preview
<img width="1361" alt="Screenshot 2021-01-29 at 13 12 43" src="https://user-images.githubusercontent.com/59654112/106274022-e765e180-6233-11eb-885e-5e5509ad4af7.png">
<img width="400" alt="Screenshot 2021-01-29 at 13 16 17" src="https://user-images.githubusercontent.com/59654112/106274194-2e53d700-6234-11eb-9cd4-3456fa38fcc2.png">
Line breaks should be as follows
Fillet of a fenny snake,
In the cauldron boil and bake;
Eye of newt, and toe of frog,
Wool of bat, and tongue of dog,
Adderโs fork, and blind-wormโs sting,
Lizardโs leg, and howletโs wing,
For a charm of powerful trouble,
Like a hell-broth boil and bubble.
<img width="532" alt="Screenshot 2021-01-29 at 13 28 33" src="https://user-images.githubusercontent.com/59654112/106275336-edf55880-6235-11eb-8f73-1757f8418aba.png">
- [x] place highlighter closer to text and remove outer circle, make default highlighter now dark blue
- [x] when I hover on default highlighter, circle appears so hover makes it looks like default currently does - dark blue outer circle, grey/no fill inside
- [x] In popup when when clicked, pink selector should be dark blue
<img width="208" alt="Screenshot 2021-01-29 at 13 34 15" src="https://user-images.githubusercontent.com/59654112/106275775-b044ff80-6236-11eb-9094-62fb5b2a2cc3.png">
- [x] Add a circle to the right of highlighter with 'br' inside
- [x] Default is off - colours: f5f6f7 circle, acacac text
- [x] background of circle goes white on hover
- [x] if 'yes' clicked for line break, circle goes green, text white
- [x] cool make 'br' text bigger, so same as text on left (click the highlighter.....) | non_infrastructure | tessa ticket line breaks for highlighting make text for word and line highlighting smaller in phone preview same as reg text prevent hyphens from causing line breaks in phone preview img width alt screenshot at src img width alt screenshot at src line breaks should be as follows fillet of a fenny snake in the cauldron boil and bake eye of newt and toe of frog wool of bat and tongue of dog adderโs fork and blind wormโs sting lizardโs leg and howletโs wing for a charm of powerful trouble like a hell broth boil and bubble img width alt screenshot at src place highlighter closer to text and remove outer circle make default highlighter now dark blue when i hover on default highlighter circle appears so hover makes it looks like default currently does dark blue outer circle grey no fill inside in popup when when clicked pink selector should be dark blue img width alt screenshot at src add a circle to the right of highlighter with br inside default is off colours circle acacac text background of circle goes white on hover if yes clicked for line break circle goes green text white cool make br text bigger so same as text on left click the highlighter | 0 |
22,409 | 15,169,340,584 | IssuesEvent | 2021-02-12 20:58:19 | dotnet/fsharp | https://api.github.com/repos/dotnet/fsharp | closed | Question: running tests on .NET 5 | Area-Testing Infrastructure | I'm trying to run FSharp.Compiler.Service.Tests on .NET 5, I thought all I'd need is this
ย ย ย ย <TargetFrameworks>netcoreapp3.1;net5.0</TargetFrameworks>
However this fails with the usual kind of weirdness we all expect from MSBuild, with rubbish about .NET Framework
```
Severity Code Description Project File Line Suppression State
Error MSB3644 The reference assemblies for .NETFramework,Version=v5.0 were not found. To resolve this, install the Developer Pack (SDK/Targeting Pack) for this framework version or retarget your application. You can download .NET Framework Developer Packs at https://aka.ms/msbuild/developerpacks FSharp.Compiler.Service.Tests E:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Microsoft.Common.CurrentVersion.targets 1186
```
I think this is because our global.json is pinning us to .NET SDK 3.1.302 and this toolchain doesn't know about .NET 5.
```
"sdk": {
"version": "3.1.302",
"rollForward": "minor"
},
```
I'm concerned that we've got no in-repo coverage of running our tests. Also some tests like those for `GetProjectOptionsFromScript` and anything to do with F# Interactive are fundamentally sensitive to the .NET SDK we are connecting to to find default reference for F# scripts.
This is related to https://github.com/dotnet/fsharp/pull/10575, where I was trying to add net5.0 testing for FSHarp.Compiler.Service.Tests.fsproj | 1.0 | Question: running tests on .NET 5 - I'm trying to run FSharp.Compiler.Service.Tests on .NET 5, I thought all I'd need is this
ย ย ย ย <TargetFrameworks>netcoreapp3.1;net5.0</TargetFrameworks>
However this fails with the usual kind of weirdness we all expect from MSBuild, with rubbish about .NET Framework
```
Severity Code Description Project File Line Suppression State
Error MSB3644 The reference assemblies for .NETFramework,Version=v5.0 were not found. To resolve this, install the Developer Pack (SDK/Targeting Pack) for this framework version or retarget your application. You can download .NET Framework Developer Packs at https://aka.ms/msbuild/developerpacks FSharp.Compiler.Service.Tests E:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Microsoft.Common.CurrentVersion.targets 1186
```
I think this is because our global.json is pinning us to .NET SDK 3.1.302 and this toolchain doesn't know about .NET 5.
```
"sdk": {
"version": "3.1.302",
"rollForward": "minor"
},
```
I'm concerned that we've got no in-repo coverage of running our tests. Also some tests like those for `GetProjectOptionsFromScript` and anything to do with F# Interactive are fundamentally sensitive to the .NET SDK we are connecting to to find default reference for F# scripts.
This is related to https://github.com/dotnet/fsharp/pull/10575, where I was trying to add net5.0 testing for FSHarp.Compiler.Service.Tests.fsproj | infrastructure | question running tests on net i m trying to run fsharp compiler service tests on net i thought all i d need is this ย ย ย ย however this fails with the usual kind of weirdness we all expect from msbuild with rubbish about net framework severity code description project file line suppression state error the reference assemblies for netframework version were not found to resolve this install the developer pack sdk targeting pack for this framework version or retarget your application you can download net framework developer packs at fsharp compiler service tests e program files microsoft visual studio preview msbuild current bin microsoft common currentversion targets i think this is because our global json is pinning us to net sdk and this toolchain doesn t know about net sdk version rollforward minor i m concerned that we ve got no in repo coverage of running our tests also some tests like those for getprojectoptionsfromscript and anything to do with f interactive are fundamentally sensitive to the net sdk we are connecting to to find default reference for f scripts this is related to where i was trying to add testing for fsharp compiler service tests fsproj | 1 |
488,411 | 14,077,119,076 | IssuesEvent | 2020-11-04 11:30:41 | OpenSRP/opensrp-client-reveal | https://api.github.com/repos/OpenSRP/opensrp-client-reveal | opened | Zambia Full APK - Remove "Not Visited Yet" from Spray Status | Priority: High | How to reproduce on IRS Lite
Login to reveal
Select other forms, enter verification form
On the Spray status, remove second "Not sprayed" and, the "Not visited yet" responses
Current behavior:
The option on spray status, reading "Not Visited Yet" should be removed. | 1.0 | Zambia Full APK - Remove "Not Visited Yet" from Spray Status - How to reproduce on IRS Lite
Login to reveal
Select other forms, enter verification form
On the Spray status, remove second "Not sprayed" and, the "Not visited yet" responses
Current behavior:
The option on spray status, reading "Not Visited Yet" should be removed. | non_infrastructure | zambia full apk remove not visited yet from spray status how to reproduce on irs lite login to reveal select other forms enter verification form on the spray status remove second not sprayed and the not visited yet responses current behavior the option on spray status reading not visited yet should be removed | 0 |
12,682 | 9,914,344,328 | IssuesEvent | 2019-06-28 14:10:06 | tempesta-tech/tempesta-test | https://api.github.com/repos/tempesta-tech/tempesta-test | opened | Multy-threaded Deproxy | Infrastructure | The Deproxy is fully written in Python and has a poor performance. As far as i remember, 20-30RPS was its maximum (while wrk+nginx chain can produce 10-12 KRPS). It's not a load generator, but it validates received messages very strictly. But the problem is that we have to develop two types of tests: deproxy-tests with one single message chain and workload tests using `wrk`. These two kind of tests are completely different.
Issues we have faced with Deproxy during pat year:
- Too slow sometimes that we have to increase some timeouts (response wasn't received in time)
- All requests and responses are processed asynchronously in one thread, while some tests require background operations that blocks (especially paramiko library). We've manually added new threads in that tests.
Issues in wrk tests:
- wrk can possibly overload VM, especially if everything is started on the same VM since it doesn't support constant throughput mode.
- need to thoroughly set asserts on wrk counters, it's very painful to find balance between false-negatives and allowed errors rate. Some errors will happen if backend configuration isn't optimal. And it isn't since we try to use backend defaults as long as it possible. This was the most painful work for me.
- we had to add response code counters and check them after tests to identify suspicious situations.
- we had to add socket error counters (connect, read, write) since it happens some time.
- context of lua scripts in wrk is thread-global, used by multiple connections, but per-connection context is more desired to write simple and reliable tests.
- Backend used in the tests is nginx, it's a real server, but something like server-side wrk is highly desired.
- requests|responses|payload may be spoilt but nothing cares about it. Only Http message integrity is checked.
As for me, we should rework the deproxy to make it possible to serve at least 5-10KRPS to test all the possible multy-threaded effects in Tempesta. The tool MUST do full Http message validation as current deproxy does. It's not very hard to do quite fast: wrk already has full (but simple) http parser inside, so does nginx, while the validation doesn't require advanced http parsers, pattern matching is works in the deproxy for a long time. Each deproxy client and server must be a separate process, that gets a script as input (like wrk), writes all the sent and received messages in it's log file, make some assertions and returns eror code (0=Test passed, not-NULL=failed). The test framework must analyse only that returns codes.
That will allow to build both functional and workload code with one piece of code, only difference - is the concurrent connections parameter.
TfwBomber was the tool intended to replace the wrk and give more abilities for fuzzing, but it doesn't care about server side. It reuses a lot of Tempesta code. It can be useful. | 1.0 | Multy-threaded Deproxy - The Deproxy is fully written in Python and has a poor performance. As far as i remember, 20-30RPS was its maximum (while wrk+nginx chain can produce 10-12 KRPS). It's not a load generator, but it validates received messages very strictly. But the problem is that we have to develop two types of tests: deproxy-tests with one single message chain and workload tests using `wrk`. These two kind of tests are completely different.
Issues we have faced with Deproxy during pat year:
- Too slow sometimes that we have to increase some timeouts (response wasn't received in time)
- All requests and responses are processed asynchronously in one thread, while some tests require background operations that blocks (especially paramiko library). We've manually added new threads in that tests.
Issues in wrk tests:
- wrk can possibly overload VM, especially if everything is started on the same VM since it doesn't support constant throughput mode.
- need to thoroughly set asserts on wrk counters, it's very painful to find balance between false-negatives and allowed errors rate. Some errors will happen if backend configuration isn't optimal. And it isn't since we try to use backend defaults as long as it possible. This was the most painful work for me.
- we had to add response code counters and check them after tests to identify suspicious situations.
- we had to add socket error counters (connect, read, write) since it happens some time.
- context of lua scripts in wrk is thread-global, used by multiple connections, but per-connection context is more desired to write simple and reliable tests.
- Backend used in the tests is nginx, it's a real server, but something like server-side wrk is highly desired.
- requests|responses|payload may be spoilt but nothing cares about it. Only Http message integrity is checked.
As for me, we should rework the deproxy to make it possible to serve at least 5-10KRPS to test all the possible multy-threaded effects in Tempesta. The tool MUST do full Http message validation as current deproxy does. It's not very hard to do quite fast: wrk already has full (but simple) http parser inside, so does nginx, while the validation doesn't require advanced http parsers, pattern matching is works in the deproxy for a long time. Each deproxy client and server must be a separate process, that gets a script as input (like wrk), writes all the sent and received messages in it's log file, make some assertions and returns eror code (0=Test passed, not-NULL=failed). The test framework must analyse only that returns codes.
That will allow to build both functional and workload code with one piece of code, only difference - is the concurrent connections parameter.
TfwBomber was the tool intended to replace the wrk and give more abilities for fuzzing, but it doesn't care about server side. It reuses a lot of Tempesta code. It can be useful. | infrastructure | multy threaded deproxy the deproxy is fully written in python and has a poor performance as far as i remember was its maximum while wrk nginx chain can produce krps it s not a load generator but it validates received messages very strictly but the problem is that we have to develop two types of tests deproxy tests with one single message chain and workload tests using wrk these two kind of tests are completely different issues we have faced with deproxy during pat year too slow sometimes that we have to increase some timeouts response wasn t received in time all requests and responses are processed asynchronously in one thread while some tests require background operations that blocks especially paramiko library we ve manually added new threads in that tests issues in wrk tests wrk can possibly overload vm especially if everything is started on the same vm since it doesn t support constant throughput mode need to thoroughly set asserts on wrk counters it s very painful to find balance between false negatives and allowed errors rate some errors will happen if backend configuration isn t optimal and it isn t since we try to use backend defaults as long as it possible this was the most painful work for me we had to add response code counters and check them after tests to identify suspicious situations we had to add socket error counters connect read write since it happens some time context of lua scripts in wrk is thread global used by multiple connections but per connection context is more desired to write simple and reliable tests backend used in the tests is nginx it s a real server but something like server side wrk is highly desired requests responses payload may be spoilt but nothing cares about it only http message integrity is checked as for me we should rework the deproxy to make it possible to serve at least to test all the possible multy threaded effects in tempesta the tool must do full http message validation as current deproxy does it s not very hard to do quite fast wrk already has full but simple http parser inside so does nginx while the validation doesn t require advanced http parsers pattern matching is works in the deproxy for a long time each deproxy client and server must be a separate process that gets a script as input like wrk writes all the sent and received messages in it s log file make some assertions and returns eror code test passed not null failed the test framework must analyse only that returns codes that will allow to build both functional and workload code with one piece of code only difference is the concurrent connections parameter tfwbomber was the tool intended to replace the wrk and give more abilities for fuzzing but it doesn t care about server side it reuses a lot of tempesta code it can be useful | 1 |
30,684 | 24,994,251,806 | IssuesEvent | 2022-11-02 21:53:04 | FTBTeam/FTB-Modpack-Issues | https://api.github.com/repos/FTBTeam/FTB-Modpack-Issues | closed | [Bug]: kubejs error & warnings everytime I join my server | state: Fixed :heavy_check_mark: 3: Infrastructure FTB Inferno | ### Modpack
FTB Inferno
### Modpack version
1.0.0
### Log Files
[server.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9690752/server.txt)
### Describe the bug
kubejs errors every time I join my server.
### Steps to reproduce
Join server
### Expected behaviour
No errors
### Screenshots

### Additional information
Server running on ubuntu20 server VM. Installed using official ftb linux installer | 1.0 | [Bug]: kubejs error & warnings everytime I join my server - ### Modpack
FTB Inferno
### Modpack version
1.0.0
### Log Files
[server.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9690752/server.txt)
### Describe the bug
kubejs errors every time I join my server.
### Steps to reproduce
Join server
### Expected behaviour
No errors
### Screenshots

### Additional information
Server running on ubuntu20 server VM. Installed using official ftb linux installer | infrastructure | kubejs error warnings everytime i join my server modpack ftb inferno modpack version log files describe the bug kubejs errors every time i join my server steps to reproduce join server expected behaviour no errors screenshots additional information server running on server vm installed using official ftb linux installer | 1 |
442 | 2,533,283,219 | IssuesEvent | 2015-01-23 22:06:01 | shirtsgroup/InterMol | https://api.github.com/repos/shirtsgroup/InterMol | closed | Refactoring to common naming convention? | code style | I think we started out writing things with some kind of Java-like naming convention but it hasn't remained particularly consistent (partially my fault when I started working on the LAMMPS parser but the Desmond parser is now also more python-like like than the GROMACS parser).
Do we want to refactor, at least the filenames and maybe some major variable names, to a common naming convention (e.g. [PEP8](http://legacy.python.org/dev/peps/pep-0008/#naming-conventions))? I could easily do this before the first tag this weekend if this is of interest (requires a bit of tedious refactoring in something like PyCharm but probably no more than 15-30 minutes of work for the big things) | 1.0 | Refactoring to common naming convention? - I think we started out writing things with some kind of Java-like naming convention but it hasn't remained particularly consistent (partially my fault when I started working on the LAMMPS parser but the Desmond parser is now also more python-like like than the GROMACS parser).
Do we want to refactor, at least the filenames and maybe some major variable names, to a common naming convention (e.g. [PEP8](http://legacy.python.org/dev/peps/pep-0008/#naming-conventions))? I could easily do this before the first tag this weekend if this is of interest (requires a bit of tedious refactoring in something like PyCharm but probably no more than 15-30 minutes of work for the big things) | non_infrastructure | refactoring to common naming convention i think we started out writing things with some kind of java like naming convention but it hasn t remained particularly consistent partially my fault when i started working on the lammps parser but the desmond parser is now also more python like like than the gromacs parser do we want to refactor at least the filenames and maybe some major variable names to a common naming convention e g i could easily do this before the first tag this weekend if this is of interest requires a bit of tedious refactoring in something like pycharm but probably no more than minutes of work for the big things | 0 |
82,828 | 10,289,682,451 | IssuesEvent | 2019-08-27 10:26:41 | elastic/kibana | https://api.github.com/repos/elastic/kibana | reopened | [APM] Show user-agent distribution information for error exceptions | Team:apm [zube]: Done design | # Summary
It came up from designing the browser distribution breakdown for transactions (page loads) in RUM, that we could add the same chart for showing browser distribution for errors.
What might this look like? | 1.0 | [APM] Show user-agent distribution information for error exceptions - # Summary
It came up from designing the browser distribution breakdown for transactions (page loads) in RUM, that we could add the same chart for showing browser distribution for errors.
What might this look like? | non_infrastructure | show user agent distribution information for error exceptions summary it came up from designing the browser distribution breakdown for transactions page loads in rum that we could add the same chart for showing browser distribution for errors what might this look like | 0 |
79,899 | 29,507,906,785 | IssuesEvent | 2023-06-03 14:47:15 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | Diff view in 'Review changes' doesn't show difference for deleted website tag | Defect Medium | ## Vespucci Version
19.0.0
## Download source
F-Droid
## Device (Manufacturer and Model)
Google Pixel 6a
## Android Version
Android 13
## Behaviour/Symptoms
Diff view doesn't show `website` tag as changed and doesn't show the old value.

## Expected Behaviour
Diff view shows `website` tag as changed and shows the old value.
## How to recreate
Is reproducible on multiple nodes. Example https://www.openstreetmap.org/node/5915334152. Edit the tags, mark the checkbox of the website line and delete it. Open the diff view (review changes).
## Crash dump submitted (no or yes + date)
No
## Any other potentially relevant information
\-
| 1.0 | Diff view in 'Review changes' doesn't show difference for deleted website tag - ## Vespucci Version
19.0.0
## Download source
F-Droid
## Device (Manufacturer and Model)
Google Pixel 6a
## Android Version
Android 13
## Behaviour/Symptoms
Diff view doesn't show `website` tag as changed and doesn't show the old value.

## Expected Behaviour
Diff view shows `website` tag as changed and shows the old value.
## How to recreate
Is reproducible on multiple nodes. Example https://www.openstreetmap.org/node/5915334152. Edit the tags, mark the checkbox of the website line and delete it. Open the diff view (review changes).
## Crash dump submitted (no or yes + date)
No
## Any other potentially relevant information
\-
| non_infrastructure | diff view in review changes doesn t show difference for deleted website tag vespucci version download source f droid device manufacturer and model google pixel android version android behaviour symptoms diff view doesn t show website tag as changed and doesn t show the old value expected behaviour diff view shows website tag as changed and shows the old value how to recreate is reproducible on multiple nodes example edit the tags mark the checkbox of the website line and delete it open the diff view review changes crash dump submitted no or yes date no any other potentially relevant information | 0 |
338,579 | 10,231,881,293 | IssuesEvent | 2019-08-18 13:22:21 | bbottema/simple-java-mail | https://api.github.com/repos/bbottema/simple-java-mail | opened | Add config property support for trusting hosts and verifying server identity | Priority-Low enhancement | Currently a mailer instance can be configured for these using the builder api, but not through properties yet. | 1.0 | Add config property support for trusting hosts and verifying server identity - Currently a mailer instance can be configured for these using the builder api, but not through properties yet. | non_infrastructure | add config property support for trusting hosts and verifying server identity currently a mailer instance can be configured for these using the builder api but not through properties yet | 0 |
28,352 | 23,175,757,799 | IssuesEvent | 2022-07-31 11:51:31 | ffhacksterex/FFHacksterEx | https://api.github.com/repos/ffhacksterex/FFHacksterEx | closed | Assembly Project - Ensure that FF1.nes is removed regardless of compile result | infrastructure | Currently, a batch file is used internally to compile the NES image.
However, it targets a temp file in the asm/ folder, then copies that to the main directory.
- Since the batch file takes a full pathname to a file as a parameter, pass the work ROM file path to the batch file instead of 'ff1.nes'. | 1.0 | Assembly Project - Ensure that FF1.nes is removed regardless of compile result - Currently, a batch file is used internally to compile the NES image.
However, it targets a temp file in the asm/ folder, then copies that to the main directory.
- Since the batch file takes a full pathname to a file as a parameter, pass the work ROM file path to the batch file instead of 'ff1.nes'. | infrastructure | assembly project ensure that nes is removed regardless of compile result currently a batch file is used internally to compile the nes image however it targets a temp file in the asm folder then copies that to the main directory since the batch file takes a full pathname to a file as a parameter pass the work rom file path to the batch file instead of nes | 1 |
2,776 | 3,863,476,102 | IssuesEvent | 2016-04-08 09:35:01 | OpenSCAP/scap-security-guide | https://api.github.com/repos/OpenSCAP/scap-security-guide | opened | [BugFix] "The value of @schematron-version should be โ1.2โ instead โ1.0โ. (please see the readme.txt http://scap.nist.gov/revision/1.2/schematron.html -- http://scap.nist.gov/revision/1.2/scap-schematron-rules.zip): โAll of the checks have been added to source-data-stream-1.2._1.2.sch; content must update @schematron-version to 1.2 in order to use this updateโ." error | bugfix Infrastructure | Third issue reported by Dragos Prisaca of NIST while performing SSG content ScapVal validation:
```
The value of @schematron-version should be โ1.2โ instead โ1.0โ.
(please see the readme.txt http://scap.nist.gov/revision/1.2/schematron.html -- http://scap.nist.gov
/revision/1.2/scap-schematron-rules.zip): โAll of the checks have been added to source-data-stream-
1.2._1.2.sch; content must update @schematron-version to 1.2 in order to use this updateโ.
``` | 1.0 | [BugFix] "The value of @schematron-version should be โ1.2โ instead โ1.0โ. (please see the readme.txt http://scap.nist.gov/revision/1.2/schematron.html -- http://scap.nist.gov/revision/1.2/scap-schematron-rules.zip): โAll of the checks have been added to source-data-stream-1.2._1.2.sch; content must update @schematron-version to 1.2 in order to use this updateโ." error - Third issue reported by Dragos Prisaca of NIST while performing SSG content ScapVal validation:
```
The value of @schematron-version should be โ1.2โ instead โ1.0โ.
(please see the readme.txt http://scap.nist.gov/revision/1.2/schematron.html -- http://scap.nist.gov
/revision/1.2/scap-schematron-rules.zip): โAll of the checks have been added to source-data-stream-
1.2._1.2.sch; content must update @schematron-version to 1.2 in order to use this updateโ.
``` | infrastructure | the value of schematron version should be โ โ instead โ โ please see the readme txt โall of the checks have been added to source data stream sch content must update schematron version to in order to use this updateโ error third issue reported by dragos prisaca of nist while performing ssg content scapval validation the value of schematron version should be โ โ instead โ โ please see the readme txt revision scap schematron rules zip โall of the checks have been added to source data stream sch content must update schematron version to in order to use this updateโ | 1 |
24,611 | 17,467,316,194 | IssuesEvent | 2021-08-06 18:54:06 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Scan through packages for unnecessary refs | area-Infrastructure-libraries | In example System.IO.Ports:

cc @ericstj @safern @wtgodbe | 1.0 | Scan through packages for unnecessary refs - In example System.IO.Ports:

cc @ericstj @safern @wtgodbe | infrastructure | scan through packages for unnecessary refs in example system io ports cc ericstj safern wtgodbe | 1 |
18,073 | 12,749,262,733 | IssuesEvent | 2020-06-26 22:14:43 | commitdev/zero | https://api.github.com/repos/commitdev/zero | opened | Use VPC module single_nat_gateway parameter to save money in generated infra | infrastructure | Allow configuring this in `environment/<env>/main.tf` | 1.0 | Use VPC module single_nat_gateway parameter to save money in generated infra - Allow configuring this in `environment/<env>/main.tf` | infrastructure | use vpc module single nat gateway parameter to save money in generated infra allow configuring this in environment main tf | 1 |
155,170 | 13,613,583,835 | IssuesEvent | 2020-09-23 12:06:26 | liferay/clay | https://api.github.com/repos/liferay/clay | opened | Expand documentation of Layout | comp: clayui.com comp: documentation | Now that the set of Layout components has been brewing for some time in DXP we could document them a bit better.
We could describe some of the more common strategies and document the markup of the component so it's more consistent with all other components.
@bryceosterhaus You worked on this, perhaps you can do this, or if you don't have time, provide me with some bullet points of what to go over. | 1.0 | Expand documentation of Layout - Now that the set of Layout components has been brewing for some time in DXP we could document them a bit better.
We could describe some of the more common strategies and document the markup of the component so it's more consistent with all other components.
@bryceosterhaus You worked on this, perhaps you can do this, or if you don't have time, provide me with some bullet points of what to go over. | non_infrastructure | expand documentation of layout now that the set of layout components has been brewing for some time in dxp we could document them a bit better we could describe some of the more common strategies and document the markup of the component so it s more consistent with all other components bryceosterhaus you worked on this perhaps you can do this or if you don t have time provide me with some bullet points of what to go over | 0 |
49,043 | 10,317,674,780 | IssuesEvent | 2019-08-30 13:17:49 | cybercongress/cyberd | https://api.github.com/repos/cybercongress/cyberd | opened | Damping factor as onchain parameter | Code: Rank Priority: Low Status: Discussion Type: Research | The idea to allow validators with governance adjust the value of damping factor | 1.0 | Damping factor as onchain parameter - The idea to allow validators with governance adjust the value of damping factor | non_infrastructure | damping factor as onchain parameter the idea to allow validators with governance adjust the value of damping factor | 0 |
57,578 | 14,165,885,022 | IssuesEvent | 2020-11-12 08:03:07 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 77: libsndfile-1.0.28: 4 advisories | 1.severity: security 2.status: wait-for-upstream | [search](https://search.nix.gsc.io/?q=libsndfile&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libsndfile+in%3Apath&type=Code)
* [ ] [CVE-2018-19432](https://nvd.nist.gov/vuln/detail/CVE-2018-19432) (nixos-19.09)
* [ ] [CVE-2018-19661](https://nvd.nist.gov/vuln/detail/CVE-2018-19661) (nixos-19.09)
* [ ] [CVE-2018-19662](https://nvd.nist.gov/vuln/detail/CVE-2018-19662) (nixos-19.09)
* [ ] [CVE-2018-19758](https://nvd.nist.gov/vuln/detail/CVE-2018-19758) (nixos-19.09)
Scanned versions: nixos-19.09: 5f506b95f9f. May contain false positives.
| True | Vulnerability roundup 77: libsndfile-1.0.28: 4 advisories - [search](https://search.nix.gsc.io/?q=libsndfile&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libsndfile+in%3Apath&type=Code)
* [ ] [CVE-2018-19432](https://nvd.nist.gov/vuln/detail/CVE-2018-19432) (nixos-19.09)
* [ ] [CVE-2018-19661](https://nvd.nist.gov/vuln/detail/CVE-2018-19661) (nixos-19.09)
* [ ] [CVE-2018-19662](https://nvd.nist.gov/vuln/detail/CVE-2018-19662) (nixos-19.09)
* [ ] [CVE-2018-19758](https://nvd.nist.gov/vuln/detail/CVE-2018-19758) (nixos-19.09)
Scanned versions: nixos-19.09: 5f506b95f9f. May contain false positives.
| non_infrastructure | vulnerability roundup libsndfile advisories nixos nixos nixos nixos scanned versions nixos may contain false positives | 0 |
250,718 | 7,986,507,531 | IssuesEvent | 2018-07-19 02:34:13 | CarbonLDP/carbonldp-workbench | https://api.github.com/repos/CarbonLDP/carbonldp-workbench | closed | Update code to refactored SDK | priority2: required type: task | Since the SDK is being refactored to do the following:
- Change `PersistedSomething` to `Something`. e.g: `PersistedDocument` is now `Document`
- Change `Something` to `InMemorySomething`. e.g: `Document` to `InMemoryDocument`
- Change `Documents` to reflects its use to a simple document instead of using it as a service
All the changes mentioned above are reflected on the following issues:
- [#259 - Remove Persisted* convention](https://github.com/CarbonLDP/carbonldp-js-sdk/issues/259)
- [#260 - Make Documents service a Document](https://github.com/CarbonLDP/carbonldp-js-sdk/issues/260)
We need to update the actual code to be compatible with this refactored version. | 1.0 | Update code to refactored SDK - Since the SDK is being refactored to do the following:
- Change `PersistedSomething` to `Something`. e.g: `PersistedDocument` is now `Document`
- Change `Something` to `InMemorySomething`. e.g: `Document` to `InMemoryDocument`
- Change `Documents` to reflects its use to a simple document instead of using it as a service
All the changes mentioned above are reflected on the following issues:
- [#259 - Remove Persisted* convention](https://github.com/CarbonLDP/carbonldp-js-sdk/issues/259)
- [#260 - Make Documents service a Document](https://github.com/CarbonLDP/carbonldp-js-sdk/issues/260)
We need to update the actual code to be compatible with this refactored version. | non_infrastructure | update code to refactored sdk since the sdk is being refactored to do the following change persistedsomething to something e g persisteddocument is now document change something to inmemorysomething e g document to inmemorydocument change documents to reflects its use to a simple document instead of using it as a service all the changes mentioned above are reflected on the following issues we need to update the actual code to be compatible with this refactored version | 0 |
65,956 | 27,291,915,925 | IssuesEvent | 2023-02-23 17:11:43 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Certificate installed for CLAB | team/DXC ops and shared services | Install the new certificate on the cluster:
**Definition of done**
- [x] Certificate installed | 1.0 | Certificate installed for CLAB - Install the new certificate on the cluster:
**Definition of done**
- [x] Certificate installed | non_infrastructure | certificate installed for clab install the new certificate on the cluster definition of done certificate installed | 0 |
29,808 | 24,287,499,119 | IssuesEvent | 2022-09-29 00:33:14 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Cannot Step Into NuGet Source Code | Functionality:SDK Priority:2 Type:Bug Type:Engineering Category:Quality Week Infrastructure | I'm going over some simple C# .NET Core code that deals with package dependencies, using the NuGet SDK itself (code [here](https://github.com/NuGet/Home/issues/6188)). So I've created a sample project in VS 2019 and used that code.
I'd like to do a step-into with Visual Studio, to see exactly what NuGet methods are called, so need to make sure the symbols I'm using match whatever version the libraries the code uses. So I don't leave things to chance, and install the 5.6.0 version of the NuGet packages the code needs (`NuGet.Protocol` and `NuGet.DependencyResolver.Core`, with their required dependencies installed automatically) from VS's own Package Manager UI. I build the project, go to the `bin` folder and record the exact file version of the `NuGet.Protocol` .dll: 5.6.0.6591. I'm pretty sure the other assemblies included in the NuGet packages use the same version, based on the fact that there are dedicated Powershell scripts inside the NuGet's own tooling that look to rewrite the version when needed for all.
I clone the `NuGet.Client` GitHub repo locally. I switch to the `5.6.0.6591` tag, and build using the instructions in the `CONTRIBUTING.md` file. I go to the `artifacts` folder, where each project has its output assemblies and symbol files. I add the path to both folders containing the files I'm after, going with `netstandard2.0` folder due to the framework I'm targeting (.NET Core). Casually looking over one of the .dll within, I notice that the file version is 5.6.0.41820, clearly different from what I was expecting. Sure enough, adding both folders (corresponding to the 2 NuGet packages) to the symbol path in VS doesn't have any effect - as stepping into the code doesn't even trigger the symbols loading, let alone go into the NuGet source code.
Another hint that things don't quite match what I expect is that the .nupkg packages created right at the end of the build process have a names such as `NuGet.Protocol.5.6.0-zlocal.41820.nupkg`.
I'm pretty certain VS is configured to correctly process symbols, as I can step through .NET Core's own source code for methods invoked from the project mentioned above, after VS downloads automatically the required sources.
Is there any way to step into the NuGet source code in the context above ? Or is it something I've got wrong in the whole setup ? | 1.0 | Cannot Step Into NuGet Source Code - I'm going over some simple C# .NET Core code that deals with package dependencies, using the NuGet SDK itself (code [here](https://github.com/NuGet/Home/issues/6188)). So I've created a sample project in VS 2019 and used that code.
I'd like to do a step-into with Visual Studio, to see exactly what NuGet methods are called, so need to make sure the symbols I'm using match whatever version the libraries the code uses. So I don't leave things to chance, and install the 5.6.0 version of the NuGet packages the code needs (`NuGet.Protocol` and `NuGet.DependencyResolver.Core`, with their required dependencies installed automatically) from VS's own Package Manager UI. I build the project, go to the `bin` folder and record the exact file version of the `NuGet.Protocol` .dll: 5.6.0.6591. I'm pretty sure the other assemblies included in the NuGet packages use the same version, based on the fact that there are dedicated Powershell scripts inside the NuGet's own tooling that look to rewrite the version when needed for all.
I clone the `NuGet.Client` GitHub repo locally. I switch to the `5.6.0.6591` tag, and build using the instructions in the `CONTRIBUTING.md` file. I go to the `artifacts` folder, where each project has its output assemblies and symbol files. I add the path to both folders containing the files I'm after, going with `netstandard2.0` folder due to the framework I'm targeting (.NET Core). Casually looking over one of the .dll within, I notice that the file version is 5.6.0.41820, clearly different from what I was expecting. Sure enough, adding both folders (corresponding to the 2 NuGet packages) to the symbol path in VS doesn't have any effect - as stepping into the code doesn't even trigger the symbols loading, let alone go into the NuGet source code.
Another hint that things don't quite match what I expect is that the .nupkg packages created right at the end of the build process have a names such as `NuGet.Protocol.5.6.0-zlocal.41820.nupkg`.
I'm pretty certain VS is configured to correctly process symbols, as I can step through .NET Core's own source code for methods invoked from the project mentioned above, after VS downloads automatically the required sources.
Is there any way to step into the NuGet source code in the context above ? Or is it something I've got wrong in the whole setup ? | infrastructure | cannot step into nuget source code i m going over some simple c net core code that deals with package dependencies using the nuget sdk itself code so i ve created a sample project in vs and used that code i d like to do a step into with visual studio to see exactly what nuget methods are called so need to make sure the symbols i m using match whatever version the libraries the code uses so i don t leave things to chance and install the version of the nuget packages the code needs nuget protocol and nuget dependencyresolver core with their required dependencies installed automatically from vs s own package manager ui i build the project go to the bin folder and record the exact file version of the nuget protocol dll i m pretty sure the other assemblies included in the nuget packages use the same version based on the fact that there are dedicated powershell scripts inside the nuget s own tooling that look to rewrite the version when needed for all i clone the nuget client github repo locally i switch to the tag and build using the instructions in the contributing md file i go to the artifacts folder where each project has its output assemblies and symbol files i add the path to both folders containing the files i m after going with folder due to the framework i m targeting net core casually looking over one of the dll within i notice that the file version is clearly different from what i was expecting sure enough adding both folders corresponding to the nuget packages to the symbol path in vs doesn t have any effect as stepping into the code doesn t even trigger the symbols loading let alone go into the nuget source code another hint that things don t quite match what i expect is that the nupkg packages created right at the end of the build process have a names such as nuget protocol zlocal nupkg i m pretty certain vs is configured to correctly process symbols as i can step through net core s own source code for methods invoked from the project mentioned above after vs downloads automatically the required sources is there any way to step into the nuget source code in the context above or is it something i ve got wrong in the whole setup | 1 |
2,220 | 3,572,921,754 | IssuesEvent | 2016-01-27 02:15:26 | nicostouch/freeboot | https://api.github.com/repos/nicostouch/freeboot | opened | Implement Continuous Delivery Pipeline | Infrastructure | - [ ] Pre-commit hook checks successful run of test suite before commit
- [ ] GitHub commit triggers build of docker image
- [ ] Successful build triggers test suite
- [ ] Successful tests trigger push to docker hub
- [ ] Push to docker hub triggers deploy to staging
- [ ] Successful deploy to staging triggers automated browser-based functional testing
- [ ] Successful functional testing triggers deploy to production | 1.0 | Implement Continuous Delivery Pipeline - - [ ] Pre-commit hook checks successful run of test suite before commit
- [ ] GitHub commit triggers build of docker image
- [ ] Successful build triggers test suite
- [ ] Successful tests trigger push to docker hub
- [ ] Push to docker hub triggers deploy to staging
- [ ] Successful deploy to staging triggers automated browser-based functional testing
- [ ] Successful functional testing triggers deploy to production | infrastructure | implement continuous delivery pipeline pre commit hook checks successful run of test suite before commit github commit triggers build of docker image successful build triggers test suite successful tests trigger push to docker hub push to docker hub triggers deploy to staging successful deploy to staging triggers automated browser based functional testing successful functional testing triggers deploy to production | 1 |
201,966 | 15,817,848,041 | IssuesEvent | 2021-04-05 15:10:52 | ladybug-tools/honeybee-radiance-command | https://api.github.com/repos/ladybug-tools/honeybee-radiance-command | opened | Improving documentation | documentation | Hi @mostaphaRoudsari
Is there a reason to have [documentation](https://www.ladybug.tools/honeybee-radiance-command/docs/honeybee_radiance_command.options.rtrace.html#honeybee_radiance_command.options.rtrace.RtraceOptions.ab) where the default value and description are displayed in this manner? | 1.0 | Improving documentation - Hi @mostaphaRoudsari
Is there a reason to have [documentation](https://www.ladybug.tools/honeybee-radiance-command/docs/honeybee_radiance_command.options.rtrace.html#honeybee_radiance_command.options.rtrace.RtraceOptions.ab) where the default value and description are displayed in this manner? | non_infrastructure | improving documentation hi mostapharoudsari is there a reason to have where the default value and description are displayed in this manner | 0 |
58,880 | 14,499,086,360 | IssuesEvent | 2020-12-11 16:17:35 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | closed | jdknext failing to build, Java Preprocessor No configuration or non-existant configuration specified | blocker comp:build jdk17 | I suspect this is occurring because JDK16 was split and jdknext is building jdk17 now. We need to add the jdk17 preprocessor configuration, and I think a few other changes to support jdk17 build.
```
23:21:32 Building OpenJ9 Java Preprocessor
23:21:32 Building /home/jenkins/workspace/Build_JDKnext_ppc64le_linux_xl_OpenJDK/build/linux-ppc64le-server-release/support/j9tools/jpp.jar
23:21:40 Generating J9JCL sources
23:21:40 Reading preprocess instructions from xml...
23:21:40 No configuration or non-existant configuration specified (Configurations are case sensitive)
23:21:40 PREPROCESS WAS NOT SUCCESSFUL
``` | 1.0 | jdknext failing to build, Java Preprocessor No configuration or non-existant configuration specified - I suspect this is occurring because JDK16 was split and jdknext is building jdk17 now. We need to add the jdk17 preprocessor configuration, and I think a few other changes to support jdk17 build.
```
23:21:32 Building OpenJ9 Java Preprocessor
23:21:32 Building /home/jenkins/workspace/Build_JDKnext_ppc64le_linux_xl_OpenJDK/build/linux-ppc64le-server-release/support/j9tools/jpp.jar
23:21:40 Generating J9JCL sources
23:21:40 Reading preprocess instructions from xml...
23:21:40 No configuration or non-existant configuration specified (Configurations are case sensitive)
23:21:40 PREPROCESS WAS NOT SUCCESSFUL
``` | non_infrastructure | jdknext failing to build java preprocessor no configuration or non existant configuration specified i suspect this is occurring because was split and jdknext is building now we need to add the preprocessor configuration and i think a few other changes to support build building java preprocessor building home jenkins workspace build jdknext linux xl openjdk build linux server release support jpp jar generating sources reading preprocess instructions from xml no configuration or non existant configuration specified configurations are case sensitive preprocess was not successful | 0 |
736,055 | 25,455,950,347 | IssuesEvent | 2022-11-24 14:12:25 | CDLUC3/ezid | https://api.github.com/repos/CDLUC3/ezid | closed | `Internal Server Error: /account/pwreset` raised when password reset attempted | bug priority-urgent | ```
Internal Server Error: /account/pwreset
AttributeError at /account/pwreset
'str' object has no attribute 'primaryContactName'
Request Method: POST
Request URL: http://ezid.cdlib.org/account/pwreset
Django Version: 3.2.7
Python Executable: /ezid/.pyenv/versions/ezid-py3/bin/python
Python Version: 3.8.5
...
Traceback (most recent call last):
File "/ezid/.pyenv/versions/ezid-py3/lib/python3.8/site-packages/Django-3.2.7-py3.8.egg/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/ezid/.pyenv/versions/ezid-py3/lib/python3.8/site-packages/Django-3.2.7-py3.8.egg/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/ezid/ezid/impl/ui_account.py", line 396, in pwreset
r = sendPasswordResetEmail(request, username, email)
File "/ezid/ezid/impl/ui_account.py", line 442, in sendPasswordResetEmail
_sendEmail(request, username, _("EZID password reset request"), message, emailAddress)
File "/ezid/ezid/impl/ui_account.py", line 249, in _sendEmail
to_full_address = user.primaryContactName + "<" + to_address + ">"
Exception Type: AttributeError at /account/pwreset
Exception Value: 'str' object has no attribute 'primaryContactName'
Request information:
USER: AnonymousUser
GET: No GET data
POST:
username = 'ucnrs_kwong'
email = 'kathleen.wong@ucop.edu'
FILES: No FILES data
COOKIES: No cookie data
``` | 1.0 | `Internal Server Error: /account/pwreset` raised when password reset attempted - ```
Internal Server Error: /account/pwreset
AttributeError at /account/pwreset
'str' object has no attribute 'primaryContactName'
Request Method: POST
Request URL: http://ezid.cdlib.org/account/pwreset
Django Version: 3.2.7
Python Executable: /ezid/.pyenv/versions/ezid-py3/bin/python
Python Version: 3.8.5
...
Traceback (most recent call last):
File "/ezid/.pyenv/versions/ezid-py3/lib/python3.8/site-packages/Django-3.2.7-py3.8.egg/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/ezid/.pyenv/versions/ezid-py3/lib/python3.8/site-packages/Django-3.2.7-py3.8.egg/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/ezid/ezid/impl/ui_account.py", line 396, in pwreset
r = sendPasswordResetEmail(request, username, email)
File "/ezid/ezid/impl/ui_account.py", line 442, in sendPasswordResetEmail
_sendEmail(request, username, _("EZID password reset request"), message, emailAddress)
File "/ezid/ezid/impl/ui_account.py", line 249, in _sendEmail
to_full_address = user.primaryContactName + "<" + to_address + ">"
Exception Type: AttributeError at /account/pwreset
Exception Value: 'str' object has no attribute 'primaryContactName'
Request information:
USER: AnonymousUser
GET: No GET data
POST:
username = 'ucnrs_kwong'
email = 'kathleen.wong@ucop.edu'
FILES: No FILES data
COOKIES: No cookie data
``` | non_infrastructure | internal server error account pwreset raised when password reset attempted internal server error account pwreset attributeerror at account pwreset str object has no attribute primarycontactname request method post request url django version python executable ezid pyenv versions ezid bin python python version traceback most recent call last file ezid pyenv versions ezid lib site packages django egg django core handlers exception py line in inner response get response request file ezid pyenv versions ezid lib site packages django egg django core handlers base py line in get response response wrapped callback request callback args callback kwargs file ezid ezid impl ui account py line in pwreset r sendpasswordresetemail request username email file ezid ezid impl ui account py line in sendpasswordresetemail sendemail request username ezid password reset request message emailaddress file ezid ezid impl ui account py line in sendemail to full address user primarycontactname exception type attributeerror at account pwreset exception value str object has no attribute primarycontactname request information user anonymoususer get no get data post username ucnrs kwong email kathleen wong ucop edu files no files data cookies no cookie data | 0 |
20,838 | 16,092,648,643 | IssuesEvent | 2021-04-26 18:45:04 | gardener/gardener | https://api.github.com/repos/gardener/gardener | closed | Relax server-side validation for templates in ManagedSeedSet | area/ops-productivity area/usability kind/enhancement | **How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|operations|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
"/priority" identifiers: 1|2|3|4|5 (ordered from greatest to least)
-->
/area usability ops-productivity
/kind enhancement
**What would you like to be added**:
In addition to relaxing client-side validation (https://github.com/gardener/gardener/issues/3909) we could also relax the server side validation of the `shootTemplate` in `ManagedSeed`s in the following ways:
- allow specifying purpose `infrastructure` in the shoot spec:
```
$ kaf local/managedseed/57-managedseedset-seedtemplate.yaml --validate=false
The ManagedSeedSet "gcp-europe-west1" is invalid:
* spec.shootTemplate.spec.purpose: Unsupported value: "infrastructure": supported values: "development", "evaluation", "production", "testing"
```
- allow leaving out the machine image version and let it be defaulted by the `ShootValidator` admission plugin to the latest version (see [here](https://github.com/gardener/gardener/blob/95b86ca115e46bb215d5a0793fccf69c3ff27398/plugin/pkg/shoot/validator/admission.go#L744)), when the controller creates the shoot
```
$ kaf local/managedseed/57-managedseedset-seedtemplate.yaml --validate=false
The ManagedSeedSet "gcp-europe-west1" is invalid:
* spec.shootTemplate.spec.provider.workers[0].machine.image.version: Required value: must specify a machine image version
```
**Why is this needed**:
Make it easier / more flexible to consume the `ManagedSeedSet` feature.
/cc @stoyanr | True | Relax server-side validation for templates in ManagedSeedSet - **How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|operations|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
"/priority" identifiers: 1|2|3|4|5 (ordered from greatest to least)
-->
/area usability ops-productivity
/kind enhancement
**What would you like to be added**:
In addition to relaxing client-side validation (https://github.com/gardener/gardener/issues/3909) we could also relax the server side validation of the `shootTemplate` in `ManagedSeed`s in the following ways:
- allow specifying purpose `infrastructure` in the shoot spec:
```
$ kaf local/managedseed/57-managedseedset-seedtemplate.yaml --validate=false
The ManagedSeedSet "gcp-europe-west1" is invalid:
* spec.shootTemplate.spec.purpose: Unsupported value: "infrastructure": supported values: "development", "evaluation", "production", "testing"
```
- allow leaving out the machine image version and let it be defaulted by the `ShootValidator` admission plugin to the latest version (see [here](https://github.com/gardener/gardener/blob/95b86ca115e46bb215d5a0793fccf69c3ff27398/plugin/pkg/shoot/validator/admission.go#L744)), when the controller creates the shoot
```
$ kaf local/managedseed/57-managedseedset-seedtemplate.yaml --validate=false
The ManagedSeedSet "gcp-europe-west1" is invalid:
* spec.shootTemplate.spec.provider.workers[0].machine.image.version: Required value: must specify a machine image version
```
**Why is this needed**:
Make it easier / more flexible to consume the `ManagedSeedSet` feature.
/cc @stoyanr | non_infrastructure | relax server side validation for templates in managedseedset how to categorize this issue please select area kind and priority for this issue this helps the community categorizing it replace below todos or exchange the existing identifiers with those that fit best in your opinion if multiple identifiers make sense you can also state the commands multiple times e g area control plane area auto scaling area identifiers audit logging auto scaling backup certification control plane migration control plane cost delivery dev productivity disaster recovery documentation high availability logging metering monitoring networking open source operations ops productivity os performance quality robustness scalability security storage testing usability user management kind identifiers api change bug cleanup discussion enhancement epic impediment poc post mortem question regression task technical debt test priority identifiers ordered from greatest to least area usability ops productivity kind enhancement what would you like to be added in addition to relaxing client side validation we could also relax the server side validation of the shoottemplate in managedseed s in the following ways allow specifying purpose infrastructure in the shoot spec kaf local managedseed managedseedset seedtemplate yaml validate false the managedseedset gcp europe is invalid spec shoottemplate spec purpose unsupported value infrastructure supported values development evaluation production testing allow leaving out the machine image version and let it be defaulted by the shootvalidator admission plugin to the latest version see when the controller creates the shoot kaf local managedseed managedseedset seedtemplate yaml validate false the managedseedset gcp europe is invalid spec shoottemplate spec provider workers machine image version required value must specify a machine image version why is this needed make it easier more flexible to consume the managedseedset feature cc stoyanr | 0 |
12,072 | 9,574,708,294 | IssuesEvent | 2019-05-07 03:01:42 | ericmjl/pyjanitor | https://api.github.com/repos/ericmjl/pyjanitor | closed | Azure Pipeline for notebook | docfix good intermediate issue infrastructure | Make an automated process to run the notebooks (examples/notebooks) | 1.0 | Azure Pipeline for notebook - Make an automated process to run the notebooks (examples/notebooks) | infrastructure | azure pipeline for notebook make an automated process to run the notebooks examples notebooks | 1 |
34,778 | 14,518,664,270 | IssuesEvent | 2020-12-14 00:31:00 | badges/shields | https://api.github.com/repos/badges/shields | closed | Removing default style override in GitHub Sponsors badge | service-badge | ## Overview
The purpose of this issue is to inform the community that we will be removing the default overridden style in the GitHub Sponsors badge. Based on information from our observability solution and searches on GitHub we do not believe this will impact any existing badges, but are posting this for transparency and awareness.
### Action Steps
If you are currently using the GitHub Sponsors badge **_without_** the `style=social` query parameter in your badge url and would like to continue utilizing the `social` style for your GitHub Sponsors badge, then you will need to add the `?style=social` query parameter to your badge url.
E.g. `https://img.shields.io/github/sponsors/homebrew` --> `https://img.shields.io/github/sponsors/homebrew?style=social`
If you already have the `style` query parameter then there is no action needed.
## Background
Shields.io supports several different style options for our badges which can be controlled via the `?style=...` query parameter on the badge urls. More information can be found under the styles section on the shields.io site - https://shields.io/#styles
The default stye is `flat`, so a badge requested with a url that does not explicitly specify the desired style with the `style` param will default to the flat style.
For example:
The default flat style can be observed with the npm version badge.
https://img.shields.io/npm/v/badge-maker

but can be changed to a different style with an explicit query param
https://img.shields.io/npm/v/badge-maker?style=social

However, there are a couple of badges that have _overridden_ that default style for that specific badge to use a default other than `flat`.
This can be seen with the Twitter badges where the style utilized by default is `social` instead of `flat`
https://img.shields.io/twitter/follow/shields

but the style can of course still be set/overridden explicitly
https://img.shields.io/twitter/follow/shields?style=flat

We typically avoid badge-specific overrides of the default style, but there are a few such instances within our [social category of badges](https://shields.io/category/social), including the GitHub Sponsors badge, where this is done.
https://img.shields.io/github/sponsors/homebrew

## Motivation
The GitHub Sponsors badge was added relatively recently, and was added to our offering in the `social` category with the default style overridden to be `social` as well. This was done based on the initial discussion and request for the badge to look similar to the native sponsor button within the GitHub interface, and a general desire for our social style with the new badge.
However, there were some subsequent discussions and questions about the category (funding vs. social) and whether we truly want to override the default style for this badge. While there's still some ongoing discussion about the category, we have determined that we need to remove the default style override for the GitHub Sponsors badge. | 1.0 | Removing default style override in GitHub Sponsors badge - ## Overview
The purpose of this issue is to inform the community that we will be removing the default overridden style in the GitHub Sponsors badge. Based on information from our observability solution and searches on GitHub we do not believe this will impact any existing badges, but are posting this for transparency and awareness.
### Action Steps
If you are currently using the GitHub Sponsors badge **_without_** the `style=social` query parameter in your badge url and would like to continue utilizing the `social` style for your GitHub Sponsors badge, then you will need to add the `?style=social` query parameter to your badge url.
E.g. `https://img.shields.io/github/sponsors/homebrew` --> `https://img.shields.io/github/sponsors/homebrew?style=social`
If you already have the `style` query parameter then there is no action needed.
## Background
Shields.io supports several different style options for our badges which can be controlled via the `?style=...` query parameter on the badge urls. More information can be found under the styles section on the shields.io site - https://shields.io/#styles
The default stye is `flat`, so a badge requested with a url that does not explicitly specify the desired style with the `style` param will default to the flat style.
For example:
The default flat style can be observed with the npm version badge.
https://img.shields.io/npm/v/badge-maker

but can be changed to a different style with an explicit query param
https://img.shields.io/npm/v/badge-maker?style=social

However, there are a couple of badges that have _overridden_ that default style for that specific badge to use a default other than `flat`.
This can be seen with the Twitter badges where the style utilized by default is `social` instead of `flat`
https://img.shields.io/twitter/follow/shields

but the style can of course still be set/overridden explicitly
https://img.shields.io/twitter/follow/shields?style=flat

We typically avoid badge-specific overrides of the default style, but there are a few such instances within our [social category of badges](https://shields.io/category/social), including the GitHub Sponsors badge, where this is done.
https://img.shields.io/github/sponsors/homebrew

## Motivation
The GitHub Sponsors badge was added relatively recently, and was added to our offering in the `social` category with the default style overridden to be `social` as well. This was done based on the initial discussion and request for the badge to look similar to the native sponsor button within the GitHub interface, and a general desire for our social style with the new badge.
However, there were some subsequent discussions and questions about the category (funding vs. social) and whether we truly want to override the default style for this badge. While there's still some ongoing discussion about the category, we have determined that we need to remove the default style override for the GitHub Sponsors badge. | non_infrastructure | removing default style override in github sponsors badge overview the purpose of this issue is to inform the community that we will be removing the default overridden style in the github sponsors badge based on information from our observability solution and searches on github we do not believe this will impact any existing badges but are posting this for transparency and awareness action steps if you are currently using the github sponsors badge without the style social query parameter in your badge url and would like to continue utilizing the social style for your github sponsors badge then you will need to add the style social query parameter to your badge url e g if you already have the style query parameter then there is no action needed background shields io supports several different style options for our badges which can be controlled via the style query parameter on the badge urls more information can be found under the styles section on the shields io site the default stye is flat so a badge requested with a url that does not explicitly specify the desired style with the style param will default to the flat style for example the default flat style can be observed with the npm version badge but can be changed to a different style with an explicit query param however there are a couple of badges that have overridden that default style for that specific badge to use a default other than flat this can be seen with the twitter badges where the style utilized by default is social instead of flat but the style can of course still be set overridden explicitly we typically avoid badge specific overrides of the default style but there are a few such instances within our including the github sponsors badge where this is done motivation the github sponsors badge was added relatively recently and was added to our offering in the social category with the default style overridden to be social as well this was done based on the initial discussion and request for the badge to look similar to the native sponsor button within the github interface and a general desire for our social style with the new badge however there were some subsequent discussions and questions about the category funding vs social and whether we truly want to override the default style for this badge while there s still some ongoing discussion about the category we have determined that we need to remove the default style override for the github sponsors badge | 0 |
23,512 | 16,361,141,206 | IssuesEvent | 2021-05-14 09:39:07 | astrolabsoftware/fink-broker | https://api.github.com/repos/astrolabsoftware/fink-broker | closed | Bump Kafka version | apache kafka infrastructure | **Describe the issue**
Kafka 2.6.1 is no more available to direct download -- let's take the latest (2.8.0). | 1.0 | Bump Kafka version - **Describe the issue**
Kafka 2.6.1 is no more available to direct download -- let's take the latest (2.8.0). | infrastructure | bump kafka version describe the issue kafka is no more available to direct download let s take the latest | 1 |
26,708 | 6,795,339,546 | IssuesEvent | 2017-11-01 15:24:24 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | RyuJIT: Throw block optimization does not hit on methods that only throw in their body. | area-CodeGen enhancement optimization tenet-performance | I noticed that blocks that have a throw statement are moved outside of the critical path. However, after further inspection it seems like the rearranging of the blocks is done without further analysis outside of the call site (which makes sense btw).
The problem is that the optimization is defeated by a pretty well known pattern used along the way in lots of software (even CoreFX).
``` csharp
private static void ThrowException( string message )
{
throw new Exception(message);
}
[MethodImpl(MethodImplOptions.NoInlining)]
private static void DoesItReorder(int integer, string value)
{
if (integer < 0)
ThrowException("message");
Console.WriteLine("This is the body.");
for (int i = 0; i < integer; i++)
{
if (i > 100)
ThrowException("message");
Console.WriteLine($"Line {i}");
}
}
```
I tend to believe this is by-design at the moment and I don't know if there is a solution for it without whole-program optimization, but wanted to raise it (even if it works for "documentation" of an edge case).
| 1.0 | RyuJIT: Throw block optimization does not hit on methods that only throw in their body. - I noticed that blocks that have a throw statement are moved outside of the critical path. However, after further inspection it seems like the rearranging of the blocks is done without further analysis outside of the call site (which makes sense btw).
The problem is that the optimization is defeated by a pretty well known pattern used along the way in lots of software (even CoreFX).
``` csharp
private static void ThrowException( string message )
{
throw new Exception(message);
}
[MethodImpl(MethodImplOptions.NoInlining)]
private static void DoesItReorder(int integer, string value)
{
if (integer < 0)
ThrowException("message");
Console.WriteLine("This is the body.");
for (int i = 0; i < integer; i++)
{
if (i > 100)
ThrowException("message");
Console.WriteLine($"Line {i}");
}
}
```
I tend to believe this is by-design at the moment and I don't know if there is a solution for it without whole-program optimization, but wanted to raise it (even if it works for "documentation" of an edge case).
| non_infrastructure | ryujit throw block optimization does not hit on methods that only throw in their body i noticed that blocks that have a throw statement are moved outside of the critical path however after further inspection it seems like the rearranging of the blocks is done without further analysis outside of the call site which makes sense btw the problem is that the optimization is defeated by a pretty well known pattern used along the way in lots of software even corefx csharp private static void throwexception string message throw new exception message private static void doesitreorder int integer string value if integer throwexception message console writeline this is the body for int i i integer i if i throwexception message console writeline line i i tend to believe this is by design at the moment and i don t know if there is a solution for it without whole program optimization but wanted to raise it even if it works for documentation of an edge case | 0 |
12,745 | 9,936,060,979 | IssuesEvent | 2019-07-02 18:08:22 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Change pub bots to recipe bots | area-infrastructure closed-obsolete | I'm not sure what this means exactly but @whesse asked me to file an issue for it. | 1.0 | Change pub bots to recipe bots - I'm not sure what this means exactly but @whesse asked me to file an issue for it. | infrastructure | change pub bots to recipe bots i m not sure what this means exactly but whesse asked me to file an issue for it | 1 |
6,248 | 6,273,942,450 | IssuesEvent | 2017-07-18 00:01:32 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Roslyn build fails if a parent directory name contains % characters | Area-Infrastructure Pedantic ;) | **Version Used**:
branch `master`, commit `a4fa73044fe3ed967e5c7cfdfc17bbc167da6fbb`
**Steps to Reproduce**:
1. Create directory `C:\%github%` and open a command prompt in it.
2. Run `git clone https://github.com/dotnet/roslyn.git` and then change the current directory to `C:\%github%\roslyn`.
3. Run `Restore.cmd`:
```
C:\%github%\roslyn>Restore.cmd
Repo Dir C:\%github%\roslyn
Binaries Dir C:\%github%\roslyn\Binaries
Downloading NuGet.exe
Running restore
Using existing NuGet.exe at version 4.1.0
Restore using MSBuild at C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\MSBuild\15.0\Bin
Restoring Base Toolset
Restoring Roslyn
Restoring Samples
Restoring Templates
Restoring DevDivInsertionFiles
```
4. Run `Build.cmd`:
```
C:\%github%\roslyn>Build.cmd
Repo Dir C:\%github%\roslyn
Binaries Dir C:\%github%\roslyn\Binaries
C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\MSBuild\15.0\Bin\msbuild.exe /nologo /consoleloggerparameters:Verbosity=minimal;summary /m /nodeReuse:false /p:TreatWarningsAsErrors=true /p:Configuration=Debug /p:DeployExtension=false /v:diag /warnaserror Roslyn.sln
RunTests -> C:\%github%\roslyn\Binaries\Debug\Exes\RunTests\RunTests.exe
CSharpErrorFactsGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CSharpErrorFactsGenerator\CSharpErrorFactsGenerator.dll
CSharpSyntaxGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CSharpSyntaxGenerator\CSharpSyntaxGenerator.dll
CompilersBoundTreeGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CompilersBoundTreeGenerator\BoundTreeGenerator.dll
VisualBasicErrorFactsGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\VisualBasicErrorFactsGenerator\VBErrorFactsGenerator.dll
CodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\CodeAnalysis\Microsoft.CodeAnalysis.dll
VisualBasicSyntaxGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\VisualBasicSyntaxGenerator\VBSyntaxGenerator.dll
CompilerTestResources -> C:\%github%\roslyn\Binaries\Debug\Dlls\CompilerTestResources\Roslyn.Compilers.Test.Resources.dll
Workspaces -> C:\%github%\roslyn\Binaries\Debug\Dlls\Workspaces\Microsoft.CodeAnalysis.Workspaces.dll
CSharpCodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\CSharpCodeAnalysis\Microsoft.CodeAnalysis.CSharp.dll
BasicCodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\BasicCodeAnalysis\Microsoft.CodeAnalysis.VisualBasic.dll
ResultProvider.Portable -> C:\%github%\roslyn\Binaries\Debug\Dlls\ResultProvider.Portable\Microsoft.CodeAnalysis.ExpressionEvaluator.ResultProvider.dll
ResultProvider.NetFX20 -> C:\%github%\roslyn\Binaries\Debug\Dlls\ResultProvider.NetFX20\Microsoft.CodeAnalysis.ExpressionEvaluator.ResultProvider.dll
Scripting -> C:\%github%\roslyn\Binaries\Debug\Dlls\Scripting\Microsoft.CodeAnalysis.Scripting.dll
ExpressionCompiler -> C:\%github%\roslyn\Binaries\Debug\Dlls\ExpressionCompiler\Microsoft.CodeAnalysis.ExpressionEvaluator.ExpressionCompiler.dll
EXEC : error : Input file 'C:\\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml' does not exist. [C:\%github%\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProv
ider\Portable\BasicResultProvider.Portable.vbproj]
C:\%github%\roslyn\build\Targets\Vsdconfig.targets(14,5): error MSB3073: The command ""C:\Users\reshetni/.nuget/packages/\Microsoft.VSSDK.BuildTools\15.1.192\tools\VSSDK\bin\vsdconfigtool.exe" "C:\%github%\roslyn\src\Expressi
onEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml" "C:\%github%\roslyn\Binaries\Obj\BasicResultProvider.Portable\Debug\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.
dll" "C:\%github%\roslyn\Binaries\Debug\Dlls\BasicResultProvider.Portable\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.vsdconfig"" exited with code -1. [C:\%github%\roslyn\src\ExpressionEvaluator\Vis
ualBasic\Source\ResultProvider\Portable\BasicResultProvider.Portable.vbproj]
EXEC : error : Input file 'C:\\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml' does not exist. [C:\%github%\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProv
ider\NetFX20\BasicResultProvider.NetFX20.vbproj]
C:\%github%\roslyn\build\Targets\Vsdconfig.targets(14,5): error MSB3073: The command ""C:\Users\reshetni/.nuget/packages/\Microsoft.VSSDK.BuildTools\15.1.192\tools\VSSDK\bin\vsdconfigtool.exe" "C:\%github%\roslyn\src\Expressi
onEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml" "C:\%github%\roslyn\Binaries\Obj\BasicResultProvider.NetFX20\Debug\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.d
ll" "C:\%github%\roslyn\Binaries\Debug\Dlls\BasicResultProvider.NetFX20\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.vsdconfig"" exited with code -1. [C:\%github%\roslyn\src\ExpressionEvaluator\Visua
lBasic\Source\ResultProvider\NetFX20\BasicResultProvider.NetFX20.vbproj]
```
... and many more similar errors after that.
See the full build log
[Roslyn.buildlog.7z.zip](https://github.com/dotnet/roslyn/files/1147028/Roslyn.buildlog.7z.zip), it can be opened with https://github.com/KirillOsenkov/MSBuildStructuredLog.
| 1.0 | Roslyn build fails if a parent directory name contains % characters - **Version Used**:
branch `master`, commit `a4fa73044fe3ed967e5c7cfdfc17bbc167da6fbb`
**Steps to Reproduce**:
1. Create directory `C:\%github%` and open a command prompt in it.
2. Run `git clone https://github.com/dotnet/roslyn.git` and then change the current directory to `C:\%github%\roslyn`.
3. Run `Restore.cmd`:
```
C:\%github%\roslyn>Restore.cmd
Repo Dir C:\%github%\roslyn
Binaries Dir C:\%github%\roslyn\Binaries
Downloading NuGet.exe
Running restore
Using existing NuGet.exe at version 4.1.0
Restore using MSBuild at C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\MSBuild\15.0\Bin
Restoring Base Toolset
Restoring Roslyn
Restoring Samples
Restoring Templates
Restoring DevDivInsertionFiles
```
4. Run `Build.cmd`:
```
C:\%github%\roslyn>Build.cmd
Repo Dir C:\%github%\roslyn
Binaries Dir C:\%github%\roslyn\Binaries
C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\MSBuild\15.0\Bin\msbuild.exe /nologo /consoleloggerparameters:Verbosity=minimal;summary /m /nodeReuse:false /p:TreatWarningsAsErrors=true /p:Configuration=Debug /p:DeployExtension=false /v:diag /warnaserror Roslyn.sln
RunTests -> C:\%github%\roslyn\Binaries\Debug\Exes\RunTests\RunTests.exe
CSharpErrorFactsGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CSharpErrorFactsGenerator\CSharpErrorFactsGenerator.dll
CSharpSyntaxGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CSharpSyntaxGenerator\CSharpSyntaxGenerator.dll
CompilersBoundTreeGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\CompilersBoundTreeGenerator\BoundTreeGenerator.dll
VisualBasicErrorFactsGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\VisualBasicErrorFactsGenerator\VBErrorFactsGenerator.dll
CodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\CodeAnalysis\Microsoft.CodeAnalysis.dll
VisualBasicSyntaxGenerator -> C:\%github%\roslyn\Binaries\Debug\Exes\VisualBasicSyntaxGenerator\VBSyntaxGenerator.dll
CompilerTestResources -> C:\%github%\roslyn\Binaries\Debug\Dlls\CompilerTestResources\Roslyn.Compilers.Test.Resources.dll
Workspaces -> C:\%github%\roslyn\Binaries\Debug\Dlls\Workspaces\Microsoft.CodeAnalysis.Workspaces.dll
CSharpCodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\CSharpCodeAnalysis\Microsoft.CodeAnalysis.CSharp.dll
BasicCodeAnalysis -> C:\%github%\roslyn\Binaries\Debug\Dlls\BasicCodeAnalysis\Microsoft.CodeAnalysis.VisualBasic.dll
ResultProvider.Portable -> C:\%github%\roslyn\Binaries\Debug\Dlls\ResultProvider.Portable\Microsoft.CodeAnalysis.ExpressionEvaluator.ResultProvider.dll
ResultProvider.NetFX20 -> C:\%github%\roslyn\Binaries\Debug\Dlls\ResultProvider.NetFX20\Microsoft.CodeAnalysis.ExpressionEvaluator.ResultProvider.dll
Scripting -> C:\%github%\roslyn\Binaries\Debug\Dlls\Scripting\Microsoft.CodeAnalysis.Scripting.dll
ExpressionCompiler -> C:\%github%\roslyn\Binaries\Debug\Dlls\ExpressionCompiler\Microsoft.CodeAnalysis.ExpressionEvaluator.ExpressionCompiler.dll
EXEC : error : Input file 'C:\\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml' does not exist. [C:\%github%\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProv
ider\Portable\BasicResultProvider.Portable.vbproj]
C:\%github%\roslyn\build\Targets\Vsdconfig.targets(14,5): error MSB3073: The command ""C:\Users\reshetni/.nuget/packages/\Microsoft.VSSDK.BuildTools\15.1.192\tools\VSSDK\bin\vsdconfigtool.exe" "C:\%github%\roslyn\src\Expressi
onEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml" "C:\%github%\roslyn\Binaries\Obj\BasicResultProvider.Portable\Debug\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.
dll" "C:\%github%\roslyn\Binaries\Debug\Dlls\BasicResultProvider.Portable\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.vsdconfig"" exited with code -1. [C:\%github%\roslyn\src\ExpressionEvaluator\Vis
ualBasic\Source\ResultProvider\Portable\BasicResultProvider.Portable.vbproj]
EXEC : error : Input file 'C:\\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml' does not exist. [C:\%github%\roslyn\src\ExpressionEvaluator\VisualBasic\Source\ResultProv
ider\NetFX20\BasicResultProvider.NetFX20.vbproj]
C:\%github%\roslyn\build\Targets\Vsdconfig.targets(14,5): error MSB3073: The command ""C:\Users\reshetni/.nuget/packages/\Microsoft.VSSDK.BuildTools\15.1.192\tools\VSSDK\bin\vsdconfigtool.exe" "C:\%github%\roslyn\src\Expressi
onEvaluator\VisualBasic\Source\ResultProvider\VisualBasicResultProvider.vsdconfigxml" "C:\%github%\roslyn\Binaries\Obj\BasicResultProvider.NetFX20\Debug\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.d
ll" "C:\%github%\roslyn\Binaries\Debug\Dlls\BasicResultProvider.NetFX20\\Microsoft.CodeAnalysis.VisualBasic.ExpressionEvaluator.ResultProvider.vsdconfig"" exited with code -1. [C:\%github%\roslyn\src\ExpressionEvaluator\Visua
lBasic\Source\ResultProvider\NetFX20\BasicResultProvider.NetFX20.vbproj]
```
... and many more similar errors after that.
See the full build log
[Roslyn.buildlog.7z.zip](https://github.com/dotnet/roslyn/files/1147028/Roslyn.buildlog.7z.zip), it can be opened with https://github.com/KirillOsenkov/MSBuildStructuredLog.
| infrastructure | roslyn build fails if a parent directory name contains characters version used branch master commit steps to reproduce create directory c github and open a command prompt in it run git clone and then change the current directory to c github roslyn run restore cmd c github roslyn restore cmd repo dir c github roslyn binaries dir c github roslyn binaries downloading nuget exe running restore using existing nuget exe at version restore using msbuild at c program files microsoft visual studio community msbuild bin restoring base toolset restoring roslyn restoring samples restoring templates restoring devdivinsertionfiles run build cmd c github roslyn build cmd repo dir c github roslyn binaries dir c github roslyn binaries c program files microsoft visual studio community msbuild bin msbuild exe nologo consoleloggerparameters verbosity minimal summary m nodereuse false p treatwarningsaserrors true p configuration debug p deployextension false v diag warnaserror roslyn sln runtests c github roslyn binaries debug exes runtests runtests exe csharperrorfactsgenerator c github roslyn binaries debug exes csharperrorfactsgenerator csharperrorfactsgenerator dll csharpsyntaxgenerator c github roslyn binaries debug exes csharpsyntaxgenerator csharpsyntaxgenerator dll compilersboundtreegenerator c github roslyn binaries debug exes compilersboundtreegenerator boundtreegenerator dll visualbasicerrorfactsgenerator c github roslyn binaries debug exes visualbasicerrorfactsgenerator vberrorfactsgenerator dll codeanalysis c github roslyn binaries debug dlls codeanalysis microsoft codeanalysis dll visualbasicsyntaxgenerator c github roslyn binaries debug exes visualbasicsyntaxgenerator vbsyntaxgenerator dll compilertestresources c github roslyn binaries debug dlls compilertestresources roslyn compilers test resources dll workspaces c github roslyn binaries debug dlls workspaces microsoft codeanalysis workspaces dll csharpcodeanalysis c github roslyn binaries debug dlls csharpcodeanalysis microsoft codeanalysis csharp dll basiccodeanalysis c github roslyn binaries debug dlls basiccodeanalysis microsoft codeanalysis visualbasic dll resultprovider portable c github roslyn binaries debug dlls resultprovider portable microsoft codeanalysis expressionevaluator resultprovider dll resultprovider c github roslyn binaries debug dlls resultprovider microsoft codeanalysis expressionevaluator resultprovider dll scripting c github roslyn binaries debug dlls scripting microsoft codeanalysis scripting dll expressioncompiler c github roslyn binaries debug dlls expressioncompiler microsoft codeanalysis expressionevaluator expressioncompiler dll exec error input file c roslyn src expressionevaluator visualbasic source resultprovider visualbasicresultprovider vsdconfigxml does not exist c github roslyn src expressionevaluator visualbasic source resultprov ider portable basicresultprovider portable vbproj c github roslyn build targets vsdconfig targets error the command c users reshetni nuget packages microsoft vssdk buildtools tools vssdk bin vsdconfigtool exe c github roslyn src expressi onevaluator visualbasic source resultprovider visualbasicresultprovider vsdconfigxml c github roslyn binaries obj basicresultprovider portable debug microsoft codeanalysis visualbasic expressionevaluator resultprovider dll c github roslyn binaries debug dlls basicresultprovider portable microsoft codeanalysis visualbasic expressionevaluator resultprovider vsdconfig exited with code c github roslyn src expressionevaluator vis ualbasic source resultprovider portable basicresultprovider portable vbproj exec error input file c roslyn src expressionevaluator visualbasic source resultprovider visualbasicresultprovider vsdconfigxml does not exist c github roslyn src expressionevaluator visualbasic source resultprov ider basicresultprovider vbproj c github roslyn build targets vsdconfig targets error the command c users reshetni nuget packages microsoft vssdk buildtools tools vssdk bin vsdconfigtool exe c github roslyn src expressi onevaluator visualbasic source resultprovider visualbasicresultprovider vsdconfigxml c github roslyn binaries obj basicresultprovider debug microsoft codeanalysis visualbasic expressionevaluator resultprovider d ll c github roslyn binaries debug dlls basicresultprovider microsoft codeanalysis visualbasic expressionevaluator resultprovider vsdconfig exited with code c github roslyn src expressionevaluator visua lbasic source resultprovider basicresultprovider vbproj and many more similar errors after that see the full build log it can be opened with | 1 |
254,703 | 27,413,631,896 | IssuesEvent | 2023-03-01 12:18:50 | scm-automation-project/simple-maven-project-with-non-default-java-version | https://api.github.com/repos/scm-automation-project/simple-maven-project-with-non-default-java-version | closed | log4j-1.2.7.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.7/log4j-1.2.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/simple-maven-project-with-non-default-java-version/commit/ce1656cf6bc73035593c3c0a582f4f0943d6f771">ce1656cf6bc73035593c3c0a582f4f0943d6f771</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (log4j version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-17571](https://www.mend.io/vulnerability-database/CVE-2019-17571) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | log4j-1.2.7.jar | Direct | log4j:log4j - 1.2.13-NODEP,1.2.17-atlassian-1,1.2.16.redhat-5;org.slf4j:nlog4j - 1.2.18 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-17571</summary>
### Vulnerable Library - <b>log4j-1.2.7.jar</b></p>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.7/log4j-1.2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/simple-maven-project-with-non-default-java-version/commit/ce1656cf6bc73035593c3c0a582f4f0943d6f771">ce1656cf6bc73035593c3c0a582f4f0943d6f771</a></p>
</p>
<p></p>
### Vulnerability Details
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-17571">https://nvd.nist.gov/vuln/detail/CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j:log4j - 1.2.13-NODEP,1.2.17-atlassian-1,1.2.16.redhat-5;org.slf4j:nlog4j - 1.2.18</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | log4j-1.2.7.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.7/log4j-1.2.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/simple-maven-project-with-non-default-java-version/commit/ce1656cf6bc73035593c3c0a582f4f0943d6f771">ce1656cf6bc73035593c3c0a582f4f0943d6f771</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (log4j version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-17571](https://www.mend.io/vulnerability-database/CVE-2019-17571) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | log4j-1.2.7.jar | Direct | log4j:log4j - 1.2.13-NODEP,1.2.17-atlassian-1,1.2.16.redhat-5;org.slf4j:nlog4j - 1.2.18 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-17571</summary>
### Vulnerable Library - <b>log4j-1.2.7.jar</b></p>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.7/log4j-1.2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/simple-maven-project-with-non-default-java-version/commit/ce1656cf6bc73035593c3c0a582f4f0943d6f771">ce1656cf6bc73035593c3c0a582f4f0943d6f771</a></p>
</p>
<p></p>
### Vulnerability Details
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-17571">https://nvd.nist.gov/vuln/detail/CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j:log4j - 1.2.13-NODEP,1.2.17-atlassian-1,1.2.16.redhat-5;org.slf4j:nlog4j - 1.2.18</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_infrastructure | jar vulnerabilities highest severity is autoclosed vulnerable library jar path to dependency file pom xml path to vulnerable library home wss scanner repository jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in version remediation available high jar direct nodep atlassian redhat org details cve vulnerable library jar path to dependency file pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nodep atlassian redhat org rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
358,901 | 25,207,893,631 | IssuesEvent | 2022-11-13 22:15:41 | Krytic/Kaitiaki | https://api.github.com/repos/Krytic/Kaitiaki | opened | Code documentation | documentation enhancement good first issue | Kaitiaki is currently lacking in documentation -- someone needs to both write the manual (on RTD or similar, see me for details) and to properly docstring the code. | 1.0 | Code documentation - Kaitiaki is currently lacking in documentation -- someone needs to both write the manual (on RTD or similar, see me for details) and to properly docstring the code. | non_infrastructure | code documentation kaitiaki is currently lacking in documentation someone needs to both write the manual on rtd or similar see me for details and to properly docstring the code | 0 |
246,119 | 26,585,983,656 | IssuesEvent | 2023-01-23 01:04:21 | Vonage/vonage-ruby-code-snippets | https://api.github.com/repos/Vonage/vonage-ruby-code-snippets | opened | rack-contrib-2.3.0.gem: 3 vulnerabilities (highest severity is: 5.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-contrib-2.3.0.gem</b></p></summary>
<p></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rack-contrib version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-44571](https://www.mend.io/vulnerability-database/CVE-2022-44571) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
| [CVE-2022-44570](https://www.mend.io/vulnerability-database/CVE-2022-44570) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
| [CVE-2022-44572](https://www.mend.io/vulnerability-database/CVE-2022-44572) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44571</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a denial of service vulnerability in the Content-Disposition parsing component of Rack. Carefully crafted input can cause Content-Disposition header parsing in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. This header is used typically used in multipart parsing. Fixed Versions: 2.0.9.2, 2.1.4.2, 2.2.6.2, 3.0.4.1.
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44571>CVE-2022-44571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-93pm-5p5f-3ghx">https://github.com/advisories/GHSA-93pm-5p5f-3ghx</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44570</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a possible denial of service vulnerability in the Range header parsing component of Rack. Carefully crafted input can cause the Range header parsing component in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. Any applications that deal with Range requests (such as streaming applications, or applications that serve files) may be impacted. Fixed Versions: 2.0.9.2, 2.1.4.2, 2.2.6.2, 3.0.4.1
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44570>CVE-2022-44570</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-65f5-mfpf-vfhj">https://github.com/advisories/GHSA-65f5-mfpf-vfhj</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44572</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a denial of service vulnerability in the multipart parsing component of Rack. Carefully crafted input can cause RFC2183 multipart boundary parsing in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. Fixed Versions: 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44572>CVE-2022-44572</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-rqv2-275x-2jq5">https://github.com/advisories/GHSA-rqv2-275x-2jq5</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details> | True | rack-contrib-2.3.0.gem: 3 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-contrib-2.3.0.gem</b></p></summary>
<p></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rack-contrib version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-44571](https://www.mend.io/vulnerability-database/CVE-2022-44571) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
| [CVE-2022-44570](https://www.mend.io/vulnerability-database/CVE-2022-44570) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
| [CVE-2022-44572](https://www.mend.io/vulnerability-database/CVE-2022-44572) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | rack-2.2.4.gem | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44571</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a denial of service vulnerability in the Content-Disposition parsing component of Rack. Carefully crafted input can cause Content-Disposition header parsing in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. This header is used typically used in multipart parsing. Fixed Versions: 2.0.9.2, 2.1.4.2, 2.2.6.2, 3.0.4.1.
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44571>CVE-2022-44571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-93pm-5p5f-3ghx">https://github.com/advisories/GHSA-93pm-5p5f-3ghx</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44570</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a possible denial of service vulnerability in the Range header parsing component of Rack. Carefully crafted input can cause the Range header parsing component in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. Any applications that deal with Range requests (such as streaming applications, or applications that serve files) may be impacted. Fixed Versions: 2.0.9.2, 2.1.4.2, 2.2.6.2, 3.0.4.1
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44570>CVE-2022-44570</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-65f5-mfpf-vfhj">https://github.com/advisories/GHSA-65f5-mfpf-vfhj</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-44572</summary>
### Vulnerable Library - <b>rack-2.2.4.gem</b></p>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.4.gem">https://rubygems.org/gems/rack-2.2.4.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.4.gem</p>
<p>
Dependency Hierarchy:
- rack-contrib-2.3.0.gem (Root Library)
- :x: **rack-2.2.4.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
There is a denial of service vulnerability in the multipart parsing component of Rack. Carefully crafted input can cause RFC2183 multipart boundary parsing in Rack to take an unexpected amount of time, possibly resulting in a denial of service attack vector. Fixed Versions: 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1
<p>Publish Date: 2022-11-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-44572>CVE-2022-44572</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-rqv2-275x-2jq5">https://github.com/advisories/GHSA-rqv2-275x-2jq5</a></p>
<p>Release Date: 2022-11-02</p>
<p>Fix Resolution: rack - 2.0.9.2,2.1.4.2,2.2.6.2,3.0.4.1</p>
</p>
<p></p>
</details> | non_infrastructure | rack contrib gem vulnerabilities highest severity is vulnerable library rack contrib gem path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rack gem vulnerabilities cve severity cvss dependency type fixed in rack contrib version remediation available medium rack gem transitive n a medium rack gem transitive n a medium rack gem transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rack gem dependency hierarchy rack contrib gem root library x rack gem vulnerable library found in base branch master vulnerability details there is a denial of service vulnerability in the content disposition parsing component of rack carefully crafted input can cause content disposition header parsing in rack to take an unexpected amount of time possibly resulting in a denial of service attack vector this header is used typically used in multipart parsing fixed versions publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rack cve vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rack gem dependency hierarchy rack contrib gem root library x rack gem vulnerable library found in base branch master vulnerability details there is a possible denial of service vulnerability in the range header parsing component of rack carefully crafted input can cause the range header parsing component in rack to take an unexpected amount of time possibly resulting in a denial of service attack vector any applications that deal with range requests such as streaming applications or applications that serve files may be impacted fixed versions publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rack cve vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rack gem dependency hierarchy rack contrib gem root library x rack gem vulnerable library found in base branch master vulnerability details there is a denial of service vulnerability in the multipart parsing component of rack carefully crafted input can cause multipart boundary parsing in rack to take an unexpected amount of time possibly resulting in a denial of service attack vector fixed versions publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rack | 0 |
368,052 | 10,865,320,378 | IssuesEvent | 2019-11-14 18:44:32 | knative/website | https://api.github.com/repos/knative/website | closed | Docs version switcher uses both "development" and "Pre-release" | kind/enhancement kind/epic kind/question priority/2 size/M | In the dropdown of the version switcher, it lists "development" to see in-progress docs for the upcoming release. But, if you choose that version, the title of that same version switcher says "Pre-release". Should we choose just one of those terms and use it consistently in the version switcher? | 1.0 | Docs version switcher uses both "development" and "Pre-release" - In the dropdown of the version switcher, it lists "development" to see in-progress docs for the upcoming release. But, if you choose that version, the title of that same version switcher says "Pre-release". Should we choose just one of those terms and use it consistently in the version switcher? | non_infrastructure | docs version switcher uses both development and pre release in the dropdown of the version switcher it lists development to see in progress docs for the upcoming release but if you choose that version the title of that same version switcher says pre release should we choose just one of those terms and use it consistently in the version switcher | 0 |
59,113 | 3,103,119,212 | IssuesEvent | 2015-08-31 07:32:12 | tgstation/-tg-station | https://api.github.com/repos/tgstation/-tg-station | closed | All buttons on the station do not work | Bug Priority: CRITICAL | Problem Description: It looks like none of the buttons on the station are working. Medbay lobby, chaplain mass driver, chaplain crematory, toxins ignition switch, nothing.
What did you expect to happen: The buttons to work
What happened instead: They do nothing when you click on them.
Why is this bad/What are the consequences: Large portions of the game are unplayable simply because toxins scientists can't start a burn mix, and chaplains can't incinerate changelings.
Steps to reproduce the problem:
1. Start up game
2. Press buttons listed above
3. Watch them do nothing
Server: Sybil
Possibly related stuff: I checked all the button IDs and they are not the cause, I have no idea what is. | 1.0 | All buttons on the station do not work - Problem Description: It looks like none of the buttons on the station are working. Medbay lobby, chaplain mass driver, chaplain crematory, toxins ignition switch, nothing.
What did you expect to happen: The buttons to work
What happened instead: They do nothing when you click on them.
Why is this bad/What are the consequences: Large portions of the game are unplayable simply because toxins scientists can't start a burn mix, and chaplains can't incinerate changelings.
Steps to reproduce the problem:
1. Start up game
2. Press buttons listed above
3. Watch them do nothing
Server: Sybil
Possibly related stuff: I checked all the button IDs and they are not the cause, I have no idea what is. | non_infrastructure | all buttons on the station do not work problem description it looks like none of the buttons on the station are working medbay lobby chaplain mass driver chaplain crematory toxins ignition switch nothing what did you expect to happen the buttons to work what happened instead they do nothing when you click on them why is this bad what are the consequences large portions of the game are unplayable simply because toxins scientists can t start a burn mix and chaplains can t incinerate changelings steps to reproduce the problem start up game press buttons listed above watch them do nothing server sybil possibly related stuff i checked all the button ids and they are not the cause i have no idea what is | 0 |
34,978 | 30,646,152,277 | IssuesEvent | 2023-07-25 04:59:17 | utr1903/newrelic-as-a-service | https://api.github.com/repos/utr1903/newrelic-as-a-service | closed | Update Terraform providers | enhancement infrastructure | ## Summary
Update Terraform provider versions for `azurerm` and `newrelic`. | 1.0 | Update Terraform providers - ## Summary
Update Terraform provider versions for `azurerm` and `newrelic`. | infrastructure | update terraform providers summary update terraform provider versions for azurerm and newrelic | 1 |
21,956 | 14,941,787,372 | IssuesEvent | 2021-01-25 20:17:13 | dotnet/project-system | https://api.github.com/repos/dotnet/project-system | closed | Default branch will be renamed to main | Area-Infrastructure Triage-Approved | On Monday, 25 January 2021 we're going to rename the default branch in this repository to `main`. For more details, see [our earlier announcement](https://github.com/dotnet/announcements/issues/172).
| 1.0 | Default branch will be renamed to main - On Monday, 25 January 2021 we're going to rename the default branch in this repository to `main`. For more details, see [our earlier announcement](https://github.com/dotnet/announcements/issues/172).
| infrastructure | default branch will be renamed to main on monday january we re going to rename the default branch in this repository to main for more details see | 1 |
32,319 | 26,623,448,294 | IssuesEvent | 2023-01-24 12:59:45 | effektio/effektio | https://api.github.com/repos/effektio/effektio | opened | Bug reporting feature | bug flutter rust infrastructure | We want users (at least in nightly) to be able to quickly report bugs whenever they encounter them. For that, we decided to implement the following protocol:
- [ ] Setup a [matrix rageshake server](https://github.com/matrix-org/rageshake/)
- [ ] and private github repo to report the issues to
- [ ] Add a feature in the SDK to log to a path rather than stderr
- [ ] Allow the UI to ask for the file-path of the currently written file (or content of it)
- [ ] Implement the rageshake protocol for flutter (maybe as a standalone package!?!)
- [ ] Implement a simple button in the bottom side bar to report a bug, that opens a dialog allowing the user to give a description, allow the logs to be attached and use the rageshake protocol to report the bug
Extra:
- [ ] allow the user to attach a screenshot
- [ ] attach flutter logs, too
- [ ] detect that the app crashed on the last run and ask the user to report the bug - attaching the logs from the last run
- [ ] provide a trackable ID for the user at the end of it (github issue no would be best) | 1.0 | Bug reporting feature - We want users (at least in nightly) to be able to quickly report bugs whenever they encounter them. For that, we decided to implement the following protocol:
- [ ] Setup a [matrix rageshake server](https://github.com/matrix-org/rageshake/)
- [ ] and private github repo to report the issues to
- [ ] Add a feature in the SDK to log to a path rather than stderr
- [ ] Allow the UI to ask for the file-path of the currently written file (or content of it)
- [ ] Implement the rageshake protocol for flutter (maybe as a standalone package!?!)
- [ ] Implement a simple button in the bottom side bar to report a bug, that opens a dialog allowing the user to give a description, allow the logs to be attached and use the rageshake protocol to report the bug
Extra:
- [ ] allow the user to attach a screenshot
- [ ] attach flutter logs, too
- [ ] detect that the app crashed on the last run and ask the user to report the bug - attaching the logs from the last run
- [ ] provide a trackable ID for the user at the end of it (github issue no would be best) | infrastructure | bug reporting feature we want users at least in nightly to be able to quickly report bugs whenever they encounter them for that we decided to implement the following protocol setup a and private github repo to report the issues to add a feature in the sdk to log to a path rather than stderr allow the ui to ask for the file path of the currently written file or content of it implement the rageshake protocol for flutter maybe as a standalone package implement a simple button in the bottom side bar to report a bug that opens a dialog allowing the user to give a description allow the logs to be attached and use the rageshake protocol to report the bug extra allow the user to attach a screenshot attach flutter logs too detect that the app crashed on the last run and ask the user to report the bug attaching the logs from the last run provide a trackable id for the user at the end of it github issue no would be best | 1 |
27,345 | 21,645,908,278 | IssuesEvent | 2022-05-06 01:49:51 | Stellarium/stellarium | https://api.github.com/repos/Stellarium/stellarium | closed | Can't join the French translation team on Transifex | translations infrastructure | Hello there. For some reason, I do not have the option to join the French translation team. I can select other languages such as Spanish or German, but not French (France). Is this on purpose ?
thanks. | 1.0 | Can't join the French translation team on Transifex - Hello there. For some reason, I do not have the option to join the French translation team. I can select other languages such as Spanish or German, but not French (France). Is this on purpose ?
thanks. | infrastructure | can t join the french translation team on transifex hello there for some reason i do not have the option to join the french translation team i can select other languages such as spanish or german but not french france is this on purpose thanks | 1 |
27,898 | 22,588,180,624 | IssuesEvent | 2022-06-28 17:04:59 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | closed | Remove duplicates from the build files | refactoring infrastructure | **Description**
It is better to take out the common parts from `build.gradle` files to avoid duplication.
We have the code below in several build files, but it can be written once in the root `build.gradle` file.
```Gradle
publishing {
repositories {
maven {
name = "GitHubPackages"
url = "https://maven.pkg.github.com/UnitTestBot/UTBotJava"
credentials {
username = System.getenv("GITHUB_ACTOR")
password = System.getenv("GITHUB_TOKEN")
}
}
}
}
``` | 1.0 | Remove duplicates from the build files - **Description**
It is better to take out the common parts from `build.gradle` files to avoid duplication.
We have the code below in several build files, but it can be written once in the root `build.gradle` file.
```Gradle
publishing {
repositories {
maven {
name = "GitHubPackages"
url = "https://maven.pkg.github.com/UnitTestBot/UTBotJava"
credentials {
username = System.getenv("GITHUB_ACTOR")
password = System.getenv("GITHUB_TOKEN")
}
}
}
}
``` | infrastructure | remove duplicates from the build files description it is better to take out the common parts from build gradle files to avoid duplication we have the code below in several build files but it can be written once in the root build gradle file gradle publishing repositories maven name githubpackages url credentials username system getenv github actor password system getenv github token | 1 |
239,876 | 7,800,105,868 | IssuesEvent | 2018-06-09 04:53:48 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0007330:
customfield type record: error when trying to add self | Bug Mantis Tinebase high priority | **Reported by pschuele on 30 Oct 2012 20:22**
**Version:** git master
customfield type record: error when trying to add self
**Steps to reproduce:** 1. create customfield type record for any application (with same model)
2. edit/create record of app with customfield filled with record
3. add self as cf value + save record
-> crash
**Additional information:** Fatal error: Maximum function nesting level of '100' reached, aborting! in /usr/local/workspace/tine20/tine20/library/Zend/Db/Select.php on line 862 Call Stack: 0.0002 676528 1. {main}() /usr/local/workspace/tine20/tine20/index.php:0 0.0053 1618544 2. Tinebase_Core::dispatchRequest() /usr/local/workspace/tine20/tine20/index.php:17 0.0746 6959632 3. Tinebase_Server_Json->handle() /usr/local/workspace/tine20/tine20/Tinebase/Core.php:229 0.1437 14392408 4. Tinebase_Server_Json->_handle() /usr/local/workspace/tine20/tine20/Tinebase/Server/Json.php:78 0.1670 17962944 5. Zend_Json_Server->handle() /usr/local/workspace/tine20/tine20/Tinebase/Server/Json.php:206 0.1670 17962864 6. Zend_Json_Server->_handle() /usr/local/workspace/tine20/tine20/Zend/Json/Server.php:197 0.1671 17963296 7. Zend_Server_Abstract->_dispatch() /usr/local/workspace/tine20/tine20/Zend/Json/Server.php:558 0.1773 19324528 8. call_user_func_array() /usr/local/workspace/tine20/tine20/library/Zend/Server/Abstract.php:232 0.1773 19325032 9. Crm_Frontend_Json->saveLead() /usr/local/workspace/tine20/tine20/library/Zend/Server/Abstract.php:232 0.1773 19325168 10. Tinebase_Frontend_Json_Abstract->_save() /usr/local/workspace/tine20/tine20/Crm/Frontend/Json.php:82 0.1870 20077888 11. call_user_func_array() /usr/local/workspace/tine20/tine20/Tinebase/Frontend/Json/Abstract.php:268 0.1870 20078376 12. Tinebase_Controller_Record_Abstract->update() /usr/local/workspace/tine20/tine20/Tinebase/Frontend/Json/Abstract.php:268 0.3965 27231248 13. Tinebase_Controller_Record_Abstract->_setRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:640 0.4880 27755160 14. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:764 0.5205 27841272 15. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.5244 27852664 16. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.5276 27869896 17. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.5611 27970792 18. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.5649 27982184 19. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.5678 27999416 20. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.6004 28100152 21. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.6045 28111544 22. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.6079 28128776 23. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.6450 28229512 24. Tinebase_CustomField->resolveRecordCustomFields()
| 1.0 | 0007330:
customfield type record: error when trying to add self - **Reported by pschuele on 30 Oct 2012 20:22**
**Version:** git master
customfield type record: error when trying to add self
**Steps to reproduce:** 1. create customfield type record for any application (with same model)
2. edit/create record of app with customfield filled with record
3. add self as cf value + save record
-> crash
**Additional information:** Fatal error: Maximum function nesting level of '100' reached, aborting! in /usr/local/workspace/tine20/tine20/library/Zend/Db/Select.php on line 862 Call Stack: 0.0002 676528 1. {main}() /usr/local/workspace/tine20/tine20/index.php:0 0.0053 1618544 2. Tinebase_Core::dispatchRequest() /usr/local/workspace/tine20/tine20/index.php:17 0.0746 6959632 3. Tinebase_Server_Json->handle() /usr/local/workspace/tine20/tine20/Tinebase/Core.php:229 0.1437 14392408 4. Tinebase_Server_Json->_handle() /usr/local/workspace/tine20/tine20/Tinebase/Server/Json.php:78 0.1670 17962944 5. Zend_Json_Server->handle() /usr/local/workspace/tine20/tine20/Tinebase/Server/Json.php:206 0.1670 17962864 6. Zend_Json_Server->_handle() /usr/local/workspace/tine20/tine20/Zend/Json/Server.php:197 0.1671 17963296 7. Zend_Server_Abstract->_dispatch() /usr/local/workspace/tine20/tine20/Zend/Json/Server.php:558 0.1773 19324528 8. call_user_func_array() /usr/local/workspace/tine20/tine20/library/Zend/Server/Abstract.php:232 0.1773 19325032 9. Crm_Frontend_Json->saveLead() /usr/local/workspace/tine20/tine20/library/Zend/Server/Abstract.php:232 0.1773 19325168 10. Tinebase_Frontend_Json_Abstract->_save() /usr/local/workspace/tine20/tine20/Crm/Frontend/Json.php:82 0.1870 20077888 11. call_user_func_array() /usr/local/workspace/tine20/tine20/Tinebase/Frontend/Json/Abstract.php:268 0.1870 20078376 12. Tinebase_Controller_Record_Abstract->update() /usr/local/workspace/tine20/tine20/Tinebase/Frontend/Json/Abstract.php:268 0.3965 27231248 13. Tinebase_Controller_Record_Abstract->_setRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:640 0.4880 27755160 14. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:764 0.5205 27841272 15. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.5244 27852664 16. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.5276 27869896 17. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.5611 27970792 18. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.5649 27982184 19. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.5678 27999416 20. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.6004 28100152 21. Tinebase_CustomField->resolveRecordCustomFields() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:346 0.6045 28111544 22. Tinebase_Controller_Record_Abstract->get() /usr/local/workspace/tine20/tine20/Tinebase/CustomField.php:440 0.6079 28128776 23. Tinebase_Controller_Record_Abstract->_getRelatedData() /usr/local/workspace/tine20/tine20/Tinebase/Controller/Record/Abstract.php:319 0.6450 28229512 24. Tinebase_CustomField->resolveRecordCustomFields()
| non_infrastructure | customfield type record error when trying to add self reported by pschuele on oct version git master customfield type record error when trying to add self steps to reproduce create customfield type record for any application with same model edit create record of app with customfield filled with record add self as cf value save record gt crash additional information fatal error maximum function nesting level of reached aborting in usr local workspace library zend db select php on line call stack main usr local workspace index php tinebase core dispatchrequest usr local workspace index php tinebase server json gt handle usr local workspace tinebase core php tinebase server json gt handle usr local workspace tinebase server json php zend json server gt handle usr local workspace tinebase server json php zend json server gt handle usr local workspace zend json server php zend server abstract gt dispatch usr local workspace zend json server php call user func array usr local workspace library zend server abstract php crm frontend json gt savelead usr local workspace library zend server abstract php tinebase frontend json abstract gt save usr local workspace crm frontend json php call user func array usr local workspace tinebase frontend json abstract php tinebase controller record abstract gt update usr local workspace tinebase frontend json abstract php tinebase controller record abstract gt setrelateddata usr local workspace tinebase controller record abstract php tinebase controller record abstract gt getrelateddata usr local workspace tinebase controller record abstract php tinebase customfield gt resolverecordcustomfields usr local workspace tinebase controller record abstract php tinebase controller record abstract gt get usr local workspace tinebase customfield php tinebase controller record abstract gt getrelateddata usr local workspace tinebase controller record abstract php tinebase customfield gt resolverecordcustomfields usr local workspace tinebase controller record abstract php tinebase controller record abstract gt get usr local workspace tinebase customfield php tinebase controller record abstract gt getrelateddata usr local workspace tinebase controller record abstract php tinebase customfield gt resolverecordcustomfields usr local workspace tinebase controller record abstract php tinebase controller record abstract gt get usr local workspace tinebase customfield php tinebase controller record abstract gt getrelateddata usr local workspace tinebase controller record abstract php tinebase customfield gt resolverecordcustomfields | 0 |
24,448 | 17,270,593,977 | IssuesEvent | 2021-07-22 19:15:51 | google/iree | https://api.github.com/repos/google/iree | opened | Google benchmark returns partial JSON results when invoked with Python scripts | infrastructure/benchmarking | This issue is a bit weird. It happens when I'm trying to switch the benchmarking phone from my dev phone to the one in the lab (which now is properly rooted). But at least it's deterministic.
We call the `iree-benchmark-module` binary with JSON output format in the benchmark pipeline with Python subprocess. For "MobileNetV2 [fp32,imagenet] (TensorFlow) 1-thread,big-core,full-inference with IREE-Dylib @ SM-G980F (CPU-ARMv8.2-A)", looks `iree-benchmark-module` returns the output in two batches; the first batch just contains the context. It still keeps running and then after a while the results for benchmark will be returned. Somehow, Python thinks `iree-benchmark-module` has completed after seeing the first batch, which is just partial and invalid JSON. So we are seeing errors like [this one](https://buildkite.com/iree/iree-benchmark/builds/505#ab1d9078-8b91-40c4-a09e-3d72f44d7564).
We are [calling into subprocess.run](https://github.com/google/iree/blob/471a36d7b4edfa69b8bd78e97adc216f4a083041/build_tools/android/common/benchmark_description.py#L36-L45), so it can be that I'm using subprocess.run wrongly or there is an issue with subprocess.run itself.. | 1.0 | Google benchmark returns partial JSON results when invoked with Python scripts - This issue is a bit weird. It happens when I'm trying to switch the benchmarking phone from my dev phone to the one in the lab (which now is properly rooted). But at least it's deterministic.
We call the `iree-benchmark-module` binary with JSON output format in the benchmark pipeline with Python subprocess. For "MobileNetV2 [fp32,imagenet] (TensorFlow) 1-thread,big-core,full-inference with IREE-Dylib @ SM-G980F (CPU-ARMv8.2-A)", looks `iree-benchmark-module` returns the output in two batches; the first batch just contains the context. It still keeps running and then after a while the results for benchmark will be returned. Somehow, Python thinks `iree-benchmark-module` has completed after seeing the first batch, which is just partial and invalid JSON. So we are seeing errors like [this one](https://buildkite.com/iree/iree-benchmark/builds/505#ab1d9078-8b91-40c4-a09e-3d72f44d7564).
We are [calling into subprocess.run](https://github.com/google/iree/blob/471a36d7b4edfa69b8bd78e97adc216f4a083041/build_tools/android/common/benchmark_description.py#L36-L45), so it can be that I'm using subprocess.run wrongly or there is an issue with subprocess.run itself.. | infrastructure | google benchmark returns partial json results when invoked with python scripts this issue is a bit weird it happens when i m trying to switch the benchmarking phone from my dev phone to the one in the lab which now is properly rooted but at least it s deterministic we call the iree benchmark module binary with json output format in the benchmark pipeline with python subprocess for tensorflow thread big core full inference with iree dylib sm cpu a looks iree benchmark module returns the output in two batches the first batch just contains the context it still keeps running and then after a while the results for benchmark will be returned somehow python thinks iree benchmark module has completed after seeing the first batch which is just partial and invalid json so we are seeing errors like we are so it can be that i m using subprocess run wrongly or there is an issue with subprocess run itself | 1 |
12,455 | 9,791,877,744 | IssuesEvent | 2019-06-10 16:00:36 | desktop/desktop | https://api.github.com/repos/desktop/desktop | closed | How to set up auto update server? | infrastructure meta | Currently I'm reading the code on auto update of GitHub Desktop, I can understand the part in GitHub Desktop client, I found it used the autoUpdater of Electron. For the server side, I found it used an API endpoint to see if there is a new version of the app. If there is no new version, the endpoint will return HTTP 204. If there is a new version, it will return a HTTP 200 JSON-formatted response, containing a url to a .zip file, and I found both Wndows and MacOS use the same API endpoint, but I have some questions on the server of the auto update: Since the API endpoint only returns a .zip file for macos, how does the server response for Windows requests? What's the difference for Windows and Macos in the server side? Can you share some knowledge on how can I build my own auto update server to let GitHub Desktop work? Thank you. | 1.0 | How to set up auto update server? - Currently I'm reading the code on auto update of GitHub Desktop, I can understand the part in GitHub Desktop client, I found it used the autoUpdater of Electron. For the server side, I found it used an API endpoint to see if there is a new version of the app. If there is no new version, the endpoint will return HTTP 204. If there is a new version, it will return a HTTP 200 JSON-formatted response, containing a url to a .zip file, and I found both Wndows and MacOS use the same API endpoint, but I have some questions on the server of the auto update: Since the API endpoint only returns a .zip file for macos, how does the server response for Windows requests? What's the difference for Windows and Macos in the server side? Can you share some knowledge on how can I build my own auto update server to let GitHub Desktop work? Thank you. | infrastructure | how to set up auto update server currently i m reading the code on auto update of github desktop i can understand the part in github desktop client i found it used the autoupdater of electron for the server side i found it used an api endpoint to see if there is a new version of the app if there is no new version the endpoint will return http if there is a new version it will return a http json formatted response containing a url to a zip file and i found both wndows and macos use the same api endpoint but i have some questions on the server of the auto update since the api endpoint only returns a zip file for macos how does the server response for windows requests what s the difference for windows and macos in the server side can you share some knowledge on how can i build my own auto update server to let github desktop work thank you | 1 |
21,947 | 14,939,966,452 | IssuesEvent | 2021-01-25 17:37:42 | cabforum/forum | https://api.github.com/repos/cabforum/forum | opened | Infra: Create a GitHub Bot Account | infrastructure | We want to be able to automate the release of Docker Images, such as https://github.com/cabforum/build-guidelines-action .
This currently [requires a PAT](https://docs.github.com/en/packages/guides/about-github-container-registry), and we should ideally use a GitHub [Bot Account](https://docs.github.com/en/github/getting-started-with-github/types-of-github-accounts) with scoped permissions. | 1.0 | Infra: Create a GitHub Bot Account - We want to be able to automate the release of Docker Images, such as https://github.com/cabforum/build-guidelines-action .
This currently [requires a PAT](https://docs.github.com/en/packages/guides/about-github-container-registry), and we should ideally use a GitHub [Bot Account](https://docs.github.com/en/github/getting-started-with-github/types-of-github-accounts) with scoped permissions. | infrastructure | infra create a github bot account we want to be able to automate the release of docker images such as this currently and we should ideally use a github with scoped permissions | 1 |
13,500 | 8,518,914,415 | IssuesEvent | 2018-11-01 13:16:41 | sandstorm-io/sandstorm | https://api.github.com/repos/sandstorm-io/sandstorm | closed | Sharing visibility | usability | It would be nice to see the Whom shared with button on the top IF a document has been shared.
Now first the "+ Share access" drop down has to be open to be able to open it. (If it is opened it shows a button on the topbar, but removes the button if the shared with view is closed).
| True | Sharing visibility - It would be nice to see the Whom shared with button on the top IF a document has been shared.
Now first the "+ Share access" drop down has to be open to be able to open it. (If it is opened it shows a button on the topbar, but removes the button if the shared with view is closed).
| non_infrastructure | sharing visibility it would be nice to see the whom shared with button on the top if a document has been shared now first the share access drop down has to be open to be able to open it if it is opened it shows a button on the topbar but removes the button if the shared with view is closed | 0 |
33,512 | 27,533,641,304 | IssuesEvent | 2023-03-07 00:48:11 | OpenHistoricalMap/issues | https://api.github.com/repos/OpenHistoricalMap/issues | closed | Deployment approach for Nominatim UI and other frontend applications | infrastructure | We have this:
https://nominatim-ohm-production.surge.sh/search.html
Code at https://github.com/OpenHistoricalMap/nominatim-ui
Which eventually we'd want to deploy to AWS infrastructure and live at nominatim.openhistoricalmap.org.
We could do that either into S3 buckets via GH Actions in the `nominatim-ui` repo, which is simple and easy but also disjointed from our main `ohm-deploy` repo and process.
In discussing this, we realized we also have other front-end only apps and should probably deploy them all in a similar way:
* https://github.com/OpenHistoricalMap/tasking-manager is deploying to S3 via GH Actions, https://github.com/OpenHistoricalMap/tasking-manager/tree/staging/.github/workflows
* https://github.com/openhistoricalmap/overpass-turbo/ is deployed via GH Pages
Since we want Production and Staging environments, I think GH Pages-only deploy is not a good option, so this ticket would be to decide between OSM Seed to containers vs GH Actions to S3.
| 1.0 | Deployment approach for Nominatim UI and other frontend applications - We have this:
https://nominatim-ohm-production.surge.sh/search.html
Code at https://github.com/OpenHistoricalMap/nominatim-ui
Which eventually we'd want to deploy to AWS infrastructure and live at nominatim.openhistoricalmap.org.
We could do that either into S3 buckets via GH Actions in the `nominatim-ui` repo, which is simple and easy but also disjointed from our main `ohm-deploy` repo and process.
In discussing this, we realized we also have other front-end only apps and should probably deploy them all in a similar way:
* https://github.com/OpenHistoricalMap/tasking-manager is deploying to S3 via GH Actions, https://github.com/OpenHistoricalMap/tasking-manager/tree/staging/.github/workflows
* https://github.com/openhistoricalmap/overpass-turbo/ is deployed via GH Pages
Since we want Production and Staging environments, I think GH Pages-only deploy is not a good option, so this ticket would be to decide between OSM Seed to containers vs GH Actions to S3.
| infrastructure | deployment approach for nominatim ui and other frontend applications we have this code at which eventually we d want to deploy to aws infrastructure and live at nominatim openhistoricalmap org we could do that either into buckets via gh actions in the nominatim ui repo which is simple and easy but also disjointed from our main ohm deploy repo and process in discussing this we realized we also have other front end only apps and should probably deploy them all in a similar way is deploying to via gh actions is deployed via gh pages since we want production and staging environments i think gh pages only deploy is not a good option so this ticket would be to decide between osm seed to containers vs gh actions to | 1 |
619,580 | 19,529,953,196 | IssuesEvent | 2021-12-30 14:57:27 | amplication/amplication | https://api.github.com/repos/amplication/amplication | closed | E2E tests are failing - release 0.11.0 | type: bug priority: high | ### What happened?
GitHub `CI `Workflow is failing in `build` job in `Run test data service's tests` step.
### What you expected to happen
FAIL src/user/base/user.controller.base.spec.ts
โ Test suite failed to run
src/user/base/user.controller.base.ts:63:7 - error TS2322: Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type '(Without<UserCreateInput, UserUncheckedCreateInput> & UserUncheckedCreateInput) | (Without<...> & UserCreateInput)'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type 'Without<UserUncheckedCreateInput, UserCreateInput> & UserCreateInput'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type 'UserCreateInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonValue' is not assignable to type 'InputJsonValue'.
Type 'null' is not assignable to type 'InputJsonValue'.
63 data: {
~~~~
node_modules/.prisma/client/index.d.ts:1933:5
1933 data: XOR<UserCreateInput, UserUncheckedCreateInput>
~~~~
The expected type comes from property 'data' which is declared here on type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; data: (Without<UserCreateInput, UserUncheckedCreateInput> & UserUncheckedCreateInput) | (Without<...> & UserCreateInput); }'
src/user/base/user.controller.base.ts:128:49 - error TS2345: Argument of type '{ select: { username: true; roles: true; id: true; name: true; bio: true; email: true; age: true; birthDate: true; score: true; manager: { select: { id: true; }; }; interests: true; priority: true; isCurious: true; location: true; extendedProperties: true; }; where?: UserWhereInput | undefined; orderBy?: UserOrderBy...' is not assignable to parameter of type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; where?: UserWhereInput | undefined; orderBy?: Enumerable<UserOrderByWithRelationInput> | undefined; cursor?: UserWhereUniqueInput | undefined; take?: number | undefined; skip?: number | undefined; distinct?: Enumerable<...> | undefin...'.
Types of property 'where' are incompatible.
Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput | undefined' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput | undefined'.
Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonNullableFilter | undefined' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type '{ equals?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; not?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; }'.
Types of property 'equals' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
Type 'null' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
128 const results = await this.service.findMany({
~
129 ...args,
~~~~~~~~~~~~~~
...
152 },
~~~~~~~~
153 });
~~~~~
src/user/base/user.controller.base.ts:256:9 - error TS2322: Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type '(Without<UserUpdateInput, UserUncheckedUpdateInput> & UserUncheckedUpdateInput) | (Without<...> & UserUpdateInput)'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type 'Without<UserUncheckedUpdateInput, UserUpdateInput> & UserUpdateInput'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type 'UserUpdateInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'InputJsonValue | undefined'.
Type 'null' is not assignable to type 'InputJsonValue | undefined'.
256 data: {
~~~~
node_modules/.prisma/client/index.d.ts:1964:5
1964 data: XOR<UserUpdateInput, UserUncheckedUpdateInput>
~~~~
The expected type comes from property 'data' which is declared here on type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; data: (Without<UserUpdateInput, UserUncheckedUpdateInput> & UserUncheckedUpdateInput) | (Without<...> & UserUpdateInput); where: UserWhereUniqueInput; }'
src/user/base/user.controller.base.ts:382:7 - error TS2322: Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
382 where: query,
~~~~~
node_modules/.prisma/client/index.d.ts:1882:5
1882 where?: UserWhereInput
~~~~~
The expected type comes from property 'where' which is declared here on type 'UserFindManyArgs'
FAIL src/organization/base/organization.controller.base.spec.ts
โ Test suite failed to run
src/organization/base/organization.controller.base.ts:283:7 - error TS2322: Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonNullableFilter | undefined' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type '{ equals?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; not?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; }'.
Types of property 'equals' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
Type 'null' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
283 where: query,
~~~~~
node_modules/.prisma/client/index.d.ts:1882:5
1882 where?: UserWhereInput
~~~~~
The expected type comes from property 'where' which is declared here on type 'UserFindManyArgs'
### How to reproduce
Run CI workflow in the master branch - release 0.11.0
### Amplication version
0.11.0
### Environment
node: 16.3.1
npm: 8.2.1
Platform: GitHub Actions
OS: Ubuntu-20
### Anything else
_No response_
### Are you willing to submit PR?
_No response_
### Code of Conduct
I agree to follow this project's Code of Conduct | 1.0 | E2E tests are failing - release 0.11.0 - ### What happened?
GitHub `CI `Workflow is failing in `build` job in `Run test data service's tests` step.
### What you expected to happen
FAIL src/user/base/user.controller.base.spec.ts
โ Test suite failed to run
src/user/base/user.controller.base.ts:63:7 - error TS2322: Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type '(Without<UserCreateInput, UserUncheckedCreateInput> & UserUncheckedCreateInput) | (Without<...> & UserCreateInput)'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type 'Without<UserUncheckedCreateInput, UserCreateInput> & UserCreateInput'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username: string; password: string; roles: string[]; name: string; bio: string; email: string; age: number; birthDate: Date; ... 5 more ...; extendedProperties: JsonValue; }' is not assignable to type 'UserCreateInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonValue' is not assignable to type 'InputJsonValue'.
Type 'null' is not assignable to type 'InputJsonValue'.
63 data: {
~~~~
node_modules/.prisma/client/index.d.ts:1933:5
1933 data: XOR<UserCreateInput, UserUncheckedCreateInput>
~~~~
The expected type comes from property 'data' which is declared here on type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; data: (Without<UserCreateInput, UserUncheckedCreateInput> & UserUncheckedCreateInput) | (Without<...> & UserCreateInput); }'
src/user/base/user.controller.base.ts:128:49 - error TS2345: Argument of type '{ select: { username: true; roles: true; id: true; name: true; bio: true; email: true; age: true; birthDate: true; score: true; manager: { select: { id: true; }; }; interests: true; priority: true; isCurious: true; location: true; extendedProperties: true; }; where?: UserWhereInput | undefined; orderBy?: UserOrderBy...' is not assignable to parameter of type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; where?: UserWhereInput | undefined; orderBy?: Enumerable<UserOrderByWithRelationInput> | undefined; cursor?: UserWhereUniqueInput | undefined; take?: number | undefined; skip?: number | undefined; distinct?: Enumerable<...> | undefin...'.
Types of property 'where' are incompatible.
Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput | undefined' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput | undefined'.
Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonNullableFilter | undefined' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type '{ equals?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; not?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; }'.
Types of property 'equals' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
Type 'null' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
128 const results = await this.service.findMany({
~
129 ...args,
~~~~~~~~~~~~~~
...
152 },
~~~~~~~~
153 });
~~~~~
src/user/base/user.controller.base.ts:256:9 - error TS2322: Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type '(Without<UserUpdateInput, UserUncheckedUpdateInput> & UserUncheckedUpdateInput) | (Without<...> & UserUpdateInput)'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type 'Without<UserUncheckedUpdateInput, UserUpdateInput> & UserUpdateInput'.
Type '{ manager: { connect: UserWhereUniqueInput; } | undefined; username?: string | undefined; password?: string | undefined; roles?: string[] | undefined; name?: string | undefined; ... 9 more ...; extendedProperties?: JsonValue | undefined; }' is not assignable to type 'UserUpdateInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'InputJsonValue | undefined'.
Type 'null' is not assignable to type 'InputJsonValue | undefined'.
256 data: {
~~~~
node_modules/.prisma/client/index.d.ts:1964:5
1964 data: XOR<UserUpdateInput, UserUncheckedUpdateInput>
~~~~
The expected type comes from property 'data' which is declared here on type '{ select?: UserSelect | null | undefined; include?: UserInclude | null | undefined; data: (Without<UserUpdateInput, UserUncheckedUpdateInput> & UserUncheckedUpdateInput) | (Without<...> & UserUpdateInput); where: UserWhereUniqueInput; }'
src/user/base/user.controller.base.ts:382:7 - error TS2322: Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
382 where: query,
~~~~~
node_modules/.prisma/client/index.d.ts:1882:5
1882 where?: UserWhereInput
~~~~~
The expected type comes from property 'where' which is declared here on type 'UserFindManyArgs'
FAIL src/organization/base/organization.controller.base.spec.ts
โ Test suite failed to run
src/organization/base/organization.controller.base.ts:283:7 - error TS2322: Type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/src/user/base/UserWhereInput").UserWhereInput' is not assignable to type 'import("/home/runner/work/amplication-hosted/amplication-hosted/packages/amplication-data-service-generator/generated/server/node_modules/.prisma/client/index").Prisma.UserWhereInput'.
Types of property 'extendedProperties' are incompatible.
Type 'JsonNullableFilter | undefined' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type 'JsonFilter | undefined'.
Type 'JsonNullableFilter' is not assignable to type '{ equals?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; not?: string | number | boolean | InputJsonObject | InputJsonArray | undefined; }'.
Types of property 'equals' are incompatible.
Type 'JsonValue | undefined' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
Type 'null' is not assignable to type 'string | number | boolean | InputJsonObject | InputJsonArray | undefined'.
283 where: query,
~~~~~
node_modules/.prisma/client/index.d.ts:1882:5
1882 where?: UserWhereInput
~~~~~
The expected type comes from property 'where' which is declared here on type 'UserFindManyArgs'
### How to reproduce
Run CI workflow in the master branch - release 0.11.0
### Amplication version
0.11.0
### Environment
node: 16.3.1
npm: 8.2.1
Platform: GitHub Actions
OS: Ubuntu-20
### Anything else
_No response_
### Are you willing to submit PR?
_No response_
### Code of Conduct
I agree to follow this project's Code of Conduct | non_infrastructure | tests are failing release what happened github ci workflow is failing in build job in run test data service s tests step what you expected to happen fail src user base user controller base spec ts โ test suite failed to run src user base user controller base ts error type manager connect userwhereuniqueinput undefined username string password string roles string name string bio string email string age number birthdate date more extendedproperties jsonvalue is not assignable to type without useruncheckedcreateinput without usercreateinput type manager connect userwhereuniqueinput undefined username string password string roles string name string bio string email string age number birthdate date more extendedproperties jsonvalue is not assignable to type without usercreateinput type manager connect userwhereuniqueinput undefined username string password string roles string name string bio string email string age number birthdate date more extendedproperties jsonvalue is not assignable to type usercreateinput types of property extendedproperties are incompatible type jsonvalue is not assignable to type inputjsonvalue type null is not assignable to type inputjsonvalue data node modules prisma client index d ts data xor the expected type comes from property data which is declared here on type select userselect null undefined include userinclude null undefined data without useruncheckedcreateinput without usercreateinput src user base user controller base ts error argument of type select username true roles true id true name true bio true email true age true birthdate true score true manager select id true interests true priority true iscurious true location true extendedproperties true where userwhereinput undefined orderby userorderby is not assignable to parameter of type select userselect null undefined include userinclude null undefined where userwhereinput undefined orderby enumerable undefined cursor userwhereuniqueinput undefined take number undefined skip number undefined distinct enumerable undefin types of property where are incompatible type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server src user base userwhereinput userwhereinput undefined is not assignable to type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server node modules prisma client index prisma userwhereinput undefined type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server src user base userwhereinput userwhereinput is not assignable to type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server node modules prisma client index prisma userwhereinput types of property extendedproperties are incompatible type jsonnullablefilter undefined is not assignable to type jsonfilter undefined type jsonnullablefilter is not assignable to type jsonfilter undefined type jsonnullablefilter is not assignable to type equals string number boolean inputjsonobject inputjsonarray undefined not string number boolean inputjsonobject inputjsonarray undefined types of property equals are incompatible type jsonvalue undefined is not assignable to type string number boolean inputjsonobject inputjsonarray undefined type null is not assignable to type string number boolean inputjsonobject inputjsonarray undefined const results await this service findmany args src user base user controller base ts error type manager connect userwhereuniqueinput undefined username string undefined password string undefined roles string undefined name string undefined more extendedproperties jsonvalue undefined is not assignable to type without useruncheckedupdateinput without userupdateinput type manager connect userwhereuniqueinput undefined username string undefined password string undefined roles string undefined name string undefined more extendedproperties jsonvalue undefined is not assignable to type without userupdateinput type manager connect userwhereuniqueinput undefined username string undefined password string undefined roles string undefined name string undefined more extendedproperties jsonvalue undefined is not assignable to type userupdateinput types of property extendedproperties are incompatible type jsonvalue undefined is not assignable to type inputjsonvalue undefined type null is not assignable to type inputjsonvalue undefined data node modules prisma client index d ts data xor the expected type comes from property data which is declared here on type select userselect null undefined include userinclude null undefined data without useruncheckedupdateinput without userupdateinput where userwhereuniqueinput src user base user controller base ts error type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server src user base userwhereinput userwhereinput is not assignable to type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server node modules prisma client index prisma userwhereinput where query node modules prisma client index d ts where userwhereinput the expected type comes from property where which is declared here on type userfindmanyargs fail src organization base organization controller base spec ts โ test suite failed to run src organization base organization controller base ts error type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server src user base userwhereinput userwhereinput is not assignable to type import home runner work amplication hosted amplication hosted packages amplication data service generator generated server node modules prisma client index prisma userwhereinput types of property extendedproperties are incompatible type jsonnullablefilter undefined is not assignable to type jsonfilter undefined type jsonnullablefilter is not assignable to type jsonfilter undefined type jsonnullablefilter is not assignable to type equals string number boolean inputjsonobject inputjsonarray undefined not string number boolean inputjsonobject inputjsonarray undefined types of property equals are incompatible type jsonvalue undefined is not assignable to type string number boolean inputjsonobject inputjsonarray undefined type null is not assignable to type string number boolean inputjsonobject inputjsonarray undefined where query node modules prisma client index d ts where userwhereinput the expected type comes from property where which is declared here on type userfindmanyargs how to reproduce run ci workflow in the master branch release amplication version environment node npm platform github actions os ubuntu anything else no response are you willing to submit pr no response code of conduct i agree to follow this project s code of conduct | 0 |
33,932 | 28,008,150,369 | IssuesEvent | 2023-03-27 16:34:35 | ministryofjustice/data-platform | https://api.github.com/repos/ministryofjustice/data-platform | closed | ๐ Create Route53 zone for Common Fate deployment | Data Platform Core Infrastructure | `access.platform.service.justice.gov.uk`
- [x] https://github.com/ministryofjustice/aws-root-account/pull/718
- [x] https://github.com/ministryofjustice/dns-iac/pull/395 | 1.0 | ๐ Create Route53 zone for Common Fate deployment - `access.platform.service.justice.gov.uk`
- [x] https://github.com/ministryofjustice/aws-root-account/pull/718
- [x] https://github.com/ministryofjustice/dns-iac/pull/395 | infrastructure | ๐ create zone for common fate deployment access platform service justice gov uk | 1 |
365,035 | 25,517,912,383 | IssuesEvent | 2022-11-28 17:50:41 | crossplane-contrib/provider-ansible | https://api.github.com/repos/crossplane-contrib/provider-ansible | closed | Docs improve with example how to use ssh key with ansible-run | documentation |
### What problem are you facing?
at the moment it is not entirely clear how to use the ssx key to connect to remote hosts while using this provider
### How could Crossplane help solve your problem?
improve the documentation by adding an example of a similar case
| 1.0 | Docs improve with example how to use ssh key with ansible-run -
### What problem are you facing?
at the moment it is not entirely clear how to use the ssx key to connect to remote hosts while using this provider
### How could Crossplane help solve your problem?
improve the documentation by adding an example of a similar case
| non_infrastructure | docs improve with example how to use ssh key with ansible run what problem are you facing at the moment it is not entirely clear how to use the ssx key to connect to remote hosts while using this provider how could crossplane help solve your problem improve the documentation by adding an example of a similar case | 0 |
53,137 | 27,984,060,427 | IssuesEvent | 2023-03-26 13:49:14 | pmmp/PocketMine-MP | https://api.github.com/repos/pmmp/PocketMine-MP | closed | Avoid triggering network block / chunk updates for non-visual changes | Category: Core Type: Enhancement Performance | <!--- tell us what you want -->
## Description
Many blocks have technical properties which don't affect their visual state. We should avoid sending changes to these properties over network.
This includes things like cactus age, leaf decay flags, daylight sensor power level and more. As such it's not necessary to notify clients that these properties have been changed when they change on the server side, meaning that we can avoid unnecessary cache destruction and reduce bandwidth costs. | True | Avoid triggering network block / chunk updates for non-visual changes - <!--- tell us what you want -->
## Description
Many blocks have technical properties which don't affect their visual state. We should avoid sending changes to these properties over network.
This includes things like cactus age, leaf decay flags, daylight sensor power level and more. As such it's not necessary to notify clients that these properties have been changed when they change on the server side, meaning that we can avoid unnecessary cache destruction and reduce bandwidth costs. | non_infrastructure | avoid triggering network block chunk updates for non visual changes description many blocks have technical properties which don t affect their visual state we should avoid sending changes to these properties over network this includes things like cactus age leaf decay flags daylight sensor power level and more as such it s not necessary to notify clients that these properties have been changed when they change on the server side meaning that we can avoid unnecessary cache destruction and reduce bandwidth costs | 0 |
105,745 | 11,457,337,194 | IssuesEvent | 2020-02-06 23:26:35 | phantomlei3/InfoRoots | https://api.github.com/repos/phantomlei3/InfoRoots | closed | Sprint 1: Vision Statements | documentation good first issue | The vision statement should include an executive summary of the project, including a problem/solution statement as well as an elevator summary, a business case for the project (including a short analysis of the competition, market space or similar commercial off the shelf software, with links), a detailed description of the project stakeholders, a list of the major features of the completed project, and a list of the major risks of the project. | 1.0 | Sprint 1: Vision Statements - The vision statement should include an executive summary of the project, including a problem/solution statement as well as an elevator summary, a business case for the project (including a short analysis of the competition, market space or similar commercial off the shelf software, with links), a detailed description of the project stakeholders, a list of the major features of the completed project, and a list of the major risks of the project. | non_infrastructure | sprint vision statements the vision statement should include an executive summary of the project including a problem solution statement as well as an elevator summary a business case for the project including a short analysis of the competition market space or similar commercial off the shelf software with links a detailed description of the project stakeholders a list of the major features of the completed project and a list of the major risks of the project | 0 |
151,326 | 23,799,431,927 | IssuesEvent | 2022-09-03 03:35:52 | heeyoung-c/react-kdt-mini-project-team1 | https://api.github.com/repos/heeyoung-c/react-kdt-mini-project-team1 | closed | SignIn/๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ | ๐design | ## <img src='https://emojis.slackmojis.com/emojis/images/1643514738/7421/typingcat.gif?1643514738' alt='์์
์ฌํญ' width=30px> ์์
์ฌํญ
design
- [x] ๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ
chore
- [x] ์ฌ์ฉํ์ง ์๋ ์ฝ๋, ์ฝ์๋ก๊ทธ, ์ฃผ์ ์ ๊ฑฐ
| 1.0 | SignIn/๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ - ## <img src='https://emojis.slackmojis.com/emojis/images/1643514738/7421/typingcat.gif?1643514738' alt='์์
์ฌํญ' width=30px> ์์
์ฌํญ
design
- [x] ๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ
chore
- [x] ์ฌ์ฉํ์ง ์๋ ์ฝ๋, ์ฝ์๋ก๊ทธ, ์ฃผ์ ์ ๊ฑฐ
| non_infrastructure | signin ๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ ์์
์ฌํญ design ๋ก๊ทธ์ธ ๋ฐ ํ์๊ฐ์
์ฐฝ ์ค์ ์ ๋ ฌ chore ์ฌ์ฉํ์ง ์๋ ์ฝ๋ ์ฝ์๋ก๊ทธ ์ฃผ์ ์ ๊ฑฐ | 0 |
101,972 | 12,734,694,584 | IssuesEvent | 2020-06-25 14:16:39 | cds-snc/notification-api | https://api.github.com/repos/cds-snc/notification-api | closed | Adjust content on 'Email resent' page | Accessibility Design S/P :tshirt: | Link text is unclear with โNot received an emailโ should be a question or call to action in English, be more descriptive of the outcome.
Heading is read by immersive screen reader as resent (resentment) rather than re-sent.
https://notification.alpha.canada.ca/email-not-received
https://notification.alpha.canada.ca/two-factor-email-sent?email_resent=True

| 1.0 | Adjust content on 'Email resent' page - Link text is unclear with โNot received an emailโ should be a question or call to action in English, be more descriptive of the outcome.
Heading is read by immersive screen reader as resent (resentment) rather than re-sent.
https://notification.alpha.canada.ca/email-not-received
https://notification.alpha.canada.ca/two-factor-email-sent?email_resent=True

| non_infrastructure | adjust content on email resent page link text is unclear with โnot received an emailโ should be a question or call to action in english be more descriptive of the outcome heading is read by immersive screen reader as resent resentment rather than re sent | 0 |
829,052 | 31,852,285,371 | IssuesEvent | 2023-09-15 03:24:39 | 4paradigm/OpenMLDB | https://api.github.com/repos/4paradigm/OpenMLDB | closed | Remove the config of `namenode.uri` in TaskManager | high-priority task-manager | We may support reading HDFS without setting `namenode.uri`. | 1.0 | Remove the config of `namenode.uri` in TaskManager - We may support reading HDFS without setting `namenode.uri`. | non_infrastructure | remove the config of namenode uri in taskmanager we may support reading hdfs without setting namenode uri | 0 |
18,645 | 13,060,068,321 | IssuesEvent | 2020-07-30 11:47:43 | epigraphic-database-heidelberg/website | https://api.github.com/repos/epigraphic-database-heidelberg/website | closed | Cantaloupe IIIF: symlink tiffs to /home/frgri/images | infrastructure | create symlinks for all existing tiffs in /home/frgri/images => image folder for Cantaloupe IIIF server, as defined in cantaloupe.properties:
FilesystemSource.BasicLookupStrategy.path_prefix = /home/frgri/images/ | 1.0 | Cantaloupe IIIF: symlink tiffs to /home/frgri/images - create symlinks for all existing tiffs in /home/frgri/images => image folder for Cantaloupe IIIF server, as defined in cantaloupe.properties:
FilesystemSource.BasicLookupStrategy.path_prefix = /home/frgri/images/ | infrastructure | cantaloupe iiif symlink tiffs to home frgri images create symlinks for all existing tiffs in home frgri images image folder for cantaloupe iiif server as defined in cantaloupe properties filesystemsource basiclookupstrategy path prefix home frgri images | 1 |
33,633 | 27,638,231,813 | IssuesEvent | 2023-03-10 15:59:48 | rusefi/rusefi | https://api.github.com/repos/rusefi/rusefi | closed | Even 'only' build should produce proper bundle/autoupdate/upload .ini | INFRASTRUCTURE | The amazing 'only:xxx' feature should behave closer to normal build, just only for one matrix entry | 1.0 | Even 'only' build should produce proper bundle/autoupdate/upload .ini - The amazing 'only:xxx' feature should behave closer to normal build, just only for one matrix entry | infrastructure | even only build should produce proper bundle autoupdate upload ini the amazing only xxx feature should behave closer to normal build just only for one matrix entry | 1 |
9,484 | 2,615,153,235 | IssuesEvent | 2015-03-01 06:30:54 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | This is a query not an issue: regarding saving position. | auto-migrated Priority-Triage Type-Defect | ```
Good morning and please pardon my ignorance.
My query is regarding the saving of the session.
Does Reaver automatically save position as it running or only when you stop the
app and it says it has saved the session?
My main reason of question is that the router I am testing at the moment
(Provided by Sky broadband UK) allows between 20 and 24 (unsure why it changes)
attempts and then will always lock out for 1 hour.
So my estimated maximum calculation going based on 20 keys per hour to reach
11000 keys is 500 hours which is about 21 days to crack this password.
Now that is a very long time! and if my computer was to crash and it hadnt
saved automatically that could be a devastating loss of time.
I appreciate you response greatly since I know this isnt an issue of the
application.
Thanks,
Dean.
```
Original issue reported on code.google.com by `sa...@phoenx.co.uk` on 4 Feb 2012 at 7:14 | 1.0 | This is a query not an issue: regarding saving position. - ```
Good morning and please pardon my ignorance.
My query is regarding the saving of the session.
Does Reaver automatically save position as it running or only when you stop the
app and it says it has saved the session?
My main reason of question is that the router I am testing at the moment
(Provided by Sky broadband UK) allows between 20 and 24 (unsure why it changes)
attempts and then will always lock out for 1 hour.
So my estimated maximum calculation going based on 20 keys per hour to reach
11000 keys is 500 hours which is about 21 days to crack this password.
Now that is a very long time! and if my computer was to crash and it hadnt
saved automatically that could be a devastating loss of time.
I appreciate you response greatly since I know this isnt an issue of the
application.
Thanks,
Dean.
```
Original issue reported on code.google.com by `sa...@phoenx.co.uk` on 4 Feb 2012 at 7:14 | non_infrastructure | this is a query not an issue regarding saving position good morning and please pardon my ignorance my query is regarding the saving of the session does reaver automatically save position as it running or only when you stop the app and it says it has saved the session my main reason of question is that the router i am testing at the moment provided by sky broadband uk allows between and unsure why it changes attempts and then will always lock out for hour so my estimated maximum calculation going based on keys per hour to reach keys is hours which is about days to crack this password now that is a very long time and if my computer was to crash and it hadnt saved automatically that could be a devastating loss of time i appreciate you response greatly since i know this isnt an issue of the application thanks dean original issue reported on code google com by sa phoenx co uk on feb at | 0 |
10,881 | 2,622,512,094 | IssuesEvent | 2015-03-04 03:39:30 | tswast/pywiiuse | https://api.github.com/repos/tswast/pywiiuse | closed | example.py does not recogonize which button I pressed on wii | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. I just connect my wii and ran the program
What is the expected output? What do you see instead?
example.py should tell me which button I pressed, but it does not. It only
knows that and event has happened, but does not know which button I pressed.
What version of the product are you using? On what operating system?
I'm using windows 7 64 bits. But I installed 32 bits python 2.5
Thanks.
```
Original issue reported on code.google.com by `yangyh0...@gmail.com` on 14 Jun 2011 at 8:33 | 1.0 | example.py does not recogonize which button I pressed on wii - ```
What steps will reproduce the problem?
1. I just connect my wii and ran the program
What is the expected output? What do you see instead?
example.py should tell me which button I pressed, but it does not. It only
knows that and event has happened, but does not know which button I pressed.
What version of the product are you using? On what operating system?
I'm using windows 7 64 bits. But I installed 32 bits python 2.5
Thanks.
```
Original issue reported on code.google.com by `yangyh0...@gmail.com` on 14 Jun 2011 at 8:33 | non_infrastructure | example py does not recogonize which button i pressed on wii what steps will reproduce the problem i just connect my wii and ran the program what is the expected output what do you see instead example py should tell me which button i pressed but it does not it only knows that and event has happened but does not know which button i pressed what version of the product are you using on what operating system i m using windows bits but i installed bits python thanks original issue reported on code google com by gmail com on jun at | 0 |
490,219 | 14,116,925,258 | IssuesEvent | 2020-11-08 06:23:12 | TerryCavanagh/diceydungeons.com | https://api.github.com/repos/TerryCavanagh/diceydungeons.com | closed | Cornelius could be optimised | 2 - Has Negative Effects B - Very likely/most players will encounter Priority for v1.11 reported in v0.14 | As soon as you get 6 or more dice, the AI starts to really struggle, and you can notice it taking time to think. Cornelius in particular could be massively optimised, because they're just trying to do a very simple thing - dump the highest values they can on Nightmare.
Create an alternate AI mode that simplifies these decisions to speed up the game. | 1.0 | Cornelius could be optimised - As soon as you get 6 or more dice, the AI starts to really struggle, and you can notice it taking time to think. Cornelius in particular could be massively optimised, because they're just trying to do a very simple thing - dump the highest values they can on Nightmare.
Create an alternate AI mode that simplifies these decisions to speed up the game. | non_infrastructure | cornelius could be optimised as soon as you get or more dice the ai starts to really struggle and you can notice it taking time to think cornelius in particular could be massively optimised because they re just trying to do a very simple thing dump the highest values they can on nightmare create an alternate ai mode that simplifies these decisions to speed up the game | 0 |
35,535 | 2,790,049,077 | IssuesEvent | 2015-05-08 23:36:11 | unt-libraries/django-name | https://api.github.com/repos/unt-libraries/django-name | opened | Remove the /media/ URL | enhancement priority low | Media is setup to be served in `DEBUG`. (See [here](https://github.com/unt-libraries/django-name/blob/master/name/urls.py#L37-L43))
The app doesn't have any media, so we don't need it! :smile: | 1.0 | Remove the /media/ URL - Media is setup to be served in `DEBUG`. (See [here](https://github.com/unt-libraries/django-name/blob/master/name/urls.py#L37-L43))
The app doesn't have any media, so we don't need it! :smile: | non_infrastructure | remove the media url media is setup to be served in debug see the app doesn t have any media so we don t need it smile | 0 |
9,055 | 27,437,186,253 | IssuesEvent | 2023-03-02 08:26:28 | Azure/azure-sdk-tools | https://api.github.com/repos/Azure/azure-sdk-tools | closed | SDK Automation filter bug: should skip all sdk jobs if there is no valid sdk automation jobs | SDK Automation | Example: https://github.com/Azure/azure-rest-api-specs-pr/pull/8964 | 1.0 | SDK Automation filter bug: should skip all sdk jobs if there is no valid sdk automation jobs - Example: https://github.com/Azure/azure-rest-api-specs-pr/pull/8964 | non_infrastructure | sdk automation filter bug should skip all sdk jobs if there is no valid sdk automation jobs example | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.