Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
61,375
| 17,023,679,167
|
IssuesEvent
|
2021-07-03 03:15:43
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
[landcover] Car park area does not render on top of man_made=pier
|
Component: mapnik Priority: major Resolution: duplicate Type: defect
|
**[Submitted to the original trac issue database at 8.33pm, Sunday, 6th February 2011]**
If you look at this car park [1] it has not rendered with a yellow area. I imagine this is because of the man_made=pier that it is on top of.
See also [2]
[1] http://www.openstreetmap.org/browse/way/10519366
[2] http://help.openstreetmap.org/questions/2733/why-are-these-car-parks-refusing-to-render
|
1.0
|
[landcover] Car park area does not render on top of man_made=pier - **[Submitted to the original trac issue database at 8.33pm, Sunday, 6th February 2011]**
If you look at this car park [1] it has not rendered with a yellow area. I imagine this is because of the man_made=pier that it is on top of.
See also [2]
[1] http://www.openstreetmap.org/browse/way/10519366
[2] http://help.openstreetmap.org/questions/2733/why-are-these-car-parks-refusing-to-render
|
defect
|
car park area does not render on top of man made pier if you look at this car park it has not rendered with a yellow area i imagine this is because of the man made pier that it is on top of see also
| 1
|
181,332
| 14,860,165,997
|
IssuesEvent
|
2021-01-18 19:57:16
|
wando-advanced-robotics/infinite-recharge-robot
|
https://api.github.com/repos/wando-advanced-robotics/infinite-recharge-robot
|
opened
|
Enumerate minimum auto modes
|
documentation task
|
How many different auto modes do we think we'll need initially? How many for competition? Let's document those here.
|
1.0
|
Enumerate minimum auto modes - How many different auto modes do we think we'll need initially? How many for competition? Let's document those here.
|
non_defect
|
enumerate minimum auto modes how many different auto modes do we think we ll need initially how many for competition let s document those here
| 0
|
73,422
| 24,619,736,479
|
IssuesEvent
|
2022-10-15 19:21:42
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG:
|
defect
|
### Describe your issue.
cython commit https://github.com/cython/cython/commit/77918c57b5f36bee708f625e2499bf05b23a87f1 in https://github.com/cython/cython/pull/4670 broke scipy installation. This is currently only on cython's master branch and not in an alpha release yet.
I'm not sure if this is best fixed on the scipy or cython side.
I reported this via back-channels to @rgommers.
### Reproducing Code Example
```python
pip install -v --no-build-isolation .
```
with the development version of cython installed.
```
### Error message
```shell
The relevant error from compilation is :
Error compiling Cython file:
------------------------------------------------------------
...
if name == NULL:
name_copy = name
else:
name_copy = strdup(name)
capsule = PyCapsule_New(func, name_copy, &raw_capsule_destructor)
^
------------------------------------------------------------
/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx:80:45: Cannot assign type 'void (*)(object) except *' to 'PyCapsule_Destructor'
Traceback (most recent call last):
File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 28, in <module>
main()
File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 20, in main
sbp.run(['cython', '-3', '--fast-fail',
File "/home/tcaswell/.pybuild/bleeding/lib/python3.12/subprocess.py", line 571, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['cython', '-3', '--fast-fail', '--output-file', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build/scipy/_lib/_ccallback_c.cpython-312-x86_64-linux-gnu.so.p/_ccallback_c.c', '--include-dir', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build', '/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx']' returned non-zero exit status 1.
```
```
### SciPy/NumPy/Python version information
numpy main, cpython main, scipy main
|
1.0
|
BUG: - ### Describe your issue.
cython commit https://github.com/cython/cython/commit/77918c57b5f36bee708f625e2499bf05b23a87f1 in https://github.com/cython/cython/pull/4670 broke scipy installation. This is currently only on cython's master branch and not in an alpha release yet.
I'm not sure if this is best fixed on the scipy or cython side.
I reported this via back-channels to @rgommers.
### Reproducing Code Example
```python
pip install -v --no-build-isolation .
```
with the development version of cython installed.
```
### Error message
```shell
The relevant error from compilation is :
Error compiling Cython file:
------------------------------------------------------------
...
if name == NULL:
name_copy = name
else:
name_copy = strdup(name)
capsule = PyCapsule_New(func, name_copy, &raw_capsule_destructor)
^
------------------------------------------------------------
/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx:80:45: Cannot assign type 'void (*)(object) except *' to 'PyCapsule_Destructor'
Traceback (most recent call last):
File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 28, in <module>
main()
File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 20, in main
sbp.run(['cython', '-3', '--fast-fail',
File "/home/tcaswell/.pybuild/bleeding/lib/python3.12/subprocess.py", line 571, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['cython', '-3', '--fast-fail', '--output-file', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build/scipy/_lib/_ccallback_c.cpython-312-x86_64-linux-gnu.so.p/_ccallback_c.c', '--include-dir', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build', '/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx']' returned non-zero exit status 1.
```
```
### SciPy/NumPy/Python version information
numpy main, cpython main, scipy main
|
defect
|
bug describe your issue cython commit in broke scipy installation this is currently only on cython s master branch and not in an alpha release yet i m not sure if this is best fixed on the scipy or cython side i reported this via back channels to rgommers reproducing code example python pip install v no build isolation with the development version of cython installed error message shell the relevant error from compilation is error compiling cython file if name null name copy name else name copy strdup name capsule pycapsule new func name copy raw capsule destructor home tcaswell source p scipy scipy scipy lib ccallback c pyx cannot assign type void object except to pycapsule destructor traceback most recent call last file home tcaswell source p scipy scipy scipy build utils cythoner py line in main file home tcaswell source p scipy scipy scipy build utils cythoner py line in main sbp run cython fast fail file home tcaswell pybuild bleeding lib subprocess py line in run raise calledprocesserror retcode process args subprocess calledprocesserror command returned non zero exit status scipy numpy python version information numpy main cpython main scipy main
| 1
|
20,201
| 3,314,991,510
|
IssuesEvent
|
2015-11-06 09:31:29
|
OpenMS/OpenMS
|
https://api.github.com/repos/OpenMS/OpenMS
|
closed
|
PeakPicker algorithms don't use Factory pattern [194]
|
defect Good Ideas for upcoming releases minor OpenMS - library
|
Submitted by hendrikweisser on 2010-09-25 17:04:34
In contrast to all (?) other tools offering different algorithms (FeatureFinder, MapAligner, FeatureLinker etc.), the PeakPicker does not use the Factory design pattern.
|
1.0
|
PeakPicker algorithms don't use Factory pattern [194] - Submitted by hendrikweisser on 2010-09-25 17:04:34
In contrast to all (?) other tools offering different algorithms (FeatureFinder, MapAligner, FeatureLinker etc.), the PeakPicker does not use the Factory design pattern.
|
defect
|
peakpicker algorithms don t use factory pattern submitted by hendrikweisser on in contrast to all other tools offering different algorithms featurefinder mapaligner featurelinker etc the peakpicker does not use the factory design pattern
| 1
|
49,799
| 13,187,273,843
|
IssuesEvent
|
2020-08-13 02:53:40
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[hdfwriter] gets fatal RunTimeError, file not properly closed, message: "This is a BUG" (Trac #2104)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2104">https://code.icecube.wisc.edu/ticket/2104</a>, reported by lisa and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-27T16:30:49",
"description": "I'm not sure if this is problem of the hdfwriter of the processing before. I didn't get any errors when I produced the i3 files. I can open them without problem and use them as i3 files without issues, but when I try to convert them to hdf5 I get this error.\n\nHere is the complete traceback:\n\nTraceback (most recent call last):\n File \"s0040-m_divide_data_years_in_LE_HE_low_high_samples_i3_to_hdf5_s0233.py\", line 128, in <module>\n tray.Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 209, in Physics\n return self.handle_frame(frame)\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 205, in handle_frame\n self.writer.convert(frame)\nRuntimeError: trying to get the address of unknown field OFUFilter_14 (in T* I3TableRow::GetPointer(const std::string&, size_t) [with T = bool])\nFATAL (I3TableService): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (I3TableService.cxx:219 in virtual I3TableService::~I3TableService())\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (in virtual I3TableService::~I3TableService())\n",
"reporter": "lisa",
"cc": "",
"resolution": "fixed",
"_ts": "1553704249811581",
"component": "combo reconstruction",
"summary": "[hdfwriter] gets fatal RunTimeError, file not properly closed, message: \"This is a BUG\"",
"priority": "critical",
"keywords": "",
"time": "2017-10-22T10:57:04",
"milestone": "Vernal Equinox 2019",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[hdfwriter] gets fatal RunTimeError, file not properly closed, message: "This is a BUG" (Trac #2104) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2104">https://code.icecube.wisc.edu/ticket/2104</a>, reported by lisa and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-27T16:30:49",
"description": "I'm not sure if this is problem of the hdfwriter of the processing before. I didn't get any errors when I produced the i3 files. I can open them without problem and use them as i3 files without issues, but when I try to convert them to hdf5 I get this error.\n\nHere is the complete traceback:\n\nTraceback (most recent call last):\n File \"s0040-m_divide_data_years_in_LE_HE_low_high_samples_i3_to_hdf5_s0233.py\", line 128, in <module>\n tray.Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 209, in Physics\n return self.handle_frame(frame)\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 205, in handle_frame\n self.writer.convert(frame)\nRuntimeError: trying to get the address of unknown field OFUFilter_14 (in T* I3TableRow::GetPointer(const std::string&, size_t) [with T = bool])\nFATAL (I3TableService): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (I3TableService.cxx:219 in virtual I3TableService::~I3TableService())\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (in virtual I3TableService::~I3TableService())\n",
"reporter": "lisa",
"cc": "",
"resolution": "fixed",
"_ts": "1553704249811581",
"component": "combo reconstruction",
"summary": "[hdfwriter] gets fatal RunTimeError, file not properly closed, message: \"This is a BUG\"",
"priority": "critical",
"keywords": "",
"time": "2017-10-22T10:57:04",
"milestone": "Vernal Equinox 2019",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
gets fatal runtimeerror file not properly closed message this is a bug trac migrated from json status closed changetime description i m not sure if this is problem of the hdfwriter of the processing before i didn t get any errors when i produced the files i can open them without problem and use them as files without issues but when i try to convert them to i get this error n nhere is the complete traceback n ntraceback most recent call last n file m divide data years in le he low high samples to py line in n tray execute n file home lunger data meta projects processing release trunk build lib py line in execute n super self execute n file home lunger data meta projects processing release trunk build lib icecube tableio py line in physics n return self handle frame frame n file home lunger data meta projects processing release trunk build lib icecube tableio py line in handle frame n self writer convert frame nruntimeerror trying to get the address of unknown field ofufilter in t getpointer const std string size t nfatal table atmcscdenergyreco is still connected which means that the output file was never properly closed this is a bug n cxx in virtual nterminate called after throwing an instance of std runtime error n what table atmcscdenergyreco is still connected which means that the output file was never properly closed this is a bug n in virtual n reporter lisa cc resolution fixed ts component combo reconstruction summary gets fatal runtimeerror file not properly closed message this is a bug priority critical keywords time milestone vernal equinox owner jvansanten type defect
| 1
|
57,641
| 15,889,780,805
|
IssuesEvent
|
2021-04-10 12:58:07
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
Highlight new users who join the voice/web
|
design: user experience module: client priority: normal type: defect
|
Originally reported on Google Code with ID 757
```
What steps will reproduce the problem?
1. Have 20 users in the Listener's window
2. A new user joins
3.
It's really hard to see who was the 21st user who joined.
If BigBlueButton could highlight the new user for a few moments, such as a different
color that fades after a few moments, it would be much easier to pick out the new users
in the list.
Same goes for the Users window.
```
Reported by `ffdixon` on 2010-11-23 23:47:42
|
1.0
|
Highlight new users who join the voice/web - Originally reported on Google Code with ID 757
```
What steps will reproduce the problem?
1. Have 20 users in the Listener's window
2. A new user joins
3.
It's really hard to see who was the 21st user who joined.
If BigBlueButton could highlight the new user for a few moments, such as a different
color that fades after a few moments, it would be much easier to pick out the new users
in the list.
Same goes for the Users window.
```
Reported by `ffdixon` on 2010-11-23 23:47:42
|
defect
|
highlight new users who join the voice web originally reported on google code with id what steps will reproduce the problem have users in the listener s window a new user joins it s really hard to see who was the user who joined if bigbluebutton could highlight the new user for a few moments such as a different color that fades after a few moments it would be much easier to pick out the new users in the list same goes for the users window reported by ffdixon on
| 1
|
437,309
| 12,576,527,270
|
IssuesEvent
|
2020-06-09 08:02:56
|
kubeflow/manifests
|
https://api.github.com/repos/kubeflow/manifests
|
closed
|
Jupyter web app config not compatible with latest jupyter web app image
|
area/jupyter kind/bug lifecycle/stale priority/p2
|
Got error when click `New Server`
`KeyError: 'notebook-name'`
`notebook-name` is not in client side request.
Related:
https://github.com/kubeflow/manifests/pull/163
|
1.0
|
Jupyter web app config not compatible with latest jupyter web app image - Got error when click `New Server`
`KeyError: 'notebook-name'`
`notebook-name` is not in client side request.
Related:
https://github.com/kubeflow/manifests/pull/163
|
non_defect
|
jupyter web app config not compatible with latest jupyter web app image got error when click new server keyerror notebook name notebook name is not in client side request related
| 0
|
10,067
| 2,618,934,329
|
IssuesEvent
|
2015-03-03 00:01:28
|
chrsmith/open-ig
|
https://api.github.com/repos/chrsmith/open-ig
|
closed
|
Impossible to create new fleet
|
auto-migrated Priority-Medium Type-Defect
|
```
Game version:
0.95.130
Operating System: (e.g., Windows 7 x86, Windows XP 64-bit)
Fedora 19 x86_64, but I suspect this is unrelated to the operating system.
Java runtime version: (run java -version)
1.6 and 1.7 openJDK, but I suspect this has nothing to do with the JVM
Installed using the Launcher? (yes, no)
Yes
Game language (en, hu, de):
en (I suspect all)
What steps will reproduce the problem?
1. Install the game
2. Run the game
3. Try to find some way to create a fleet without an already existing fleet
4. If 3 fails, give up on the game.
What is the expected output? What do you see instead?
There should be some -easy to find- button which will allow me to create a
fleet. No such button is present
Please provide any additional information below.
Please upload any save before and/or after the problem happened. Please
attach the open-ig.log file found in the
application's directory.
```
Original issue reported on code.google.com by `frederik...@gmail.com` on 24 Jul 2013 at 7:40
|
1.0
|
Impossible to create new fleet - ```
Game version:
0.95.130
Operating System: (e.g., Windows 7 x86, Windows XP 64-bit)
Fedora 19 x86_64, but I suspect this is unrelated to the operating system.
Java runtime version: (run java -version)
1.6 and 1.7 openJDK, but I suspect this has nothing to do with the JVM
Installed using the Launcher? (yes, no)
Yes
Game language (en, hu, de):
en (I suspect all)
What steps will reproduce the problem?
1. Install the game
2. Run the game
3. Try to find some way to create a fleet without an already existing fleet
4. If 3 fails, give up on the game.
What is the expected output? What do you see instead?
There should be some -easy to find- button which will allow me to create a
fleet. No such button is present
Please provide any additional information below.
Please upload any save before and/or after the problem happened. Please
attach the open-ig.log file found in the
application's directory.
```
Original issue reported on code.google.com by `frederik...@gmail.com` on 24 Jul 2013 at 7:40
|
defect
|
impossible to create new fleet game version operating system e g windows windows xp bit fedora but i suspect this is unrelated to the operating system java runtime version run java version and openjdk but i suspect this has nothing to do with the jvm installed using the launcher yes no yes game language en hu de en i suspect all what steps will reproduce the problem install the game run the game try to find some way to create a fleet without an already existing fleet if fails give up on the game what is the expected output what do you see instead there should be some easy to find button which will allow me to create a fleet no such button is present please provide any additional information below please upload any save before and or after the problem happened please attach the open ig log file found in the application s directory original issue reported on code google com by frederik gmail com on jul at
| 1
|
363,893
| 25,472,705,770
|
IssuesEvent
|
2022-11-25 11:37:26
|
UnBArqDsw2022-2/2022.2_G4_IDotPet
|
https://api.github.com/repos/UnBArqDsw2022-2/2022.2_G4_IDotPet
|
opened
|
Correção do Léxico
|
documentation
|
<!-- Certifique-se de ser uma tarefa bem contida. Certifique-se da possibilidade de desmembrar a issue em issues melhores e, caso seja possível, o faça -->
# Descrição:
<!-- Dê os objetivos da issue, para quê serve, qual issue afeta, que artefato é criado/alterado. Ao citar outras issues, utilize o código delas: #Número_da_issue -->
Essa issue tem como objetivo corrigir o Léxico conforme recomendações da apresentação base.
# Tarefas:
<!-- Não economize na listagem de tarefas -->
- [ ] Revisar e corrigir Léxico (Nicolas e Herick)
# Critério de aceitação:
<!-- O revisor da issue deve marcar esses critérios antes de fechá-la -->
- [ ] Documento de Léxico revisado e corrigido.
|
1.0
|
Correção do Léxico - <!-- Certifique-se de ser uma tarefa bem contida. Certifique-se da possibilidade de desmembrar a issue em issues melhores e, caso seja possível, o faça -->
# Descrição:
<!-- Dê os objetivos da issue, para quê serve, qual issue afeta, que artefato é criado/alterado. Ao citar outras issues, utilize o código delas: #Número_da_issue -->
Essa issue tem como objetivo corrigir o Léxico conforme recomendações da apresentação base.
# Tarefas:
<!-- Não economize na listagem de tarefas -->
- [ ] Revisar e corrigir Léxico (Nicolas e Herick)
# Critério de aceitação:
<!-- O revisor da issue deve marcar esses critérios antes de fechá-la -->
- [ ] Documento de Léxico revisado e corrigido.
|
non_defect
|
correção do léxico descrição essa issue tem como objetivo corrigir o léxico conforme recomendações da apresentação base tarefas revisar e corrigir léxico nicolas e herick critério de aceitação documento de léxico revisado e corrigido
| 0
|
653,463
| 21,582,845,927
|
IssuesEvent
|
2022-05-02 20:44:25
|
RAF-SI-2021/Banka-Back
|
https://api.github.com/repos/RAF-SI-2021/Banka-Back
|
closed
|
Promena naziva klasa, metoda i atributa da koriste camelCase
|
area/backend priority/critical
|
Java generalno koristi camelCase za nazive klasa, metoda, atributa. Mi trenutno koristimo snake_case, što prouzrokuje dosta problema pogotova sa Hibernate-om. Potrebno je promeniti sve nazive u camelCase (npr. `hartija_od_vrednosti_id` u `hartijaOdVrednostiId`).
|
1.0
|
Promena naziva klasa, metoda i atributa da koriste camelCase - Java generalno koristi camelCase za nazive klasa, metoda, atributa. Mi trenutno koristimo snake_case, što prouzrokuje dosta problema pogotova sa Hibernate-om. Potrebno je promeniti sve nazive u camelCase (npr. `hartija_od_vrednosti_id` u `hartijaOdVrednostiId`).
|
non_defect
|
promena naziva klasa metoda i atributa da koriste camelcase java generalno koristi camelcase za nazive klasa metoda atributa mi trenutno koristimo snake case što prouzrokuje dosta problema pogotova sa hibernate om potrebno je promeniti sve nazive u camelcase npr hartija od vrednosti id u hartijaodvrednostiid
| 0
|
27,885
| 5,117,890,683
|
IssuesEvent
|
2017-01-07 21:46:35
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
restart self video window again after closing fails
|
Accepted Defect Low Priority Usability Video
|
Originally reported on Google Code with ID 1554
```
closing self video, (use X on video window,) and immediately trying to start video
again, will cause the video button to change icon but video will not start, and button
functionality stops.
```
Reported by `yossia6` on 2013-06-12 07:09:21
|
1.0
|
restart self video window again after closing fails - Originally reported on Google Code with ID 1554
```
closing self video, (use X on video window,) and immediately trying to start video
again, will cause the video button to change icon but video will not start, and button
functionality stops.
```
Reported by `yossia6` on 2013-06-12 07:09:21
|
defect
|
restart self video window again after closing fails originally reported on google code with id closing self video use x on video window and immediately trying to start video again will cause the video button to change icon but video will not start and button functionality stops reported by on
| 1
|
2,272
| 2,603,992,125
|
IssuesEvent
|
2015-02-24 19:06:53
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳好的疱疹专科医院
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:36
|
1.0
|
沈阳好的疱疹专科医院 - ```
沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:36
|
defect
|
沈阳好的疱疹专科医院 沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
| 1
|
30,990
| 6,388,361,896
|
IssuesEvent
|
2017-08-03 15:25:41
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
closed
|
Firefox: synthetic keyboard events do not set KeyboardEvent.key property properly
|
D-firefox E-less easy I-defect
|
Selenium 2.45.0 uses synthetic keyboard events for Firefox 36.0.4 on all(?) platforms. These synthetic keyboard events set the javascript [KeyboardEvent.key](https://developer.mozilla.org/en-US/docs/Web/API/KeyboardEvent/key) property always to 'Unidentified' instead of the expected character.
Using native events with Firefox <= 34.0.5 on Windows 7 the following phpunit test run successfully:
``` php
<?php
class KeyTest extends PHPUnit_Extensions_Selenium2TestCase
{
protected function setUp()
{
$this->setBrowser('firefox');
$this->setHost('localhost');
$this->setBrowserUrl('http://unixpapa.com/js/testkey.html');
}
public function testKeys()
{
$this->url('http://unixpapa.com/js/testkey.html');
$this->byName('classic')->click();
$this->byName('dom3')->click();
$this->keys('z');
$this->assertContains('key=z', $this->byTag('textarea')->value());
}
}
```
As far as I can see, it's related to the use of `sendKeyEvent` in favour to `sendNativeKeyEvent` in https://github.com/SeleniumHQ/selenium/blob/cf26e94543b7c838996a21995fcb8bbe11a800bd/javascript/firefox-driver/js/utils.js#L671.
While testing `sendNativeKeyEvent` in Firefox's Scratchpad I had mixed results:
- Windows: working
- Linux: Exception
- MacOS: obviously working (key is send to scratchpad window instead of webpage)
|
1.0
|
Firefox: synthetic keyboard events do not set KeyboardEvent.key property properly - Selenium 2.45.0 uses synthetic keyboard events for Firefox 36.0.4 on all(?) platforms. These synthetic keyboard events set the javascript [KeyboardEvent.key](https://developer.mozilla.org/en-US/docs/Web/API/KeyboardEvent/key) property always to 'Unidentified' instead of the expected character.
Using native events with Firefox <= 34.0.5 on Windows 7 the following phpunit test run successfully:
``` php
<?php
class KeyTest extends PHPUnit_Extensions_Selenium2TestCase
{
protected function setUp()
{
$this->setBrowser('firefox');
$this->setHost('localhost');
$this->setBrowserUrl('http://unixpapa.com/js/testkey.html');
}
public function testKeys()
{
$this->url('http://unixpapa.com/js/testkey.html');
$this->byName('classic')->click();
$this->byName('dom3')->click();
$this->keys('z');
$this->assertContains('key=z', $this->byTag('textarea')->value());
}
}
```
As far as I can see, it's related to the use of `sendKeyEvent` in favour to `sendNativeKeyEvent` in https://github.com/SeleniumHQ/selenium/blob/cf26e94543b7c838996a21995fcb8bbe11a800bd/javascript/firefox-driver/js/utils.js#L671.
While testing `sendNativeKeyEvent` in Firefox's Scratchpad I had mixed results:
- Windows: working
- Linux: Exception
- MacOS: obviously working (key is send to scratchpad window instead of webpage)
|
defect
|
firefox synthetic keyboard events do not set keyboardevent key property properly selenium uses synthetic keyboard events for firefox on all platforms these synthetic keyboard events set the javascript property always to unidentified instead of the expected character using native events with firefox on windows the following phpunit test run successfully php php class keytest extends phpunit extensions protected function setup this setbrowser firefox this sethost localhost this setbrowserurl public function testkeys this url this byname classic click this byname click this keys z this assertcontains key z this bytag textarea value as far as i can see it s related to the use of sendkeyevent in favour to sendnativekeyevent in while testing sendnativekeyevent in firefox s scratchpad i had mixed results windows working linux exception macos obviously working key is send to scratchpad window instead of webpage
| 1
|
225,707
| 24,881,192,684
|
IssuesEvent
|
2022-10-28 01:21:39
|
tomdgl397/juice-shop
|
https://api.github.com/repos/tomdgl397/juice-shop
|
opened
|
CVE-2022-40304 (High) detected in reactos0.4.13-dev
|
security vulnerability
|
## CVE-2022-40304 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>reactos0.4.13-dev</b></p></summary>
<p>
<p>A free Windows-compatible Operating System</p>
<p>Library home page: <a href=https://github.com/reactos/reactos.git>https://github.com/reactos/reactos.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/libxmljs2/vendor/libxml/entities.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-40304 libxml2: dict corruption caused by entity reference cycles
<p>Publish Date: 2022-09-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40304>CVE-2022-40304</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-09</p>
<p>Fix Resolution: v2.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-40304 (High) detected in reactos0.4.13-dev - ## CVE-2022-40304 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>reactos0.4.13-dev</b></p></summary>
<p>
<p>A free Windows-compatible Operating System</p>
<p>Library home page: <a href=https://github.com/reactos/reactos.git>https://github.com/reactos/reactos.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/libxmljs2/vendor/libxml/entities.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-40304 libxml2: dict corruption caused by entity reference cycles
<p>Publish Date: 2022-09-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40304>CVE-2022-40304</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-09</p>
<p>Fix Resolution: v2.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in dev cve high severity vulnerability vulnerable library dev a free windows compatible operating system library home page a href found in head commit a href vulnerable source files node modules vendor libxml entities c vulnerability details cve dict corruption caused by entity reference cycles publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend
| 0
|
58,381
| 16,515,154,407
|
IssuesEvent
|
2021-05-26 09:06:59
|
galasa-dev/projectmanagement
|
https://api.github.com/repos/galasa-dev/projectmanagement
|
opened
|
Find a way to have the pom name/derscription set correctly using gradle build convention plugins
|
Framework defect
|
For the 16 release, needs to change the gradle builds so that the name and description is set correctly for every project, even when using the new build convention plugions.
Assigning to James so that he learns more about the gradle plugins and the build convention method.
|
1.0
|
Find a way to have the pom name/derscription set correctly using gradle build convention plugins - For the 16 release, needs to change the gradle builds so that the name and description is set correctly for every project, even when using the new build convention plugions.
Assigning to James so that he learns more about the gradle plugins and the build convention method.
|
defect
|
find a way to have the pom name derscription set correctly using gradle build convention plugins for the release needs to change the gradle builds so that the name and description is set correctly for every project even when using the new build convention plugions assigning to james so that he learns more about the gradle plugins and the build convention method
| 1
|
74,578
| 25,188,478,689
|
IssuesEvent
|
2022-11-11 20:48:44
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
Design | Profile | Accessibility | Mobile #
|
design 508/Accessibility authenticated-experience profile 508-defect-4
|
## Background
Review design changes for disability rating and full name being added to the main sections of profile. We will need to discuss these changes with the content and IA folks and get their feedback, then make adjustments to the design as necessary.
Also, determine if we should only show this if we know user is a Veteran or always show.
## Tasks
- [ ] Review designs with Content/IA
- [ ] Document feedback
- [ ] Adjust designs as necessary
## Acceptance Criteria
- [ ] Designs for full name/disability rating are reviewed by content and IA
## Associated Tickets
- #48264
|
1.0
|
Design | Profile | Accessibility | Mobile # - ## Background
Review design changes for disability rating and full name being added to the main sections of profile. We will need to discuss these changes with the content and IA folks and get their feedback, then make adjustments to the design as necessary.
Also, determine if we should only show this if we know user is a Veteran or always show.
## Tasks
- [ ] Review designs with Content/IA
- [ ] Document feedback
- [ ] Adjust designs as necessary
## Acceptance Criteria
- [ ] Designs for full name/disability rating are reviewed by content and IA
## Associated Tickets
- #48264
|
defect
|
design profile accessibility mobile background review design changes for disability rating and full name being added to the main sections of profile we will need to discuss these changes with the content and ia folks and get their feedback then make adjustments to the design as necessary also determine if we should only show this if we know user is a veteran or always show tasks review designs with content ia document feedback adjust designs as necessary acceptance criteria designs for full name disability rating are reviewed by content and ia associated tickets
| 1
|
73,932
| 15,286,689,075
|
IssuesEvent
|
2021-02-23 14:58:35
|
Baneeishaque/Android-Common-Utils4
|
https://api.github.com/repos/Baneeishaque/Android-Common-Utils4
|
opened
|
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar
|
security vulnerability
|
## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: Android-Common-Utils4/common4/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.10/4b95f4897fa13f2cd904aee711aeafc0c5295cd8/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-27.1.2.jar (Root Library)
- sdk-common-27.1.2.jar
- sdklib-27.1.2.jar
- httpmime-4.5.6.jar
- httpclient-4.5.6.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils4/commit/e8328a1ad0239aaf16726916e083bb49de9642ec">e8328a1ad0239aaf16726916e083bb49de9642ec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-20</p>
<p>Fix Resolution: commons-codec:commons-codec:1.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar - ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: Android-Common-Utils4/common4/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.10/4b95f4897fa13f2cd904aee711aeafc0c5295cd8/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-27.1.2.jar (Root Library)
- sdk-common-27.1.2.jar
- sdklib-27.1.2.jar
- httpmime-4.5.6.jar
- httpclient-4.5.6.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils4/commit/e8328a1ad0239aaf16726916e083bb49de9642ec">e8328a1ad0239aaf16726916e083bb49de9642ec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-20</p>
<p>Fix Resolution: commons-codec:commons-codec:1.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file android common build gradle path to vulnerable library home wss scanner gradle caches modules files commons codec commons codec commons codec jar dependency hierarchy lint gradle jar root library sdk common jar sdklib jar httpmime jar httpclient jar x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons codec commons codec step up your open source security game with whitesource
| 0
|
53,853
| 13,262,390,222
|
IssuesEvent
|
2020-08-20 21:41:53
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__' (Trac #2197)
|
Migrated from Trac defect infrastructure
|
Hi,
I saw this error while experimenting with the trunk of filterscripts:
```text
Traceback (most recent call last):
File "./IceCube_BaseProc_only_MM_trunk.py", line 309, in <module>
tray.Execute()
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py", line 256, in Execute
super(I3Tray, self).Execute()
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log
record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__
self.threadName = threading.current_thread().name
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread
return _active[_get_ident()]
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log
record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__
self.threadName = threading.current_thread().name
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread
return _active[_get_ident()]
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py", line 26, in MaskMaker
g = omgeo[omkey]
TypeError: 'I3Geometry' object has no attribute '__getitem__'
```
I think the problem is line 26 in
http://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py
and should be fixed by replacing
```
g = omgeo[omkey]
```
with
```text
g = omgeo.omgeo[omkey]
```
(thanks to alex olivias and jvs for pointing this out to me)
Cheers,
Frederik
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2197">https://code.icecube.wisc.edu/projects/icecube/ticket/2197</a>, reported by flauberand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T03:53:01",
"_ts": "1547265181974434",
"description": "Hi,\n\nI saw this error while experimenting with the trunk of filterscripts:\n\n{{{\nTraceback (most recent call last):\n File \"./IceCube_BaseProc_only_MM_trunk.py\", line 309, in <module>\n tray.Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py\", line 26, in MaskMaker\n g = omgeo[omkey]\nTypeError: 'I3Geometry' object has no attribute '__getitem__'\n}}}\n\nI think the problem is line 26 in \nhttp://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py\n\nand should be fixed by replacing\n{{{ \ng = omgeo[omkey]\n}}}\nwith \n{{{\ng = omgeo.omgeo[omkey]\n}}}\n(thanks to alex olivias and jvs for pointing this out to me)\nCheers,\nFrederik\n\n\n",
"reporter": "flauber",
"cc": "",
"resolution": "fixed",
"time": "2018-10-09T14:06:46",
"component": "infrastructure",
"summary": "filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__'",
"priority": "normal",
"keywords": "filterscripts",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__' (Trac #2197) - Hi,
I saw this error while experimenting with the trunk of filterscripts:
```text
Traceback (most recent call last):
File "./IceCube_BaseProc_only_MM_trunk.py", line 309, in <module>
tray.Execute()
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py", line 256, in Execute
super(I3Tray, self).Execute()
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log
record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__
self.threadName = threading.current_thread().name
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread
return _active[_get_ident()]
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log
record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__
self.threadName = threading.current_thread().name
File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread
return _active[_get_ident()]
File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py", line 26, in MaskMaker
g = omgeo[omkey]
TypeError: 'I3Geometry' object has no attribute '__getitem__'
```
I think the problem is line 26 in
http://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py
and should be fixed by replacing
```
g = omgeo[omkey]
```
with
```text
g = omgeo.omgeo[omkey]
```
(thanks to alex olivias and jvs for pointing this out to me)
Cheers,
Frederik
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2197">https://code.icecube.wisc.edu/projects/icecube/ticket/2197</a>, reported by flauberand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T03:53:01",
"_ts": "1547265181974434",
"description": "Hi,\n\nI saw this error while experimenting with the trunk of filterscripts:\n\n{{{\nTraceback (most recent call last):\n File \"./IceCube_BaseProc_only_MM_trunk.py\", line 309, in <module>\n tray.Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py\", line 26, in MaskMaker\n g = omgeo[omkey]\nTypeError: 'I3Geometry' object has no attribute '__getitem__'\n}}}\n\nI think the problem is line 26 in \nhttp://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py\n\nand should be fixed by replacing\n{{{ \ng = omgeo[omkey]\n}}}\nwith \n{{{\ng = omgeo.omgeo[omkey]\n}}}\n(thanks to alex olivias and jvs for pointing this out to me)\nCheers,\nFrederik\n\n\n",
"reporter": "flauber",
"cc": "",
"resolution": "fixed",
"time": "2018-10-09T14:06:46",
"component": "infrastructure",
"summary": "filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__'",
"priority": "normal",
"keywords": "filterscripts",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
filterscripts typeerror object has no attribute getitem trac hi i saw this error while experimenting with the trunk of filterscripts text traceback most recent call last file icecube baseproc only mm trunk py line in tray execute file data user flauber software icerec trunk build lib py line in execute super self execute file data user flauber software icerec trunk build lib icecube icetray py line in log record logging logrecord logger name self pylevels file line msg tuple none none file cvmfs icecube opensciencegrid org rhel lib logging init py line in init self threadname threading current thread name file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread return active file data user flauber software icerec trunk build lib icecube icetray py line in log record logging logrecord logger name self pylevels file line msg tuple none none file cvmfs icecube opensciencegrid org rhel lib logging init py line in init self threadname threading current thread name file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread return active file data user flauber software icerec trunk build lib icecube filterscripts baseproc superdst py line in maskmaker g omgeo typeerror object has no attribute getitem i think the problem is line in and should be fixed by replacing g omgeo with text g omgeo omgeo thanks to alex olivias and jvs for pointing this out to me cheers frederik migrated from json status closed changetime ts description hi n ni saw this error while experimenting with the trunk of filterscripts n n ntraceback most recent call last n file icecube baseproc only mm trunk py line in n tray execute n file data user flauber software icerec trunk build lib py line in execute n super self execute n file data user flauber software icerec trunk build lib icecube icetray py line in log n record logging logrecord logger name self pylevels file line msg tuple none none n file cvmfs icecube opensciencegrid org rhel lib logging init py line in init n self threadname threading current thread name n file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread n return active n file data user flauber software icerec trunk build lib icecube icetray py line in log n record logging logrecord logger name self pylevels file line msg tuple none none n file cvmfs icecube opensciencegrid org rhel lib logging init py line in init n self threadname threading current thread name n file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread n return active n file data user flauber software icerec trunk build lib icecube filterscripts baseproc superdst py line in maskmaker n g omgeo ntypeerror object has no attribute getitem n n ni think the problem is line in n should be fixed by replacing n ng omgeo n nwith n ng omgeo omgeo n n thanks to alex olivias and jvs for pointing this out to me ncheers nfrederik n n n reporter flauber cc resolution fixed time component infrastructure summary filterscripts typeerror object has no attribute getitem priority normal keywords filterscripts milestone owner olivas type defect
| 1
|
56,332
| 15,023,514,986
|
IssuesEvent
|
2021-02-01 18:20:09
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
508-defect-3 [COGNITION]: COVID Vaccine Trials - Checkboxes should not be separated from additional information inputs
|
508-defect-3 508-issue-cognition 508/Accessibility covid-vaccine-trials
|
# [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. -->
**VFS Point of Contact:** _Trevor_
## User Story or Problem Statement
<!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. -->
As a user, I want to make sure all of my answers are recorded. When additional information fields appear away from the question I am answering, I might miss them.
## Details
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The "What is your gender?" question gives users an option to self-describe. If users check this option, an additional text input appears after the next checkbox. The input should come directly after this checkbox. An option might be to put the "Prefer to self-describe" checkbox last, after Prefer not to answer. Screenshot attached below.
## Acceptance Criteria
- [ ] Checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input
## Environment
* https://staging.va.gov/coronavirus-research/volunteer/sign-up
## WCAG or Vendor Guidance (optional)
* [Meaningful Sequence: Understanding SC 1.3.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-sequence.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

|
1.0
|
508-defect-3 [COGNITION]: COVID Vaccine Trials - Checkboxes should not be separated from additional information inputs - # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. -->
**VFS Point of Contact:** _Trevor_
## User Story or Problem Statement
<!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. -->
As a user, I want to make sure all of my answers are recorded. When additional information fields appear away from the question I am answering, I might miss them.
## Details
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The "What is your gender?" question gives users an option to self-describe. If users check this option, an additional text input appears after the next checkbox. The input should come directly after this checkbox. An option might be to put the "Prefer to self-describe" checkbox last, after Prefer not to answer. Screenshot attached below.
## Acceptance Criteria
- [ ] Checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input
## Environment
* https://staging.va.gov/coronavirus-research/volunteer/sign-up
## WCAG or Vendor Guidance (optional)
* [Meaningful Sequence: Understanding SC 1.3.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-sequence.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

|
defect
|
defect covid vaccine trials checkboxes should not be separated from additional information inputs enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact trevor user story or problem statement as a user i want to make sure all of my answers are recorded when additional information fields appear away from the question i am answering i might miss them details the what is your gender question gives users an option to self describe if users check this option an additional text input appears after the next checkbox the input should come directly after this checkbox an option might be to put the prefer to self describe checkbox last after prefer not to answer screenshot attached below acceptance criteria checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input environment wcag or vendor guidance optional screenshots or trace logs
| 1
|
38,722
| 8,527,328,041
|
IssuesEvent
|
2018-11-02 19:07:13
|
ccharlier/orionbanque
|
https://api.github.com/repos/ccharlier/orionbanque
|
closed
|
Fix 1 Maintainability, 6 Style issues in OrionBanque\Classe\Configuration.cs
|
codefactor
|
[CodeFactor](https://www.codefactor.io/repository/github/ccharlier/orionbanque/overview/master) found multiple issues:
#### A C# document may only contain a single class at the root level unless all of the classes are partial and are of the same type.
[OrionBanque\Classe\Configuration.cs:19
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L19)
#### A closing curly bracket must not be preceded by a blank line.
[OrionBanque\Classe\Configuration.cs:287
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L287)
#### The code must not contain multiple blank lines in a row.
[OrionBanque\Classe\Configuration.cs:5
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L5)[OrionBanque\Classe\Configuration.cs:199
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L199)
#### An opening curly bracket must not be followed by a blank line.
[OrionBanque\Classe\Configuration.cs:219
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L219)
#### Invalid spacing around the semicolon.
[OrionBanque\Classe\Configuration.cs:195
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)[OrionBanque\Classe\Configuration.cs:195
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)
|
1.0
|
Fix 1 Maintainability, 6 Style issues in OrionBanque\Classe\Configuration.cs - [CodeFactor](https://www.codefactor.io/repository/github/ccharlier/orionbanque/overview/master) found multiple issues:
#### A C# document may only contain a single class at the root level unless all of the classes are partial and are of the same type.
[OrionBanque\Classe\Configuration.cs:19
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L19)
#### A closing curly bracket must not be preceded by a blank line.
[OrionBanque\Classe\Configuration.cs:287
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L287)
#### The code must not contain multiple blank lines in a row.
[OrionBanque\Classe\Configuration.cs:5
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L5)[OrionBanque\Classe\Configuration.cs:199
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L199)
#### An opening curly bracket must not be followed by a blank line.
[OrionBanque\Classe\Configuration.cs:219
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L219)
#### Invalid spacing around the semicolon.
[OrionBanque\Classe\Configuration.cs:195
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)[OrionBanque\Classe\Configuration.cs:195
](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)
|
non_defect
|
fix maintainability style issues in orionbanque classe configuration cs found multiple issues a c document may only contain a single class at the root level unless all of the classes are partial and are of the same type orionbanque classe configuration cs a closing curly bracket must not be preceded by a blank line orionbanque classe configuration cs the code must not contain multiple blank lines in a row orionbanque classe configuration cs an opening curly bracket must not be followed by a blank line orionbanque classe configuration cs invalid spacing around the semicolon orionbanque classe configuration cs
| 0
|
78,730
| 15,586,068,790
|
IssuesEvent
|
2021-03-18 01:06:01
|
Pr0methean/s5
|
https://api.github.com/repos/Pr0methean/s5
|
opened
|
CVE-2019-3875 (Medium) detected in keycloak-core-4.7.0.Final.jar
|
security vulnerability
|
## CVE-2019-3875 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>keycloak-core-4.7.0.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://keycloak.org/keycloak-core">http://keycloak.org/keycloak-core</a></p>
<p>Path to dependency file: /s5/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/keycloak/keycloak-core/4.7.0.Final/keycloak-core-4.7.0.Final.jar</p>
<p>
Dependency Hierarchy:
- keycloak-spring-boot-starter-4.7.0.Final.jar (Root Library)
- keycloak-spring-boot-2-adapter-4.7.0.Final.jar
- :x: **keycloak-core-4.7.0.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in keycloak before 6.0.2. The X.509 authenticator supports the verification of client certificates through the CRL, where the CRL list can be obtained from the URL provided in the certificate itself (CDP) or through the separately configured path. The CRL are often available over the network through unsecured protocols ('http' or 'ldap') and hence the caller should verify the signature and possibly the certification path. Keycloak currently doesn't validate signatures on CRL, which can result in a possibility of various attacks like man-in-the-middle.
<p>Publish Date: 2019-06-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3875>CVE-2019-3875</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875</a></p>
<p>Release Date: 2019-06-12</p>
<p>Fix Resolution: 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-3875 (Medium) detected in keycloak-core-4.7.0.Final.jar - ## CVE-2019-3875 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>keycloak-core-4.7.0.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://keycloak.org/keycloak-core">http://keycloak.org/keycloak-core</a></p>
<p>Path to dependency file: /s5/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/keycloak/keycloak-core/4.7.0.Final/keycloak-core-4.7.0.Final.jar</p>
<p>
Dependency Hierarchy:
- keycloak-spring-boot-starter-4.7.0.Final.jar (Root Library)
- keycloak-spring-boot-2-adapter-4.7.0.Final.jar
- :x: **keycloak-core-4.7.0.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in keycloak before 6.0.2. The X.509 authenticator supports the verification of client certificates through the CRL, where the CRL list can be obtained from the URL provided in the certificate itself (CDP) or through the separately configured path. The CRL are often available over the network through unsecured protocols ('http' or 'ldap') and hence the caller should verify the signature and possibly the certification path. Keycloak currently doesn't validate signatures on CRL, which can result in a possibility of various attacks like man-in-the-middle.
<p>Publish Date: 2019-06-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3875>CVE-2019-3875</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875</a></p>
<p>Release Date: 2019-06-12</p>
<p>Fix Resolution: 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in keycloak core final jar cve medium severity vulnerability vulnerable library keycloak core final jar library home page a href path to dependency file pom xml path to vulnerable library root repository org keycloak keycloak core final keycloak core final jar dependency hierarchy keycloak spring boot starter final jar root library keycloak spring boot adapter final jar x keycloak core final jar vulnerable library vulnerability details a vulnerability was found in keycloak before the x authenticator supports the verification of client certificates through the crl where the crl list can be obtained from the url provided in the certificate itself cdp or through the separately configured path the crl are often available over the network through unsecured protocols http or ldap and hence the caller should verify the signature and possibly the certification path keycloak currently doesn t validate signatures on crl which can result in a possibility of various attacks like man in the middle publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
775,983
| 27,242,614,973
|
IssuesEvent
|
2023-02-21 21:57:21
|
o3de/o3de.org
|
https://api.github.com/repos/o3de/o3de.org
|
closed
|
Error code index
|
kind/reference kind/feature priority/critical
|
It would be useful to have several different error code indexes - things like editor errors, command-line tool errors, core engine or atom errors - and provide an index that maps them to human-readable information so that it's possible to do a quick lookup online. This also acts as a **great** search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something.
This is also a very good candidate for automation, if there's a way to extract error codes reliably from the O3DE source (such as if they're contained in, say, a filemap somewhere.)
|
1.0
|
Error code index - It would be useful to have several different error code indexes - things like editor errors, command-line tool errors, core engine or atom errors - and provide an index that maps them to human-readable information so that it's possible to do a quick lookup online. This also acts as a **great** search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something.
This is also a very good candidate for automation, if there's a way to extract error codes reliably from the O3DE source (such as if they're contained in, say, a filemap somewhere.)
|
non_defect
|
error code index it would be useful to have several different error code indexes things like editor errors command line tool errors core engine or atom errors and provide an index that maps them to human readable information so that it s possible to do a quick lookup online this also acts as a great search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something this is also a very good candidate for automation if there s a way to extract error codes reliably from the source such as if they re contained in say a filemap somewhere
| 0
|
80,766
| 30,523,453,788
|
IssuesEvent
|
2023-07-19 09:36:47
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
SelectCheckboxMenu: unable to partial search for an option without following value order
|
:lady_beetle: defect :bangbang: needs-triage
|
### Describe the bug
This also doesn't work for the showcase;
When using selectCheckboxMenu with options like "London City", "Leeds City", "Manchester City" "Manchester United" .
when using the search feature of selectcheckboxmenu, I can't just type City and filter the options that way, I currently have to start at the beginning of the value/string like "Manchester" which would show Manchester city and Manchester United.
### Reproducer
can be checked using the primefaces showcase, if you wanted to look for Miami but typed iami you dont get any result
### Expected behavior
the search should filter the list based on any matched letters within the values
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
All
### JSF version
2.1
### Java version
19
### Browser(s)
_No response_
|
1.0
|
SelectCheckboxMenu: unable to partial search for an option without following value order - ### Describe the bug
This also doesn't work for the showcase;
When using selectCheckboxMenu with options like "London City", "Leeds City", "Manchester City" "Manchester United" .
when using the search feature of selectcheckboxmenu, I can't just type City and filter the options that way, I currently have to start at the beginning of the value/string like "Manchester" which would show Manchester city and Manchester United.
### Reproducer
can be checked using the primefaces showcase, if you wanted to look for Miami but typed iami you dont get any result
### Expected behavior
the search should filter the list based on any matched letters within the values
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
All
### JSF version
2.1
### Java version
19
### Browser(s)
_No response_
|
defect
|
selectcheckboxmenu unable to partial search for an option without following value order describe the bug this also doesn t work for the showcase when using selectcheckboxmenu with options like london city leeds city manchester city manchester united when using the search feature of selectcheckboxmenu i can t just type city and filter the options that way i currently have to start at the beginning of the value string like manchester which would show manchester city and manchester united reproducer can be checked using the primefaces showcase if you wanted to look for miami but typed iami you dont get any result expected behavior the search should filter the list based on any matched letters within the values primefaces edition community primefaces version theme no response jsf implementation all jsf version java version browser s no response
| 1
|
38,218
| 8,701,643,514
|
IssuesEvent
|
2018-12-05 12:12:00
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Reflection on MethodHandles$Lookup fails in JDK 9
|
C: Functionality P: Medium T: Defect
|
This is the same issue as encountered in jOOR: https://github.com/jOOQ/jOOR/issues/49. The `DefaultRecordMapper.ProxyMapper` cannot call default methods on such proxies, starting from JDK 9
|
1.0
|
Reflection on MethodHandles$Lookup fails in JDK 9 - This is the same issue as encountered in jOOR: https://github.com/jOOQ/jOOR/issues/49. The `DefaultRecordMapper.ProxyMapper` cannot call default methods on such proxies, starting from JDK 9
|
defect
|
reflection on methodhandles lookup fails in jdk this is the same issue as encountered in joor the defaultrecordmapper proxymapper cannot call default methods on such proxies starting from jdk
| 1
|
315,566
| 23,587,115,878
|
IssuesEvent
|
2022-08-23 12:33:33
|
r-lib/cli
|
https://api.github.com/repos/r-lib/cli
|
closed
|
"Building a Semantic CLI" vignette has misplaced asciicast svgs.
|
documentation
|
For some reason, the code present at the `asciicast` block in [line 207](https://github.com/r-lib/cli/blob/HEAD/vignettes/semantic-cli.Rmd#L207) isn't rendered, and what's rendered instead is the next `asciicast` block. All remaning `asciicast` blocks after this point render before they were supposed to, resulting in a confusing vignette, like in the following screenshot:

I have tried this in multiple browsers, it does not seem to be a local issue.
|
1.0
|
"Building a Semantic CLI" vignette has misplaced asciicast svgs. - For some reason, the code present at the `asciicast` block in [line 207](https://github.com/r-lib/cli/blob/HEAD/vignettes/semantic-cli.Rmd#L207) isn't rendered, and what's rendered instead is the next `asciicast` block. All remaning `asciicast` blocks after this point render before they were supposed to, resulting in a confusing vignette, like in the following screenshot:

I have tried this in multiple browsers, it does not seem to be a local issue.
|
non_defect
|
building a semantic cli vignette has misplaced asciicast svgs for some reason the code present at the asciicast block in isn t rendered and what s rendered instead is the next asciicast block all remaning asciicast blocks after this point render before they were supposed to resulting in a confusing vignette like in the following screenshot i have tried this in multiple browsers it does not seem to be a local issue
| 0
|
40,040
| 9,810,129,174
|
IssuesEvent
|
2019-06-12 19:42:49
|
IBM/CAST
|
https://api.github.com/repos/IBM/CAST
|
closed
|
standalone_ib_and_switch_collection error after upgraded to 1.6.0-2583
|
Comp: CSM.api Comp: CSM.db PhaseFound: IST Sev: 2 Type: Defect
|
**Describe the bug**
Running the command `standalone_ib_and_switch_collection` (1.6.0-2583) produced csmapi and database errors at the end.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to `/opt/ibm/csm/sbin/`
2. Run `./standalone_ib_and_switch_collection`
3. See errors:
```
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
(null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
(null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
```
**Expected behavior**
The command should not display error messages.
**Screenshots**
```
[root@f4n05 sbin]# ./standalone_ib_and_switch_collection
UFM reported 190 IB records.
This report from UFM can be found in 'ufm_ib_cable_output_file.json' located at '/var/log/ibm/csm/inv'
WARNING: 9 IB cables were discovered, but are missing serial numbers and have been removed from CSM inventory collection data.
These records copied into 'bad_ib_cable_records.txt' located at '/var/log/ibm/csm/inv'
UFM reported 9 switch records.
This report from UFM can be found in 'ufm_switch_output_file.json' located at '/var/log/ibm/csm/inv'
WARNING: 2 Switches found with 'N/A' serial numbers and have been removed from CSM inventory collection data.
These records copied into 'bad_switch_records.txt' located at '/var/log/ibm/csm/inv'
---
# ib inventory successful
...
# total ib inventory collected: 181
# new ib records inserted into database: 0
# old ib records updated in database: 181
# old ib records removed from the database: 0
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
(null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
(null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
```
**Environment (please complete the following information):**
- Machine [CSM master: f4n05]
- Version [CSM 1.6.0-2583]
**Additional context**
The command worked fine with 1.6.0-2575 before upgrading to 1.6.0-2583.
**Issue Source:**
IST CSM upgrade on alt. cluster.
|
1.0
|
standalone_ib_and_switch_collection error after upgraded to 1.6.0-2583 - **Describe the bug**
Running the command `standalone_ib_and_switch_collection` (1.6.0-2583) produced csmapi and database errors at the end.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to `/opt/ibm/csm/sbin/`
2. Run `./standalone_ib_and_switch_collection`
3. See errors:
```
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
(null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
(null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
```
**Expected behavior**
The command should not display error messages.
**Screenshots**
```
[root@f4n05 sbin]# ./standalone_ib_and_switch_collection
UFM reported 190 IB records.
This report from UFM can be found in 'ufm_ib_cable_output_file.json' located at '/var/log/ibm/csm/inv'
WARNING: 9 IB cables were discovered, but are missing serial numbers and have been removed from CSM inventory collection data.
These records copied into 'bad_ib_cable_records.txt' located at '/var/log/ibm/csm/inv'
UFM reported 9 switch records.
This report from UFM can be found in 'ufm_switch_output_file.json' located at '/var/log/ibm/csm/inv'
WARNING: 2 Switches found with 'N/A' serial numbers and have been removed from CSM inventory collection data.
These records copied into 'bad_switch_records.txt' located at '/var/log/ibm/csm/inv'
---
# ib inventory successful
...
# total ib inventory collected: 181
# new ib records inserted into database: 0
# old ib records updated in database: 181
# old ib records removed from the database: 0
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
(null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory"
[csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set
[csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
(null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey"
```
**Environment (please complete the following information):**
- Machine [CSM master: f4n05]
- Version [CSM 1.6.0-2583]
**Additional context**
The command worked fine with 1.6.0-2575 before upgrading to 1.6.0-2583.
**Issue Source:**
IST CSM upgrade on alt. cluster.
|
defect
|
standalone ib and switch collection error after upgraded to describe the bug running the command standalone ib and switch collection produced csmapi and database errors at the end to reproduce steps to reproduce the behavior go to opt ibm csm sbin run standalone ib and switch collection see errors home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory null failed errcode errmsg csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey null failed errcode errmsg csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey expected behavior the command should not display error messages screenshots standalone ib and switch collection ufm reported ib records this report from ufm can be found in ufm ib cable output file json located at var log ibm csm inv warning ib cables were discovered but are missing serial numbers and have been removed from csm inventory collection data these records copied into bad ib cable records txt located at var log ibm csm inv ufm reported switch records this report from ufm can be found in ufm switch output file json located at var log ibm csm inv warning switches found with n a serial numbers and have been removed from csm inventory collection data these records copied into bad switch records txt located at var log ibm csm inv ib inventory successful total ib inventory collected new ib records inserted into database old ib records updated in database old ib records removed from the database home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory null failed errcode errmsg csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey null failed errcode errmsg csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey environment please complete the following information machine version additional context the command worked fine with before upgrading to issue source ist csm upgrade on alt cluster
| 1
|
665,001
| 22,295,559,721
|
IssuesEvent
|
2022-06-13 00:47:15
|
idom-team/idom
|
https://api.github.com/repos/idom-team/idom
|
opened
|
Fix Flaky Tests
|
priority: 2 (moderate)
|
### Current Situation
There are a number of tests in the suite that are very flaky (usually requiring 2-3 runs to pass). For new contributors this is likely to be confusing since they'll be unsure if those failures are related to their changes.
### Proposed Actions
It's likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order (or at least that's been my experience in the past). Unfortunately the problems are hard to reproduce since running the local tests rarely, if ever, fails at the rate seen in CI/CD.
|
1.0
|
Fix Flaky Tests - ### Current Situation
There are a number of tests in the suite that are very flaky (usually requiring 2-3 runs to pass). For new contributors this is likely to be confusing since they'll be unsure if those failures are related to their changes.
### Proposed Actions
It's likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order (or at least that's been my experience in the past). Unfortunately the problems are hard to reproduce since running the local tests rarely, if ever, fails at the rate seen in CI/CD.
|
non_defect
|
fix flaky tests current situation there are a number of tests in the suite that are very flaky usually requiring runs to pass for new contributors this is likely to be confusing since they ll be unsure if those failures are related to their changes proposed actions it s likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order or at least that s been my experience in the past unfortunately the problems are hard to reproduce since running the local tests rarely if ever fails at the rate seen in ci cd
| 0
|
63,843
| 18,015,927,521
|
IssuesEvent
|
2021-09-16 13:55:13
|
hazelcast/hazelcast-go-client
|
https://api.github.com/repos/hazelcast/hazelcast-go-client
|
opened
|
User callbacks for events should not block internals of the client
|
Type: Defect Source: Internal
|
Currently since events are subscribed as sync if a user blocks an event, all the client events are blocked.
|
1.0
|
User callbacks for events should not block internals of the client -
Currently since events are subscribed as sync if a user blocks an event, all the client events are blocked.
|
defect
|
user callbacks for events should not block internals of the client currently since events are subscribed as sync if a user blocks an event all the client events are blocked
| 1
|
101,022
| 8,767,448,643
|
IssuesEvent
|
2018-12-17 19:47:30
|
nuxdie/frontendweekend
|
https://api.github.com/repos/nuxdie/frontendweekend
|
closed
|
add dev chat
|
monitoring testing tracking
|
something like slack for dev notifications and web hooks
(maybe send to telegram)
|
1.0
|
add dev chat - something like slack for dev notifications and web hooks
(maybe send to telegram)
|
non_defect
|
add dev chat something like slack for dev notifications and web hooks maybe send to telegram
| 0
|
62,420
| 17,023,920,047
|
IssuesEvent
|
2021-07-03 04:33:38
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
HTTPS support for osm replication updates
|
Component: admin Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 7.54am, Tuesday, 7th April 2015]**
https://planet.openstreetmap.org/replication/day/000/000/
Since the OSM updates diff support only http, I cannot update the OSM database in my production environment due to security issue. It is vulnerable to MITM(Man-In-The-Middle) attack. When I tried https url, it redirects to http url.
Is there any specific reason for not having https support ?
It would be great, if OSM gives support https. And also OSM may add md5check sum details in the xx_state.txt files.
I am ready to take this task and contribute. Can you guide me, where to start ?
|
1.0
|
HTTPS support for osm replication updates - **[Submitted to the original trac issue database at 7.54am, Tuesday, 7th April 2015]**
https://planet.openstreetmap.org/replication/day/000/000/
Since the OSM updates diff support only http, I cannot update the OSM database in my production environment due to security issue. It is vulnerable to MITM(Man-In-The-Middle) attack. When I tried https url, it redirects to http url.
Is there any specific reason for not having https support ?
It would be great, if OSM gives support https. And also OSM may add md5check sum details in the xx_state.txt files.
I am ready to take this task and contribute. Can you guide me, where to start ?
|
defect
|
https support for osm replication updates since the osm updates diff support only http i cannot update the osm database in my production environment due to security issue it is vulnerable to mitm man in the middle attack when i tried https url it redirects to http url is there any specific reason for not having https support it would be great if osm gives support https and also osm may add sum details in the xx state txt files i am ready to take this task and contribute can you guide me where to start
| 1
|
41,122
| 10,313,853,205
|
IssuesEvent
|
2019-08-30 00:45:54
|
techo/voluntariado-eventual
|
https://api.github.com/repos/techo/voluntariado-eventual
|
closed
|
Búsqueda por varios términos en los inscriptos a una actividad
|
Defecto
|
**El pedido está relacionado con un problema, describí**
Cómo Coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones, si se pone más de un término, no trae resultados.
**Describí la solución que te gustaría**
Qué traiga los resultados que coincidan para con el nombre, apellido, email, dni de esos términos
**Describí alternativas que hayas considerado**
**Contexto adicional**
Lo mismo pasaba sobre las tablas de buscar actividad o persona. Se puede copiar de ahí prácticamente.

|
1.0
|
Búsqueda por varios términos en los inscriptos a una actividad - **El pedido está relacionado con un problema, describí**
Cómo Coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones, si se pone más de un término, no trae resultados.
**Describí la solución que te gustaría**
Qué traiga los resultados que coincidan para con el nombre, apellido, email, dni de esos términos
**Describí alternativas que hayas considerado**
**Contexto adicional**
Lo mismo pasaba sobre las tablas de buscar actividad o persona. Se puede copiar de ahí prácticamente.

|
defect
|
búsqueda por varios términos en los inscriptos a una actividad el pedido está relacionado con un problema describí cómo coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones si se pone más de un término no trae resultados describí la solución que te gustaría qué traiga los resultados que coincidan para con el nombre apellido email dni de esos términos describí alternativas que hayas considerado contexto adicional lo mismo pasaba sobre las tablas de buscar actividad o persona se puede copiar de ahí prácticamente
| 1
|
65,365
| 19,431,758,862
|
IssuesEvent
|
2021-12-21 12:50:03
|
catmaid/CATMAID
|
https://api.github.com/repos/catmaid/CATMAID
|
closed
|
Use of localStorage and cookies
|
type: defect status: done
|
Currently some semi-persistent client settings (e.g., selected stack mirror) use either/both of localStorage and cookies. Two issues:
1. Cookies are not appropriate for this purpose, as we've decided this is per-browser-client-specific data (hence not being in a `CATMAID.Settings` persistent store). Additionally the duplication/interference of cookies and localStorage is a needless complication now that all supported browsers have localStorage.
2. Our current use of localStorage does not use instance-specific keys, which can cause collisions on servers with multiple instances.
|
1.0
|
Use of localStorage and cookies - Currently some semi-persistent client settings (e.g., selected stack mirror) use either/both of localStorage and cookies. Two issues:
1. Cookies are not appropriate for this purpose, as we've decided this is per-browser-client-specific data (hence not being in a `CATMAID.Settings` persistent store). Additionally the duplication/interference of cookies and localStorage is a needless complication now that all supported browsers have localStorage.
2. Our current use of localStorage does not use instance-specific keys, which can cause collisions on servers with multiple instances.
|
defect
|
use of localstorage and cookies currently some semi persistent client settings e g selected stack mirror use either both of localstorage and cookies two issues cookies are not appropriate for this purpose as we ve decided this is per browser client specific data hence not being in a catmaid settings persistent store additionally the duplication interference of cookies and localstorage is a needless complication now that all supported browsers have localstorage our current use of localstorage does not use instance specific keys which can cause collisions on servers with multiple instances
| 1
|
58,267
| 16,457,461,514
|
IssuesEvent
|
2021-05-21 14:21:57
|
Questie/Questie
|
https://api.github.com/repos/Questie/Questie
|
opened
|
Game stutters everytime it updates a quest
|
Type - Defect
|
<!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- -->
My game stutters everytime I loot a quest item, kill a quest objective etc. Issue disappears if I disable questie.
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
6.3.11
|
1.0
|
Game stutters everytime it updates a quest - <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- -->
My game stutters everytime I loot a quest item, kill a quest objective etc. Issue disappears if I disable questie.
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
6.3.11
|
defect
|
game stutters everytime it updates a quest read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description my game stutters everytime i loot a quest item kill a quest objective etc issue disappears if i disable questie screenshots questie version
| 1
|
65,393
| 19,479,717,572
|
IssuesEvent
|
2021-12-25 01:45:45
|
unascribed/Fabrication
|
https://api.github.com/repos/unascribed/Fabrication
|
closed
|
Game crashes when i click create new world or load a world when i have extra.encroaching_emeralds enabled
|
k: Defect n: Fabric s: New
|
https://gist.github.com/gitunown/bce59b7db3a389f8881caf1afd3c4bef
|
1.0
|
Game crashes when i click create new world or load a world when i have extra.encroaching_emeralds enabled - https://gist.github.com/gitunown/bce59b7db3a389f8881caf1afd3c4bef
|
defect
|
game crashes when i click create new world or load a world when i have extra encroaching emeralds enabled
| 1
|
355,973
| 10,586,911,304
|
IssuesEvent
|
2019-10-08 20:48:09
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Annotation Problem *.pyi
|
high priority module: typing triaged
|
## 🚀 Feature
It is difficult to control the validity of function arguments when they are in different files.
Very often the information in *.pyi does not match *.py.
https://github.com/pytorch/pytorch/blob/fcd13549f923342a7474993928d86181b75734d3/torch/utils/data/dataloader.pyi#L35
I suggest using Python 2 Annotations:
```python
def optional_unwrap (self, x, y):
# type: (Optional [int], Optional [int]) -> int
```
You can expand this format to indicate in which file to place annotations.
After that, all *.pyi can be completely assembled in one script. Information will always be up to date.
cc @ezyang @gchanan @zou3519
|
1.0
|
Annotation Problem *.pyi - ## 🚀 Feature
It is difficult to control the validity of function arguments when they are in different files.
Very often the information in *.pyi does not match *.py.
https://github.com/pytorch/pytorch/blob/fcd13549f923342a7474993928d86181b75734d3/torch/utils/data/dataloader.pyi#L35
I suggest using Python 2 Annotations:
```python
def optional_unwrap (self, x, y):
# type: (Optional [int], Optional [int]) -> int
```
You can expand this format to indicate in which file to place annotations.
After that, all *.pyi can be completely assembled in one script. Information will always be up to date.
cc @ezyang @gchanan @zou3519
|
non_defect
|
annotation problem pyi 🚀 feature it is difficult to control the validity of function arguments when they are in different files very often the information in pyi does not match py i suggest using python annotations python def optional unwrap self x y type optional optional int you can expand this format to indicate in which file to place annotations after that all pyi can be completely assembled in one script information will always be up to date cc ezyang gchanan
| 0
|
58,443
| 3,089,416,162
|
IssuesEvent
|
2015-08-25 21:23:17
|
phetsims/joist
|
https://api.github.com/repos/phetsims/joist
|
opened
|
Add additional Google Analytics tracker
|
enhancement high-priority Summer 2015 redeploy
|
See https://github.com/phetsims/yotta/issues/12 for details. Will include sending data to another tracker conditionally.
|
1.0
|
Add additional Google Analytics tracker - See https://github.com/phetsims/yotta/issues/12 for details. Will include sending data to another tracker conditionally.
|
non_defect
|
add additional google analytics tracker see for details will include sending data to another tracker conditionally
| 0
|
62,189
| 17,023,869,201
|
IssuesEvent
|
2021-07-03 04:16:40
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
[roads] end of oneway unclear
|
Component: mapnik Priority: minor Resolution: wontfix Type: defect
|
**[Submitted to the original trac issue database at 8.33pm, Monday, 22nd July 2013]**
An [http://www.openstreetmap.org/?lat=50.620504&lon=5.660834&zoom=18&layers=M inhabitant of this place] draws my attention (aren't some people nice?) on the oneway street Jean Leclercq. He says that the map makes it unclear that the oneway restriction stops half way, at the crossing with street Jean Jaurs. He says that there should be an arrow there to show that end.[[BR]]
I'm taking the idea one step further. An additional arrow is a hint and not a certainty. Could it be easily possible, when a oneway stops in the middle of a street, to draw a clear double headed arrow at the beginning of the unrestricted segment?[[BR]]
This is similar to Ticket #4338, but different. They should be considered in parallel. Thanks.
|
1.0
|
[roads] end of oneway unclear - **[Submitted to the original trac issue database at 8.33pm, Monday, 22nd July 2013]**
An [http://www.openstreetmap.org/?lat=50.620504&lon=5.660834&zoom=18&layers=M inhabitant of this place] draws my attention (aren't some people nice?) on the oneway street Jean Leclercq. He says that the map makes it unclear that the oneway restriction stops half way, at the crossing with street Jean Jaurs. He says that there should be an arrow there to show that end.[[BR]]
I'm taking the idea one step further. An additional arrow is a hint and not a certainty. Could it be easily possible, when a oneway stops in the middle of a street, to draw a clear double headed arrow at the beginning of the unrestricted segment?[[BR]]
This is similar to Ticket #4338, but different. They should be considered in parallel. Thanks.
|
defect
|
end of oneway unclear an draws my attention aren t some people nice on the oneway street jean leclercq he says that the map makes it unclear that the oneway restriction stops half way at the crossing with street jean jaurs he says that there should be an arrow there to show that end i m taking the idea one step further an additional arrow is a hint and not a certainty could it be easily possible when a oneway stops in the middle of a street to draw a clear double headed arrow at the beginning of the unrestricted segment this is similar to ticket but different they should be considered in parallel thanks
| 1
|
68,120
| 21,483,720,850
|
IssuesEvent
|
2022-04-26 20:30:09
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
Enabling threads hides the voice call button
|
T-Defect X-Needs-Design Z-UI UX A-VoIP S-Minor O-Occasional
|
### Steps to reproduce
1. Voice button is now hidden behind the hamburger icon
2. Except it’s not labelled “voice call” it’s “call”, which makes it even harder to find
3. Instead, perhaps we need a single call button which then gives you voice or video as a choice
### Outcome
#### What did you expect?
Intuitive calling
#### What happened instead?
I kept getting video calls because folks can no longer find the voice call button.
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Enabling threads hides the voice call button - ### Steps to reproduce
1. Voice button is now hidden behind the hamburger icon
2. Except it’s not labelled “voice call” it’s “call”, which makes it even harder to find
3. Instead, perhaps we need a single call button which then gives you voice or video as a choice
### Outcome
#### What did you expect?
Intuitive calling
#### What happened instead?
I kept getting video calls because folks can no longer find the voice call button.
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
enabling threads hides the voice call button steps to reproduce voice button is now hidden behind the hamburger icon except it’s not labelled “voice call” it’s “call” which makes it even harder to find instead perhaps we need a single call button which then gives you voice or video as a choice outcome what did you expect intuitive calling what happened instead i kept getting video calls because folks can no longer find the voice call button your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no
| 1
|
18,502
| 2,615,172,357
|
IssuesEvent
|
2015-03-01 06:54:25
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
opened
|
Streaming audio on the web
|
auto-migrated Milestone-X New Priority-P2 Tutorial Type-Feature
|
```
- media source api
- binary web sockets
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 31 Jan 2012 at 8:26
|
1.0
|
Streaming audio on the web - ```
- media source api
- binary web sockets
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 31 Jan 2012 at 8:26
|
non_defect
|
streaming audio on the web media source api binary web sockets original issue reported on code google com by ericbide com on jan at
| 0
|
36,822
| 5,088,285,280
|
IssuesEvent
|
2016-12-31 17:41:09
|
NMGRL/pychron
|
https://api.github.com/repos/NMGRL/pychron
|
closed
|
Furnace loader OK
|
TestingRequired
|
Felix and Jan running well. Furnace sample dumped perfectly with electromagnets. Perhaps Matt did have the valves shut when he tried it. All controls working well except "Dump" button, which is seemingly not doing anything. All other button elements fine. Should dump button actuate valves and shutters, then dump? Or does it actuate the magnets and mark the position as dumped? Perhaps the button should be labelled "Load" no dump, and the "Magnets" button should be labelled "Dump". I will have a look at whether all buckets will now dump.
Could funnel position be encoded into the status word? These would allow easy animation of the funnel graphic.
|
1.0
|
Furnace loader OK - Felix and Jan running well. Furnace sample dumped perfectly with electromagnets. Perhaps Matt did have the valves shut when he tried it. All controls working well except "Dump" button, which is seemingly not doing anything. All other button elements fine. Should dump button actuate valves and shutters, then dump? Or does it actuate the magnets and mark the position as dumped? Perhaps the button should be labelled "Load" no dump, and the "Magnets" button should be labelled "Dump". I will have a look at whether all buckets will now dump.
Could funnel position be encoded into the status word? These would allow easy animation of the funnel graphic.
|
non_defect
|
furnace loader ok felix and jan running well furnace sample dumped perfectly with electromagnets perhaps matt did have the valves shut when he tried it all controls working well except dump button which is seemingly not doing anything all other button elements fine should dump button actuate valves and shutters then dump or does it actuate the magnets and mark the position as dumped perhaps the button should be labelled load no dump and the magnets button should be labelled dump i will have a look at whether all buckets will now dump could funnel position be encoded into the status word these would allow easy animation of the funnel graphic
| 0
|
83,181
| 10,329,866,973
|
IssuesEvent
|
2019-09-02 13:18:14
|
AugurProject/augur
|
https://api.github.com/repos/AugurProject/augur
|
opened
|
Trading: Updating resized components + color changes
|
Design Roadmap: Trading
|
After changes made to the Trading page, resizing and refining components.
We need to apply that to components on:
- Account summary
- Portfolio
- Reporting
We're also updating the UI color scheme
Need to be applied across desktop/mobile etc
|
1.0
|
Trading: Updating resized components + color changes - After changes made to the Trading page, resizing and refining components.
We need to apply that to components on:
- Account summary
- Portfolio
- Reporting
We're also updating the UI color scheme
Need to be applied across desktop/mobile etc
|
non_defect
|
trading updating resized components color changes after changes made to the trading page resizing and refining components we need to apply that to components on account summary portfolio reporting we re also updating the ui color scheme need to be applied across desktop mobile etc
| 0
|
13,693
| 3,176,000,096
|
IssuesEvent
|
2015-09-24 05:28:54
|
sulu-io/sulu
|
https://api.github.com/repos/sulu-io/sulu
|
closed
|
Admin Frontend Issues
|
Design
|
The following issues have been observed
Contact/Account in ContactExtensionBundle
* [x] Wrong space distribution when sidebar is activated and browser width is between 900px and 1090px
datagrid:
<img width="1082" alt="bildschirmfoto 2015-09-17 um 11 32 40" src="https://cloud.githubusercontent.com/assets/5152633/9929873/dec9c294-5d2f-11e5-819c-aaffaa157757.png">
* [x] Styling of content, when sidebar overlaps content
<img width="941" alt="bildschirmfoto 2015-09-17 um 10 55 14" src="https://cloud.githubusercontent.com/assets/5152633/9929068/c1ae8dde-5d2a-11e5-9004-5865419d0c26.png">
* [x] nested table view: error when resizing browser

* [x] Resolution 1440px does not comply with sidebar. In fact, the navigation should be collapsed:
<img width="1402" alt="bildschirmfoto 2015-09-17 um 11 26 22" src="https://cloud.githubusercontent.com/assets/5152633/9929812/8038b226-5d2f-11e5-8563-9be0e1b1b70c.png">
|
1.0
|
Admin Frontend Issues - The following issues have been observed
Contact/Account in ContactExtensionBundle
* [x] Wrong space distribution when sidebar is activated and browser width is between 900px and 1090px
datagrid:
<img width="1082" alt="bildschirmfoto 2015-09-17 um 11 32 40" src="https://cloud.githubusercontent.com/assets/5152633/9929873/dec9c294-5d2f-11e5-819c-aaffaa157757.png">
* [x] Styling of content, when sidebar overlaps content
<img width="941" alt="bildschirmfoto 2015-09-17 um 10 55 14" src="https://cloud.githubusercontent.com/assets/5152633/9929068/c1ae8dde-5d2a-11e5-9004-5865419d0c26.png">
* [x] nested table view: error when resizing browser

* [x] Resolution 1440px does not comply with sidebar. In fact, the navigation should be collapsed:
<img width="1402" alt="bildschirmfoto 2015-09-17 um 11 26 22" src="https://cloud.githubusercontent.com/assets/5152633/9929812/8038b226-5d2f-11e5-8563-9be0e1b1b70c.png">
|
non_defect
|
admin frontend issues the following issues have been observed contact account in contactextensionbundle wrong space distribution when sidebar is activated and browser width is between and datagrid img width alt bildschirmfoto um src styling of content when sidebar overlaps content img width alt bildschirmfoto um src nested table view error when resizing browser resolution does not comply with sidebar in fact the navigation should be collapsed img width alt bildschirmfoto um src
| 0
|
212,390
| 23,887,922,153
|
IssuesEvent
|
2022-09-08 09:09:40
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
[Security Solution] Any number after decimal is not allowed except zero for the timer
|
bug impact:medium Team: SecuritySolution Team:Onboarding and Lifecycle Mgt v8.5.0
|
**Description:**
Any number after decimal is not allowed except zero for the timer
**Build Details:**
```
VERSION: 8.4.0 BC1
BUILD: 54999
COMMIT: 58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d
ARTIFACTS: https://staging.elastic.co/8.4.0-91109cf0/summary-8.4.0.html
```
**Browser Details:**
All
**Preconditions:**
1. Kibana user should be logged in
**Steps to Reproduce:**
1. Navigate to the Endpoints tab under the Manage section of the Security from the left-hand side navigation bar
2. Click on the three dots for an endpoint
3. Click on the launch responder option
4. Click on the Actions log button
5. Click on the timer button
6. Try adding a value after decimal other than zero.
- Example: 5.6 or 8.3
7. Observe when we press any number to add, the decimal is removed and no value is added to the timer
**Actual Result:**
Any number after decimal is not allowed except zero for the timer
**Screen Recording:**
https://user-images.githubusercontent.com/60252716/189079409-eee117d4-aec2-4b0e-b5d4-2fb88a5a39b2.mp4
**Logs:**
N/A
|
True
|
[Security Solution] Any number after decimal is not allowed except zero for the timer - **Description:**
Any number after decimal is not allowed except zero for the timer
**Build Details:**
```
VERSION: 8.4.0 BC1
BUILD: 54999
COMMIT: 58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d
ARTIFACTS: https://staging.elastic.co/8.4.0-91109cf0/summary-8.4.0.html
```
**Browser Details:**
All
**Preconditions:**
1. Kibana user should be logged in
**Steps to Reproduce:**
1. Navigate to the Endpoints tab under the Manage section of the Security from the left-hand side navigation bar
2. Click on the three dots for an endpoint
3. Click on the launch responder option
4. Click on the Actions log button
5. Click on the timer button
6. Try adding a value after decimal other than zero.
- Example: 5.6 or 8.3
7. Observe when we press any number to add, the decimal is removed and no value is added to the timer
**Actual Result:**
Any number after decimal is not allowed except zero for the timer
**Screen Recording:**
https://user-images.githubusercontent.com/60252716/189079409-eee117d4-aec2-4b0e-b5d4-2fb88a5a39b2.mp4
**Logs:**
N/A
|
non_defect
|
any number after decimal is not allowed except zero for the timer description any number after decimal is not allowed except zero for the timer build details version build commit artifacts browser details all preconditions kibana user should be logged in steps to reproduce navigate to the endpoints tab under the manage section of the security from the left hand side navigation bar click on the three dots for an endpoint click on the launch responder option click on the actions log button click on the timer button try adding a value after decimal other than zero example or observe when we press any number to add the decimal is removed and no value is added to the timer actual result any number after decimal is not allowed except zero for the timer screen recording logs n a
| 0
|
70,346
| 9,403,723,450
|
IssuesEvent
|
2019-04-09 02:42:51
|
pyouroboros/ouroboros
|
https://api.github.com/repos/pyouroboros/ouroboros
|
reopened
|
Document swarm configuration
|
documentation
|
The documentation doesn't mention anything about Swarm, but there are configurations relevant to it:
I think `SWARM=true` needs to be passed.
It seems that then `LABEL_ENABLE` defaults to `true`, so `MONITOR` is ignored. I haven't tried setting it to false.
|
1.0
|
Document swarm configuration - The documentation doesn't mention anything about Swarm, but there are configurations relevant to it:
I think `SWARM=true` needs to be passed.
It seems that then `LABEL_ENABLE` defaults to `true`, so `MONITOR` is ignored. I haven't tried setting it to false.
|
non_defect
|
document swarm configuration the documentation doesn t mention anything about swarm but there are configurations relevant to it i think swarm true needs to be passed it seems that then label enable defaults to true so monitor is ignored i haven t tried setting it to false
| 0
|
16,834
| 2,948,322,953
|
IssuesEvent
|
2015-07-06 01:31:15
|
Winetricks/winetricks
|
https://api.github.com/repos/Winetricks/winetricks
|
closed
|
MinGW should install mingw32-make
|
auto-migrated Priority-Medium Type-Defect
|
```
MinGW is useless without make.
```
Original issue reported on code.google.com by `disin...@gmail.com` on 12 Feb 2012 at 6:28
|
1.0
|
MinGW should install mingw32-make - ```
MinGW is useless without make.
```
Original issue reported on code.google.com by `disin...@gmail.com` on 12 Feb 2012 at 6:28
|
defect
|
mingw should install make mingw is useless without make original issue reported on code google com by disin gmail com on feb at
| 1
|
267,592
| 28,509,121,450
|
IssuesEvent
|
2023-04-19 01:37:04
|
dpteam/RK3188_TABLET
|
https://api.github.com/repos/dpteam/RK3188_TABLET
|
closed
|
CVE-2011-3191 (High) detected in linux-yocto-4.12v3.1.10 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2011-3191 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary>
<p>
<p>Linux 4.12 Embedded Kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/cifs/cifssmb.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Integer signedness error in the CIFSFindNext function in fs/cifs/cifssmb.c in the Linux kernel before 3.1 allows remote CIFS servers to cause a denial of service (memory corruption) or possibly have unspecified other impact via a large length value in a response to a read request for a directory.
<p>Publish Date: 2012-05-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-3191>CVE-2011-3191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-3191">https://nvd.nist.gov/vuln/detail/CVE-2011-3191</a></p>
<p>Release Date: 2012-05-24</p>
<p>Fix Resolution: 3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2011-3191 (High) detected in linux-yocto-4.12v3.1.10 - autoclosed - ## CVE-2011-3191 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary>
<p>
<p>Linux 4.12 Embedded Kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/cifs/cifssmb.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Integer signedness error in the CIFSFindNext function in fs/cifs/cifssmb.c in the Linux kernel before 3.1 allows remote CIFS servers to cause a denial of service (memory corruption) or possibly have unspecified other impact via a large length value in a response to a read request for a directory.
<p>Publish Date: 2012-05-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-3191>CVE-2011-3191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-3191">https://nvd.nist.gov/vuln/detail/CVE-2011-3191</a></p>
<p>Release Date: 2012-05-24</p>
<p>Fix Resolution: 3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linux yocto autoclosed cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files fs cifs cifssmb c vulnerability details integer signedness error in the cifsfindnext function in fs cifs cifssmb c in the linux kernel before allows remote cifs servers to cause a denial of service memory corruption or possibly have unspecified other impact via a large length value in a response to a read request for a directory publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
53,151
| 13,261,029,907
|
IssuesEvent
|
2020-08-20 19:13:44
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[icetray] Adding services directly to the context bypasses I3TrayInfo (Trac #826)
|
Migrated from Trac combo core defect
|
With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:
```text
from I3Tray import I3Tray
from icecube import icetray, dataclasses, dataio, phys_services
tray = I3Tray()
randomService = phys_services.I3GSLRandomService(seed=12345)
tray.context["I3RandomService"] = randomService
tray.AddModule("I3InfiniteSource")
tray.AddModule("I3Writer","writer",filename="test.i3")
tray.Execute(1)
```
The seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code.
There are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/826">https://code.icecube.wisc.edu/projects/icecube/ticket/826</a>, reported by cweaverand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T13:54:32",
"_ts": "1550066072250335",
"description": "With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:\n\n{{{\nfrom I3Tray import I3Tray\nfrom icecube import icetray, dataclasses, dataio, phys_services\ntray = I3Tray()\nrandomService = phys_services.I3GSLRandomService(seed=12345)\ntray.context[\"I3RandomService\"] = randomService\ntray.AddModule(\"I3InfiniteSource\")\ntray.AddModule(\"I3Writer\",\"writer\",filename=\"test.i3\")\ntray.Execute(1)\n}}}\n\nThe seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. \n\nThere are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. ",
"reporter": "cweaver",
"cc": "",
"resolution": "insufficient resources",
"time": "2014-12-04T22:37:06",
"component": "combo core",
"summary": "[icetray] Adding services directly to the context bypasses I3TrayInfo",
"priority": "normal",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[icetray] Adding services directly to the context bypasses I3TrayInfo (Trac #826) - With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:
```text
from I3Tray import I3Tray
from icecube import icetray, dataclasses, dataio, phys_services
tray = I3Tray()
randomService = phys_services.I3GSLRandomService(seed=12345)
tray.context["I3RandomService"] = randomService
tray.AddModule("I3InfiniteSource")
tray.AddModule("I3Writer","writer",filename="test.i3")
tray.Execute(1)
```
The seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code.
There are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/826">https://code.icecube.wisc.edu/projects/icecube/ticket/826</a>, reported by cweaverand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T13:54:32",
"_ts": "1550066072250335",
"description": "With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:\n\n{{{\nfrom I3Tray import I3Tray\nfrom icecube import icetray, dataclasses, dataio, phys_services\ntray = I3Tray()\nrandomService = phys_services.I3GSLRandomService(seed=12345)\ntray.context[\"I3RandomService\"] = randomService\ntray.AddModule(\"I3InfiniteSource\")\ntray.AddModule(\"I3Writer\",\"writer\",filename=\"test.i3\")\ntray.Execute(1)\n}}}\n\nThe seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. \n\nThere are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. ",
"reporter": "cweaver",
"cc": "",
"resolution": "insufficient resources",
"time": "2014-12-04T22:37:06",
"component": "combo core",
"summary": "[icetray] Adding services directly to the context bypasses I3TrayInfo",
"priority": "normal",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
adding services directly to the context bypasses trac with the reworking of the context it has become possible to add services without the aid of a service factory which is convenient in many cases unfortunately it also does an end run around the trayinfo bookkeeping system since this discovers services and their parameters by iterating over the list of factories contained in the tray a relevant example is code like this text from import from icecube import icetray dataclasses dataio phys services tray randomservice phys services seed tray context randomservice tray addmodule tray addmodule writer filename test tray execute the seed used by the rng is a very important piece of information which the user may want to inspect after the fact but using code like the above hides it from the trayinfoservice as can be seen by examining the i frame in the file generated by running this code there are a few difficulties associated with trying to correct this first arbitrary objects may be constructed and placed in the context which may not conform to any standard interface to extract parameter values second the trayinfo may be queried at times after the service factories have run so it would have to somehow avoid double counting the services installed by the factories but these services names may be unrelated to the names of their corresponding factories migrated from json status closed changetime ts description with the reworking of the context it has become possible to add services without the aid of a service factory which is convenient in many cases unfortunately it also does an end run around the trayinfo bookkeeping system since this discovers services and their parameters by iterating over the list of factories contained in the tray a relevant example is code like this n n nfrom import nfrom icecube import icetray dataclasses dataio phys services ntray nrandomservice phys services seed ntray context randomservice ntray addmodule ntray addmodule writer filename test ntray execute n n nthe seed used by the rng is a very important piece of information which the user may want to inspect after the fact but using code like the above hides it from the trayinfoservice as can be seen by examining the i frame in the file generated by running this code n nthere are a few difficulties associated with trying to correct this first arbitrary objects may be constructed and placed in the context which may not conform to any standard interface to extract parameter values second the trayinfo may be queried at times after the service factories have run so it would have to somehow avoid double counting the services installed by the factories but these services names may be unrelated to the names of their corresponding factories reporter cweaver cc resolution insufficient resources time component combo core summary adding services directly to the context bypasses priority normal keywords milestone long term future owner olivas type defect
| 1
|
354,814
| 25,174,967,526
|
IssuesEvent
|
2022-11-11 08:24:28
|
mariavallejo20/proyecto_mvo
|
https://api.github.com/repos/mariavallejo20/proyecto_mvo
|
closed
|
[M0- Dev] Elección del gestor de dependencias
|
documentation
|
# Gestor de dependencias
Un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades. Para ello, será necesario documentarse sobre distintas posibilidades y, finalmente, realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio.
## Criterios a seguir para la elección
Para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta:
1. Estándares del lenguaje:
* Fichero "package.json", no ha sido establecido como estándar para NodeJS, sin embargo es la práctica más extendida para la gestión de dependencias.
2. Buenas prácticas:
* Debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores.
* Criterios específicos de mi proyecto:
* Se trata de un proyecto de pequeño tamaño
* Se debe tener en cuenta la velocidad del gestor.
* Emplear un único fichero para la gestión de las dependencias, en este caso "package.json", lo que permitirá no aumentar la deuda técnica en el futuro.
* Se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias.
3. Actualización del gestor: Se deberá tener en cuenta la comunidad de la que dispone el gestor, ya que se busca tener un buen mantenimiento y actualización. Un buen soporte desembocará en una mayor seguridad, criterio que debemos tener en cuenta ya que es importante.
## Posibles gestores de dependencias
* **NPM (Node Package Manager)**:
- Es un gestor aconsejado por una gran comunidad de desarrolladores, ya que viene empaquetado con NodeJS.
- Cuenta una gran comunidad y con un buen soporte y mantenimiento, con varias actualizaciones al año y siendo la última hace 5 meses.
- Para la gestión de dependencias utiliza el archivo "package.json".
- La instalación de paquetes no es demasiado rápida, sin embargo, teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá, este no será un gran inconveniente.
- NPM realiza una evaluación de seguridad a partir de la versión 6, lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible.
- NPM habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio.
* **Bower**:
- Para la gestión de dependencias emplea el archivo "boxer.json", lo que no cumple con la práctica extendida por la mayoría de los desarrolladores. Esto puede provocar deuda técnica en un futuro, por lo que lo descartaré como una opción.
* **Yarn**:
- Junto con NPM es el gestor de dependencias más utiizado por los desarrolladores de NodeJS. Sin embargo, este no viene instalado con la instalación del lenguaje.
- Cuenta con un buen soporte y mantenimiento, con varias actualizaciones anuales; la última hace 5 meses.
- Emplea para la gestión el archivo "package.json" lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro.
- Yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente, lo que aumenta la velocidad de instalación de las dependencias. Sin embargo, teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa, aunque si habría que tenerlo en cuenta para proyectos mayores.
- Mientra descarga paquetes, Yarn realiza una verificación de seguridad en segundo plano, aumentando la seguridad.
- Yarn habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio.
* **pnpm**:
- PNPM es más rápido que los anteriores a la hora de las instalación de dependencias, sobre todo para paquetes de mayor tamaño.
- En la seguridad NPM es similar a Yarn, ya que también emplea sumas de verificación. Además, pnpm
verifica la integridad de su código antes de ejecutarlo.
- PNPM cuenta con una gran comunidad de desarrolladores y un buen soporte, con actualizaciones frecuentes.
- Para la gestión de dependencias utiliza el archivo "package.json".
## Elección
Tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos, he decidido lo siguiente:
NPM, Yarn y PNPM son buenas opciones para usar en mi proyecto.
He decidido descartar NPM ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras.
Por otro lado, entre PNPM y Yarn, **he decidido escoger PNPM**, ya que ambas tienes características similares que cumplen los criterios establecidos, buen soporte y documentación, pero PNPM presenta una mayor velicidad de instalación de los paquetes.
Enlazado con la HU3[#6]
|
1.0
|
[M0- Dev] Elección del gestor de dependencias - # Gestor de dependencias
Un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades. Para ello, será necesario documentarse sobre distintas posibilidades y, finalmente, realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio.
## Criterios a seguir para la elección
Para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta:
1. Estándares del lenguaje:
* Fichero "package.json", no ha sido establecido como estándar para NodeJS, sin embargo es la práctica más extendida para la gestión de dependencias.
2. Buenas prácticas:
* Debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores.
* Criterios específicos de mi proyecto:
* Se trata de un proyecto de pequeño tamaño
* Se debe tener en cuenta la velocidad del gestor.
* Emplear un único fichero para la gestión de las dependencias, en este caso "package.json", lo que permitirá no aumentar la deuda técnica en el futuro.
* Se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias.
3. Actualización del gestor: Se deberá tener en cuenta la comunidad de la que dispone el gestor, ya que se busca tener un buen mantenimiento y actualización. Un buen soporte desembocará en una mayor seguridad, criterio que debemos tener en cuenta ya que es importante.
## Posibles gestores de dependencias
* **NPM (Node Package Manager)**:
- Es un gestor aconsejado por una gran comunidad de desarrolladores, ya que viene empaquetado con NodeJS.
- Cuenta una gran comunidad y con un buen soporte y mantenimiento, con varias actualizaciones al año y siendo la última hace 5 meses.
- Para la gestión de dependencias utiliza el archivo "package.json".
- La instalación de paquetes no es demasiado rápida, sin embargo, teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá, este no será un gran inconveniente.
- NPM realiza una evaluación de seguridad a partir de la versión 6, lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible.
- NPM habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio.
* **Bower**:
- Para la gestión de dependencias emplea el archivo "boxer.json", lo que no cumple con la práctica extendida por la mayoría de los desarrolladores. Esto puede provocar deuda técnica en un futuro, por lo que lo descartaré como una opción.
* **Yarn**:
- Junto con NPM es el gestor de dependencias más utiizado por los desarrolladores de NodeJS. Sin embargo, este no viene instalado con la instalación del lenguaje.
- Cuenta con un buen soporte y mantenimiento, con varias actualizaciones anuales; la última hace 5 meses.
- Emplea para la gestión el archivo "package.json" lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro.
- Yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente, lo que aumenta la velocidad de instalación de las dependencias. Sin embargo, teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa, aunque si habría que tenerlo en cuenta para proyectos mayores.
- Mientra descarga paquetes, Yarn realiza una verificación de seguridad en segundo plano, aumentando la seguridad.
- Yarn habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio.
* **pnpm**:
- PNPM es más rápido que los anteriores a la hora de las instalación de dependencias, sobre todo para paquetes de mayor tamaño.
- En la seguridad NPM es similar a Yarn, ya que también emplea sumas de verificación. Además, pnpm
verifica la integridad de su código antes de ejecutarlo.
- PNPM cuenta con una gran comunidad de desarrolladores y un buen soporte, con actualizaciones frecuentes.
- Para la gestión de dependencias utiliza el archivo "package.json".
## Elección
Tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos, he decidido lo siguiente:
NPM, Yarn y PNPM son buenas opciones para usar en mi proyecto.
He decidido descartar NPM ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras.
Por otro lado, entre PNPM y Yarn, **he decidido escoger PNPM**, ya que ambas tienes características similares que cumplen los criterios establecidos, buen soporte y documentación, pero PNPM presenta una mayor velicidad de instalación de los paquetes.
Enlazado con la HU3[#6]
|
non_defect
|
elección del gestor de dependencias gestor de dependencias un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades para ello será necesario documentarse sobre distintas posibilidades y finalmente realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio criterios a seguir para la elección para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta estándares del lenguaje fichero package json no ha sido establecido como estándar para nodejs sin embargo es la práctica más extendida para la gestión de dependencias buenas prácticas debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores criterios específicos de mi proyecto se trata de un proyecto de pequeño tamaño se debe tener en cuenta la velocidad del gestor emplear un único fichero para la gestión de las dependencias en este caso package json lo que permitirá no aumentar la deuda técnica en el futuro se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias actualización del gestor se deberá tener en cuenta la comunidad de la que dispone el gestor ya que se busca tener un buen mantenimiento y actualización un buen soporte desembocará en una mayor seguridad criterio que debemos tener en cuenta ya que es importante posibles gestores de dependencias npm node package manager es un gestor aconsejado por una gran comunidad de desarrolladores ya que viene empaquetado con nodejs cuenta una gran comunidad y con un buen soporte y mantenimiento con varias actualizaciones al año y siendo la última hace meses para la gestión de dependencias utiliza el archivo package json la instalación de paquetes no es demasiado rápida sin embargo teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá este no será un gran inconveniente npm realiza una evaluación de seguridad a partir de la versión lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible npm habilita espacios de trabajo lo que le permite administrar dependencias para varios proyectos desde un único repositorio bower para la gestión de dependencias emplea el archivo boxer json lo que no cumple con la práctica extendida por la mayoría de los desarrolladores esto puede provocar deuda técnica en un futuro por lo que lo descartaré como una opción yarn junto con npm es el gestor de dependencias más utiizado por los desarrolladores de nodejs sin embargo este no viene instalado con la instalación del lenguaje cuenta con un buen soporte y mantenimiento con varias actualizaciones anuales la última hace meses emplea para la gestión el archivo package json lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente lo que aumenta la velocidad de instalación de las dependencias sin embargo teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa aunque si habría que tenerlo en cuenta para proyectos mayores mientra descarga paquetes yarn realiza una verificación de seguridad en segundo plano aumentando la seguridad yarn habilita espacios de trabajo lo que le permite administrar dependencias para varios proyectos desde un único repositorio pnpm pnpm es más rápido que los anteriores a la hora de las instalación de dependencias sobre todo para paquetes de mayor tamaño en la seguridad npm es similar a yarn ya que también emplea sumas de verificación además pnpm verifica la integridad de su código antes de ejecutarlo pnpm cuenta con una gran comunidad de desarrolladores y un buen soporte con actualizaciones frecuentes para la gestión de dependencias utiliza el archivo package json elección tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos he decidido lo siguiente npm yarn y pnpm son buenas opciones para usar en mi proyecto he decidido descartar npm ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras por otro lado entre pnpm y yarn he decidido escoger pnpm ya que ambas tienes características similares que cumplen los criterios establecidos buen soporte y documentación pero pnpm presenta una mayor velicidad de instalación de los paquetes enlazado con la
| 0
|
7,219
| 2,610,358,338
|
IssuesEvent
|
2015-02-26 19:56:01
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Blogspot post, Title and URL problem
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
When we posted to blogspot, url blogspot and title almost similar but
today, when i post an article to blogspot, the url post using first paragraph
from article?
Are scribefire error problem connecting to blogger api?
I try post manually from blogspot editor and i don't get error?
What browser are you using?
firefox 3.6
What version of ScribeFire are you running?
scribefire next 1.9
```
-----
Original issue reported on code.google.com by `tonit...@gmail.com` on 20 Nov 2011 at 1:44
|
1.0
|
Blogspot post, Title and URL problem - ```
What's the problem?
When we posted to blogspot, url blogspot and title almost similar but
today, when i post an article to blogspot, the url post using first paragraph
from article?
Are scribefire error problem connecting to blogger api?
I try post manually from blogspot editor and i don't get error?
What browser are you using?
firefox 3.6
What version of ScribeFire are you running?
scribefire next 1.9
```
-----
Original issue reported on code.google.com by `tonit...@gmail.com` on 20 Nov 2011 at 1:44
|
defect
|
blogspot post title and url problem what s the problem when we posted to blogspot url blogspot and title almost similar but today when i post an article to blogspot the url post using first paragraph from article are scribefire error problem connecting to blogger api i try post manually from blogspot editor and i don t get error what browser are you using firefox what version of scribefire are you running scribefire next original issue reported on code google com by tonit gmail com on nov at
| 1
|
637,740
| 20,676,479,361
|
IssuesEvent
|
2022-03-10 09:46:23
|
AY2122S2-TIC4002-F18-5/tp2
|
https://api.github.com/repos/AY2122S2-TIC4002-F18-5/tp2
|
closed
|
Add a log command to register a brief log update on a Person
|
priority.High
|
Add a log command to register a brief text log message giving a recent or important update relevant to a team member Person.
|
1.0
|
Add a log command to register a brief log update on a Person - Add a log command to register a brief text log message giving a recent or important update relevant to a team member Person.
|
non_defect
|
add a log command to register a brief log update on a person add a log command to register a brief text log message giving a recent or important update relevant to a team member person
| 0
|
713,853
| 24,541,622,515
|
IssuesEvent
|
2022-10-12 04:38:55
|
AY2223S1-CS2113-F11-1/tp
|
https://api.github.com/repos/AY2223S1-CS2113-F11-1/tp
|
closed
|
Parse Command
|
type.Chore priority.High
|
When the user initiates a command in String, parse and execute the relevant command.
|
1.0
|
Parse Command - When the user initiates a command in String, parse and execute the relevant command.
|
non_defect
|
parse command when the user initiates a command in string parse and execute the relevant command
| 0
|
72,472
| 24,132,662,091
|
IssuesEvent
|
2022-09-21 08:44:49
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
SEVERE Exception for SQL limit clause [HZ-1436]
|
Type: Defect Source: Internal to-jira Team: Platform
|
**Describe the bug**
An exception appears in the logs for an SQL clause that stops because of LIMIT
**Expected behavior**
No exception in logs, noise
**To Reproduce**
```
public static void main(String[] args) throws Exception {
Config config = new Config();
config.getJetConfig().setEnabled(true);
HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config);
hazelcastInstance.getMap("neil").put("Hello", "World");
hazelcastInstance.getMap("neil").put("Goodbye", "World");
String mapping = "CREATE MAPPING neil TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java',"
+ " 'keyJavaClass' = '" + String.class.getName() + "'," + " 'valueFormat' = 'java',"
+ " 'valueJavaClass' = '" + String.class.getName() + "'" + " )";
String query = "SELECT * FROM neil LIMIT 1";
List.of(mapping, query).forEach(sql -> hazelcastInstance.getSql().execute(sql));
}
```
gives
```
com.hazelcast.jet.JetException: Execution on a member failed: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause
at com.hazelcast.jet.impl.LightMasterContext.findError(LightMasterContext.java:310)
at com.hazelcast.jet.impl.LightMasterContext.lambda$createContext$2(LightMasterContext.java:164)
at com.hazelcast.jet.impl.LightMasterContext.lambda$invokeOnParticipant$8(LightMasterContext.java:281)
at com.hazelcast.spi.impl.AbstractInvocationFuture$WhenCompleteNode.lambda$execute$0(AbstractInvocationFuture.java:1564)
at java.base/java.util.concurrent.ForkJoinTask$RunnableExecuteAction.exec(ForkJoinTask.java:1395)
at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:373)
at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1182)
at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1655)
at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1622)
at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165)
Caused by: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause
```
**Additional context**
Hazelcast Platform 5.2-SNAPSHOT (20220819 - 111323f)
|
1.0
|
SEVERE Exception for SQL limit clause [HZ-1436] - **Describe the bug**
An exception appears in the logs for an SQL clause that stops because of LIMIT
**Expected behavior**
No exception in logs, noise
**To Reproduce**
```
public static void main(String[] args) throws Exception {
Config config = new Config();
config.getJetConfig().setEnabled(true);
HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config);
hazelcastInstance.getMap("neil").put("Hello", "World");
hazelcastInstance.getMap("neil").put("Goodbye", "World");
String mapping = "CREATE MAPPING neil TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java',"
+ " 'keyJavaClass' = '" + String.class.getName() + "'," + " 'valueFormat' = 'java',"
+ " 'valueJavaClass' = '" + String.class.getName() + "'" + " )";
String query = "SELECT * FROM neil LIMIT 1";
List.of(mapping, query).forEach(sql -> hazelcastInstance.getSql().execute(sql));
}
```
gives
```
com.hazelcast.jet.JetException: Execution on a member failed: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause
at com.hazelcast.jet.impl.LightMasterContext.findError(LightMasterContext.java:310)
at com.hazelcast.jet.impl.LightMasterContext.lambda$createContext$2(LightMasterContext.java:164)
at com.hazelcast.jet.impl.LightMasterContext.lambda$invokeOnParticipant$8(LightMasterContext.java:281)
at com.hazelcast.spi.impl.AbstractInvocationFuture$WhenCompleteNode.lambda$execute$0(AbstractInvocationFuture.java:1564)
at java.base/java.util.concurrent.ForkJoinTask$RunnableExecuteAction.exec(ForkJoinTask.java:1395)
at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:373)
at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1182)
at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1655)
at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1622)
at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165)
Caused by: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause
```
**Additional context**
Hazelcast Platform 5.2-SNAPSHOT (20220819 - 111323f)
|
defect
|
severe exception for sql limit clause describe the bug an exception appears in the logs for an sql clause that stops because of limit expected behavior no exception in logs noise to reproduce public static void main string args throws exception config config new config config getjetconfig setenabled true hazelcastinstance hazelcastinstance hazelcast newhazelcastinstance config hazelcastinstance getmap neil put hello world hazelcastinstance getmap neil put goodbye world string mapping create mapping neil type imap options keyformat java keyjavaclass string class getname valueformat java valuejavaclass string class getname string query select from neil limit list of mapping query foreach sql hazelcastinstance getsql execute sql gives com hazelcast jet jetexception execution on a member failed com hazelcast sql impl resultlimitreachedexception done by reaching the item number in sql limit clause at com hazelcast jet impl lightmastercontext finderror lightmastercontext java at com hazelcast jet impl lightmastercontext lambda createcontext lightmastercontext java at com hazelcast jet impl lightmastercontext lambda invokeonparticipant lightmastercontext java at com hazelcast spi impl abstractinvocationfuture whencompletenode lambda execute abstractinvocationfuture java at java base java util concurrent forkjointask runnableexecuteaction exec forkjointask java at java base java util concurrent forkjointask doexec forkjointask java at java base java util concurrent forkjoinpool workqueue toplevelexec forkjoinpool java at java base java util concurrent forkjoinpool scan forkjoinpool java at java base java util concurrent forkjoinpool runworker forkjoinpool java at java base java util concurrent forkjoinworkerthread run forkjoinworkerthread java caused by com hazelcast sql impl resultlimitreachedexception done by reaching the item number in sql limit clause additional context hazelcast platform snapshot
| 1
|
24,880
| 4,117,831,080
|
IssuesEvent
|
2016-06-08 09:05:50
|
abc306396358/pytesser
|
https://api.github.com/repos/abc306396358/pytesser
|
closed
|
How to get whitelist to work with pytesseract
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
Trying to use the code that makes a whitelist for Tesseract like follows
ocr = tesseract.TessBaseAPI()
ocr.SetVariable("tessedit_char_whitelist", "0123456789;")
ocr.SetPageSegMode(tesseract.PSM_AUTO)
ocr.Init("C:\\Program Files (x86)\\Tesseract-OCR\\","eng",tesseract.OEM_DEFAULT)
What is the expected output? What do you see instead?
Intended output is to have only "0123456789;" characters be recognized when
using the image_to_string() function. Using code like what is above,
image_to_string() just ignores it and grabs whatever characters it finds.
What version of the product are you using? On what operating system?
pytesseract-0.1, Python 2.7, Windows 8.1
Please provide any additional information below.
I've been trying everything people use for Tesseract-OCR, but that doesn't work
with pytesseract. I haven't been able to find any solution or method to
whitelisting with the image_to_string() function anywhere, which would be
immensely helpful in improving the accuracy of the function.
Thanks in advance for any help on the matter.
```
Original issue reported on code.google.com by `darke...@yahoo.com` on 9 Jun 2015 at 6:58
|
1.0
|
How to get whitelist to work with pytesseract - ```
What steps will reproduce the problem?
Trying to use the code that makes a whitelist for Tesseract like follows
ocr = tesseract.TessBaseAPI()
ocr.SetVariable("tessedit_char_whitelist", "0123456789;")
ocr.SetPageSegMode(tesseract.PSM_AUTO)
ocr.Init("C:\\Program Files (x86)\\Tesseract-OCR\\","eng",tesseract.OEM_DEFAULT)
What is the expected output? What do you see instead?
Intended output is to have only "0123456789;" characters be recognized when
using the image_to_string() function. Using code like what is above,
image_to_string() just ignores it and grabs whatever characters it finds.
What version of the product are you using? On what operating system?
pytesseract-0.1, Python 2.7, Windows 8.1
Please provide any additional information below.
I've been trying everything people use for Tesseract-OCR, but that doesn't work
with pytesseract. I haven't been able to find any solution or method to
whitelisting with the image_to_string() function anywhere, which would be
immensely helpful in improving the accuracy of the function.
Thanks in advance for any help on the matter.
```
Original issue reported on code.google.com by `darke...@yahoo.com` on 9 Jun 2015 at 6:58
|
defect
|
how to get whitelist to work with pytesseract what steps will reproduce the problem trying to use the code that makes a whitelist for tesseract like follows ocr tesseract tessbaseapi ocr setvariable tessedit char whitelist ocr setpagesegmode tesseract psm auto ocr init c program files tesseract ocr eng tesseract oem default what is the expected output what do you see instead intended output is to have only characters be recognized when using the image to string function using code like what is above image to string just ignores it and grabs whatever characters it finds what version of the product are you using on what operating system pytesseract python windows please provide any additional information below i ve been trying everything people use for tesseract ocr but that doesn t work with pytesseract i haven t been able to find any solution or method to whitelisting with the image to string function anywhere which would be immensely helpful in improving the accuracy of the function thanks in advance for any help on the matter original issue reported on code google com by darke yahoo com on jun at
| 1
|
66,149
| 6,989,730,772
|
IssuesEvent
|
2017-12-14 17:03:22
|
saltstack/salt
|
https://api.github.com/repos/saltstack/salt
|
opened
|
Add timezone windows Tests to Auto Test Suite
|
auto-test Windows
|
This task is to monitor the addition of four tests to the auto test suite:
- [ ] timezone.get_zone
- [ ] timezone.get_offset
- [ ] timezone.set_zone
- [ ] timezone.zone_compare
There are unit tests but need to add integration tests
|
1.0
|
Add timezone windows Tests to Auto Test Suite - This task is to monitor the addition of four tests to the auto test suite:
- [ ] timezone.get_zone
- [ ] timezone.get_offset
- [ ] timezone.set_zone
- [ ] timezone.zone_compare
There are unit tests but need to add integration tests
|
non_defect
|
add timezone windows tests to auto test suite this task is to monitor the addition of four tests to the auto test suite timezone get zone timezone get offset timezone set zone timezone zone compare there are unit tests but need to add integration tests
| 0
|
426,599
| 12,375,000,534
|
IssuesEvent
|
2020-05-19 03:18:26
|
mozilla/fx-private-relay
|
https://api.github.com/repos/mozilla/fx-private-relay
|
closed
|
Add the ability to see all their aliases from the addon
|
enhancement priority-P1 size:M 🧩 add-on
|
It's unclear which addresses have been created without going to the dashboard on the website.
Should also consider a "manage addresses" link that directs someone to the website dashboard to view and manage their created records.
|
1.0
|
Add the ability to see all their aliases from the addon - It's unclear which addresses have been created without going to the dashboard on the website.
Should also consider a "manage addresses" link that directs someone to the website dashboard to view and manage their created records.
|
non_defect
|
add the ability to see all their aliases from the addon it s unclear which addresses have been created without going to the dashboard on the website should also consider a manage addresses link that directs someone to the website dashboard to view and manage their created records
| 0
|
31,020
| 25,258,493,804
|
IssuesEvent
|
2022-11-15 20:21:26
|
pulibrary/dspace-development
|
https://api.github.com/repos/pulibrary/dspace-development
|
closed
|
DataSpace `dev` environment server is inaccessible
|
infrastructure dataspace
|
The following was reported:
> 2:00 If someone could look at whatever's going on with https://dataspace-dev.princeton.edu/ that would be amazing because I anticipate Lynn will be asking about it at the meeting, but otherwise I can take a look when I'm back.
I currently cannot access the application using https://dataspace-dev.princeton.edu/
|
1.0
|
DataSpace `dev` environment server is inaccessible - The following was reported:
> 2:00 If someone could look at whatever's going on with https://dataspace-dev.princeton.edu/ that would be amazing because I anticipate Lynn will be asking about it at the meeting, but otherwise I can take a look when I'm back.
I currently cannot access the application using https://dataspace-dev.princeton.edu/
|
non_defect
|
dataspace dev environment server is inaccessible the following was reported if someone could look at whatever s going on with that would be amazing because i anticipate lynn will be asking about it at the meeting but otherwise i can take a look when i m back i currently cannot access the application using
| 0
|
9,574
| 2,615,162,823
|
IssuesEvent
|
2015-03-01 06:42:15
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
|
auto-migrated Priority-Triage Type-Defect
|
```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
reaver 1.4
1. What operating system are you using (Linux is the only supported OS)
vmware workstation linux 2.6
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-62
4. What is the manufacturer and model # of the device you are trying to
crack?
tp-link 321g v4
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b xx:xx:xx:xx:xx -vv
6. Please describe what you think the issue is.
no idea
7. Paste the output from Reaver below.
[+] Waiting for beacon from XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
```
Original issue reported on code.google.com by `adsa...@gmail.com` on 6 Jun 2012 at 3:15
|
1.0
|
WARNING: Failed to associate with XX:XX:XX:XX:XX:XX - ```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
reaver 1.4
1. What operating system are you using (Linux is the only supported OS)
vmware workstation linux 2.6
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-62
4. What is the manufacturer and model # of the device you are trying to
crack?
tp-link 321g v4
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b xx:xx:xx:xx:xx -vv
6. Please describe what you think the issue is.
no idea
7. Paste the output from Reaver below.
[+] Waiting for beacon from XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
```
Original issue reported on code.google.com by `adsa...@gmail.com` on 6 Jun 2012 at 3:15
|
defect
|
warning failed to associate with xx xx xx xx xx xx a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os vmware workstation linux is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack tp link what is the entire command line string you are supplying to reaver reaver i b xx xx xx xx xx vv please describe what you think the issue is no idea paste the output from reaver below waiting for beacon from xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx original issue reported on code google com by adsa gmail com on jun at
| 1
|
48,559
| 13,135,248,517
|
IssuesEvent
|
2020-08-07 02:12:00
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
ExtraNodesetGenerator does not work with distributed mesh
|
T: defect
|
## Bug Description
When specify IDs of nodes to be added in the node set, nodes with these IDs may only be on one processor, thus make call `BoundaryInfo::add_node` error out.
## Steps to Reproduce
All current tests for this mesh generator are restricted with replicated mesh. NS module only use this to pin pressure at a node. If this cannot be fixed, those NS tests cannot run with distributed mesh. It might be an easy fix by protecting `add_node` call by querying if the node exist on the mesh.
## Impact
Enable tests using this mesh generator to run in parallel with distributed mesh.
|
1.0
|
ExtraNodesetGenerator does not work with distributed mesh - ## Bug Description
When specify IDs of nodes to be added in the node set, nodes with these IDs may only be on one processor, thus make call `BoundaryInfo::add_node` error out.
## Steps to Reproduce
All current tests for this mesh generator are restricted with replicated mesh. NS module only use this to pin pressure at a node. If this cannot be fixed, those NS tests cannot run with distributed mesh. It might be an easy fix by protecting `add_node` call by querying if the node exist on the mesh.
## Impact
Enable tests using this mesh generator to run in parallel with distributed mesh.
|
defect
|
extranodesetgenerator does not work with distributed mesh bug description when specify ids of nodes to be added in the node set nodes with these ids may only be on one processor thus make call boundaryinfo add node error out steps to reproduce all current tests for this mesh generator are restricted with replicated mesh ns module only use this to pin pressure at a node if this cannot be fixed those ns tests cannot run with distributed mesh it might be an easy fix by protecting add node call by querying if the node exist on the mesh impact enable tests using this mesh generator to run in parallel with distributed mesh
| 1
|
150,332
| 19,603,612,824
|
IssuesEvent
|
2022-01-06 06:07:21
|
snykiotcubedev/react-17.0.2
|
https://api.github.com/repos/snykiotcubedev/react-17.0.2
|
opened
|
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz
|
security vulnerability
|
## CVE-2020-7645 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary>
<p>Launch latest Chrome with the Devtools Protocol port open</p>
<p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p>
<p>
Dependency Hierarchy:
- :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/react-17.0.2/commit/c546697344431dcd6c04b4fea877c488ef3a6ad5">c546697344431dcd6c04b4fea877c488ef3a6ad5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems.
<p>Publish Date: 2020-05-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-7645">https://nvd.nist.gov/vuln/detail/CVE-2020-7645</a></p>
<p>Release Date: 2020-05-07</p>
<p>Fix Resolution: chrome-launcher - 0.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz - ## CVE-2020-7645 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary>
<p>Launch latest Chrome with the Devtools Protocol port open</p>
<p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p>
<p>
Dependency Hierarchy:
- :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/react-17.0.2/commit/c546697344431dcd6c04b4fea877c488ef3a6ad5">c546697344431dcd6c04b4fea877c488ef3a6ad5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems.
<p>Publish Date: 2020-05-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-7645">https://nvd.nist.gov/vuln/detail/CVE-2020-7645</a></p>
<p>Release Date: 2020-05-07</p>
<p>Fix Resolution: chrome-launcher - 0.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in chrome launcher tgz cve high severity vulnerability vulnerable library chrome launcher tgz launch latest chrome with the devtools protocol port open library home page a href dependency hierarchy x chrome launcher tgz vulnerable library found in head commit a href found in base branch main vulnerability details all versions of chrome launcher allow execution of arbitrary commands by controlling the home environment variable in linux operating systems publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution chrome launcher step up your open source security game with whitesource
| 0
|
164,512
| 20,364,519,507
|
IssuesEvent
|
2022-02-21 02:57:46
|
mxttwoods/blog
|
https://api.github.com/repos/mxttwoods/blog
|
closed
|
WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed
|
security vulnerability
|
## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-3.2.1.tgz (Root Library)
- webpack-dev-server-3.11.2.tgz
- selfsigned-1.10.8.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.2.1</p>
<p>Direct dependency fix Resolution (gatsby): 3.11.0-alpha-qe-sm.92</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed - ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-3.2.1.tgz (Root Library)
- webpack-dev-server-3.11.2.tgz
- selfsigned-1.10.8.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.2.1</p>
<p>Direct dependency fix Resolution (gatsby): 3.11.0-alpha-qe-sm.92</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in node forge tgz autoclosed ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy gatsby tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in base branch main vulnerability details the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution gatsby alpha qe sm step up your open source security game with whitesource
| 0
|
108,132
| 23,537,849,075
|
IssuesEvent
|
2022-08-20 00:24:04
|
happycube/ld-decode
|
https://api.github.com/repos/happycube/ld-decode
|
closed
|
ld-compress help points to old wiki location
|
ld-decode-tools deficiency
|
The wiki documentation for ld-compress has moved to it's own page now:
https://github.com/happycube/ld-decode/wiki/ld-compress
but the application is still pointing to the old location:
Output:
```
ld-compress --help
Usage: /usr/local/bin/ld-compress [-c] [-a] [-u] [-v] [-p] [-h] [-l <1-12>] [-g] file(s)
Modes:
-c Compress (default): Takes one or more .lds files and compresses them to .ldf files in the current directory.
-u Uncompress: Takes one or more .ldf/.raw.oga files and uncompresses them to .lds files in the current directory.
-a GPU Acceleration. Uses OpenCL or CUDA to accelerate encoding. See https://github.com/happycube/ld-decode/wiki/ld-decode-utilities
-v Verify: Returns md5 checksums of the given .ldf/.raw.oga files and their contained .lds files for verification purposes.
Options
-p Progress: displays progress bars - requires pv to be installed.
-h Help: This dialog.
-l Compression level 1 - 12 (1 - 11 for GPU encoding). Default is 11 (10 for GPU). 6 is recommended for faster but fair compression.
-g Use .raw.oga extension instead of .ldf when compressing.
```
|
1.0
|
ld-compress help points to old wiki location - The wiki documentation for ld-compress has moved to it's own page now:
https://github.com/happycube/ld-decode/wiki/ld-compress
but the application is still pointing to the old location:
Output:
```
ld-compress --help
Usage: /usr/local/bin/ld-compress [-c] [-a] [-u] [-v] [-p] [-h] [-l <1-12>] [-g] file(s)
Modes:
-c Compress (default): Takes one or more .lds files and compresses them to .ldf files in the current directory.
-u Uncompress: Takes one or more .ldf/.raw.oga files and uncompresses them to .lds files in the current directory.
-a GPU Acceleration. Uses OpenCL or CUDA to accelerate encoding. See https://github.com/happycube/ld-decode/wiki/ld-decode-utilities
-v Verify: Returns md5 checksums of the given .ldf/.raw.oga files and their contained .lds files for verification purposes.
Options
-p Progress: displays progress bars - requires pv to be installed.
-h Help: This dialog.
-l Compression level 1 - 12 (1 - 11 for GPU encoding). Default is 11 (10 for GPU). 6 is recommended for faster but fair compression.
-g Use .raw.oga extension instead of .ldf when compressing.
```
|
non_defect
|
ld compress help points to old wiki location the wiki documentation for ld compress has moved to it s own page now but the application is still pointing to the old location output ld compress help usage usr local bin ld compress file s modes c compress default takes one or more lds files and compresses them to ldf files in the current directory u uncompress takes one or more ldf raw oga files and uncompresses them to lds files in the current directory a gpu acceleration uses opencl or cuda to accelerate encoding see v verify returns checksums of the given ldf raw oga files and their contained lds files for verification purposes options p progress displays progress bars requires pv to be installed h help this dialog l compression level for gpu encoding default is for gpu is recommended for faster but fair compression g use raw oga extension instead of ldf when compressing
| 0
|
42,550
| 11,134,925,344
|
IssuesEvent
|
2019-12-20 13:09:00
|
nanopb/nanopb
|
https://api.github.com/repos/nanopb/nanopb
|
closed
|
Passing spaces and # in protoc arguments does not work
|
Component-Generator Priority-Low Type-Defect
|
When I use the `-L` or `--library-include-format=_some_folder/pb.h`, I expect in my `.pb.h` file to have `#include <_some_folder/pb.h`. However the script generated only `_some_folder/pb.h` for the include, not wrapped in `#include <>`. When I try to instead to `--library-include-format=\"#include <_some_folder/pb.h>\", I get various parsing errors with nanopb & bash. Is this the intended behaviour of this flag? Becuase that wouldn't make much sense. Let me know if theres any code/other info you all need.
|
1.0
|
Passing spaces and # in protoc arguments does not work - When I use the `-L` or `--library-include-format=_some_folder/pb.h`, I expect in my `.pb.h` file to have `#include <_some_folder/pb.h`. However the script generated only `_some_folder/pb.h` for the include, not wrapped in `#include <>`. When I try to instead to `--library-include-format=\"#include <_some_folder/pb.h>\", I get various parsing errors with nanopb & bash. Is this the intended behaviour of this flag? Becuase that wouldn't make much sense. Let me know if theres any code/other info you all need.
|
defect
|
passing spaces and in protoc arguments does not work when i use the l or library include format some folder pb h i expect in my pb h file to have include when i try to instead to library include format include i get various parsing errors with nanopb bash is this the intended behaviour of this flag becuase that wouldn t make much sense let me know if theres any code other info you all need
| 1
|
41,404
| 10,440,130,101
|
IssuesEvent
|
2019-09-18 08:03:15
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Nine parser bugs / missing features
|
C: Parser E: All Editions P: Medium T: Defect
|
First; congratulations on the release of 3.12! I've been looking forward to seeing it :)
I've been using the `org.jooq.meta.extensions.ddl.DDLDatabase` generator to generate my code from SQL. So far I've had to manually keep two versions; one with the scripts that will be executed by Flyway on the mysql DB and another that is used by JOOQ to generate the code. This is an obvious potential source of bugs in addition to the extra work so I'd like to get rid of it.
I tried running the new 3.12 on my SQL migration scripts and ran into nine errors:
1. After clause in alter statement
```
Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [10:60] ...LE `example` ADD COLUMN `field3` varchar(32) NULL [*]AFTER `field2`;
```
My understanding is that the order of the fields can have an impact on the performance of the queries. But I don't really care that much. Removing the AFTER clause will make the parsing work.
EDIT: duplicate of #7241
2. Autoincrement in ALTER clause
```
Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:73] ...ADD COLUMN `example_id` int(10) unsigned NOT NULL [*]AUTO_INCREMENT
```
I don't see a workaround, and this feature is needed in my DB.
EDIT: Reported as #9129.
3. Duplicate foreign key name in alter statement
```
Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Token '(' expected: [14:46] ...ink`
[ERROR] ADD CONSTRAINT `fk_example_link` FOREIGN KEY [*]`fk_example_link` (`example_id`) REFERENCES `example` (`example_id`) ON DELETE C...
```
Workaround is to remove the second occurrence of the name.
EDIT: Reported as #9130
4. Alter statement with first clause
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:69] ...le` ADD COLUMN `field1` int(10) unsigned NOT NULL [*]FIRST;
```
Same as with AFTER clause.
EDIT: duplicate of #7241
5. Multiple statements in an ALTER statement
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [9:46] ...ple`
[ERROR] ADD COLUMN `field2` int(10) unsigned NOT NULL[*],
[ERROR] ADD COLUMN `field3` int(10) unsigned NOT NULL;
```
Workaround is to manually split things into individual statements.
EDIT: Duplicate of #8846.
6. A BIT field with a default
```
org.jooq.impl.ParserException: Token ')' expected: [3:35] ...BLE `example` (
`field` bit(1) NOT NULL DEFAULT b[*]'0'
);
```
Workaround is rewriting the SQL without the b and single quotes.
EDIT: Reported as #9131.
7. Creating an index
```
Error while exporting schema: SQL [create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index"))]; Unknown data type: "index_idx"; SQL statement:
[ERROR] create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index")) [50004-199]
```
I got the normal way of creating indexes to work with JOOQ. However I did run into an additional bug when the table name isn't quoted:
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Error while exporting schema: SQL [create index "index_idx" on example("index")]; Table "EXAMPLE" not found; SQL statement:
[ERROR] create index "index_idx" on example("index")
```
8. Disabling foreign-key-checks + foreign key to table further into the script
```
SET FOREIGN_KEY_CHECKS = 0;
...
```
The obvious workaround is to re-arrange the statements so that the needed tables and fields are available in the right order.
9. Using btree statement
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [6:54] ...ER TABLE `example` ADD PRIMARY KEY (`example_id`) [*]USING BTREE;
```
The workaround is simple here; just remove the USING BTREE part.
EDIT: Reported as #9132.
There is a strange tenth issue that I ran into where I think MySQL actually erroneously succeeds when it should fail. It has to do with deleting a primary key when it no longer exists, but because the clause is contained in a single ALTER statement MySQL allows it.
I created an example project here: https://github.com/codemonstur/jooq-generate-errors. Also looked through issue #8452 and couldn't find these. All SQL was generated using Schemasync.
|
1.0
|
Nine parser bugs / missing features - First; congratulations on the release of 3.12! I've been looking forward to seeing it :)
I've been using the `org.jooq.meta.extensions.ddl.DDLDatabase` generator to generate my code from SQL. So far I've had to manually keep two versions; one with the scripts that will be executed by Flyway on the mysql DB and another that is used by JOOQ to generate the code. This is an obvious potential source of bugs in addition to the extra work so I'd like to get rid of it.
I tried running the new 3.12 on my SQL migration scripts and ran into nine errors:
1. After clause in alter statement
```
Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [10:60] ...LE `example` ADD COLUMN `field3` varchar(32) NULL [*]AFTER `field2`;
```
My understanding is that the order of the fields can have an impact on the performance of the queries. But I don't really care that much. Removing the AFTER clause will make the parsing work.
EDIT: duplicate of #7241
2. Autoincrement in ALTER clause
```
Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:73] ...ADD COLUMN `example_id` int(10) unsigned NOT NULL [*]AUTO_INCREMENT
```
I don't see a workaround, and this feature is needed in my DB.
EDIT: Reported as #9129.
3. Duplicate foreign key name in alter statement
```
Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Token '(' expected: [14:46] ...ink`
[ERROR] ADD CONSTRAINT `fk_example_link` FOREIGN KEY [*]`fk_example_link` (`example_id`) REFERENCES `example` (`example_id`) ON DELETE C...
```
Workaround is to remove the second occurrence of the name.
EDIT: Reported as #9130
4. Alter statement with first clause
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:69] ...le` ADD COLUMN `field1` int(10) unsigned NOT NULL [*]FIRST;
```
Same as with AFTER clause.
EDIT: duplicate of #7241
5. Multiple statements in an ALTER statement
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [9:46] ...ple`
[ERROR] ADD COLUMN `field2` int(10) unsigned NOT NULL[*],
[ERROR] ADD COLUMN `field3` int(10) unsigned NOT NULL;
```
Workaround is to manually split things into individual statements.
EDIT: Duplicate of #8846.
6. A BIT field with a default
```
org.jooq.impl.ParserException: Token ')' expected: [3:35] ...BLE `example` (
`field` bit(1) NOT NULL DEFAULT b[*]'0'
);
```
Workaround is rewriting the SQL without the b and single quotes.
EDIT: Reported as #9131.
7. Creating an index
```
Error while exporting schema: SQL [create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index"))]; Unknown data type: "index_idx"; SQL statement:
[ERROR] create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index")) [50004-199]
```
I got the normal way of creating indexes to work with JOOQ. However I did run into an additional bug when the table name isn't quoted:
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Error while exporting schema: SQL [create index "index_idx" on example("index")]; Table "EXAMPLE" not found; SQL statement:
[ERROR] create index "index_idx" on example("index")
```
8. Disabling foreign-key-checks + foreign key to table further into the script
```
SET FOREIGN_KEY_CHECKS = 0;
...
```
The obvious workaround is to re-arrange the statements so that the needed tables and fields are available in the right order.
9. Using btree statement
```
[ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [6:54] ...ER TABLE `example` ADD PRIMARY KEY (`example_id`) [*]USING BTREE;
```
The workaround is simple here; just remove the USING BTREE part.
EDIT: Reported as #9132.
There is a strange tenth issue that I ran into where I think MySQL actually erroneously succeeds when it should fail. It has to do with deleting a primary key when it no longer exists, but because the clause is contained in a single ALTER statement MySQL allows it.
I created an example project here: https://github.com/codemonstur/jooq-generate-errors. Also looked through issue #8452 and couldn't find these. All SQL was generated using Schemasync.
|
defect
|
nine parser bugs missing features first congratulations on the release of i ve been looking forward to seeing it i ve been using the org jooq meta extensions ddl ddldatabase generator to generate my code from sql so far i ve had to manually keep two versions one with the scripts that will be executed by flyway on the mysql db and another that is used by jooq to generate the code this is an obvious potential source of bugs in addition to the extra work so i d like to get rid of it i tried running the new on my sql migration scripts and ran into nine errors after clause in alter statement error running jooq code generation tool unexpected token or missing query delimiter le example add column varchar null after my understanding is that the order of the fields can have an impact on the performance of the queries but i don t really care that much removing the after clause will make the parsing work edit duplicate of autoincrement in alter clause failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter add column example id int unsigned not null auto increment i don t see a workaround and this feature is needed in my db edit reported as duplicate foreign key name in alter statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool token expected ink add constraint fk example link foreign key fk example link example id references example example id on delete c workaround is to remove the second occurrence of the name edit reported as alter statement with first clause failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter le add column int unsigned not null first same as with after clause edit duplicate of multiple statements in an alter statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter ple add column int unsigned not null add column int unsigned not null workaround is to manually split things into individual statements edit duplicate of a bit field with a default org jooq impl parserexception token expected ble example field bit not null default b workaround is rewriting the sql without the b and single quotes edit reported as creating an index error while exporting schema sql unknown data type index idx sql statement create table example example id bigint not null auto increment index bigint not null primary key id index index idx index i got the normal way of creating indexes to work with jooq however i did run into an additional bug when the table name isn t quoted failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool error while exporting schema sql table example not found sql statement create index index idx on example index disabling foreign key checks foreign key to table further into the script set foreign key checks the obvious workaround is to re arrange the statements so that the needed tables and fields are available in the right order using btree statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter er table example add primary key example id using btree the workaround is simple here just remove the using btree part edit reported as there is a strange tenth issue that i ran into where i think mysql actually erroneously succeeds when it should fail it has to do with deleting a primary key when it no longer exists but because the clause is contained in a single alter statement mysql allows it i created an example project here also looked through issue and couldn t find these all sql was generated using schemasync
| 1
|
82,234
| 32,074,548,429
|
IssuesEvent
|
2023-09-25 10:07:49
|
vector-im/element-call
|
https://api.github.com/repos/vector-im/element-call
|
opened
|
Tiles near the bottom of the window in grid mode become un-clickable half way down.
|
T-Defect
|
### Steps to reproduce
1. Participate in a call in Grid mode
2. Resize the window if necessary to have tiles scrollable underneath the controls
### Outcome
#### What did you expect?
- Tiles not directly underneath the controls should be clickable.
- Tiles directly underneath the controls should be clickable outside the bounds of the control buttons.
#### What happened instead?
There appears to be an invisible bar housing the controls that swallows up the mouse focus.
https://github.com/vector-im/element-call/assets/6060466/ee25753b-9b65-469b-a1d4-08c43fa81689
### Operating system
macOS
### Browser information
Firefox & Safari.
### URL for webapp
call.element.io
### Will you send logs?
No
|
1.0
|
Tiles near the bottom of the window in grid mode become un-clickable half way down. - ### Steps to reproduce
1. Participate in a call in Grid mode
2. Resize the window if necessary to have tiles scrollable underneath the controls
### Outcome
#### What did you expect?
- Tiles not directly underneath the controls should be clickable.
- Tiles directly underneath the controls should be clickable outside the bounds of the control buttons.
#### What happened instead?
There appears to be an invisible bar housing the controls that swallows up the mouse focus.
https://github.com/vector-im/element-call/assets/6060466/ee25753b-9b65-469b-a1d4-08c43fa81689
### Operating system
macOS
### Browser information
Firefox & Safari.
### URL for webapp
call.element.io
### Will you send logs?
No
|
defect
|
tiles near the bottom of the window in grid mode become un clickable half way down steps to reproduce participate in a call in grid mode resize the window if necessary to have tiles scrollable underneath the controls outcome what did you expect tiles not directly underneath the controls should be clickable tiles directly underneath the controls should be clickable outside the bounds of the control buttons what happened instead there appears to be an invisible bar housing the controls that swallows up the mouse focus operating system macos browser information firefox safari url for webapp call element io will you send logs no
| 1
|
621,248
| 19,581,527,781
|
IssuesEvent
|
2022-01-04 22:05:08
|
paperclipui/paperclip
|
https://api.github.com/repos/paperclipui/paperclip
|
closed
|
[MTI] diffing sprint
|
priority: high effort: medium impact: high
|
TODOs:
- [x] generate HTML report for diffed images
- [ ] include coverage information for each screen? - should navigate to coverage report
- [x] watch mode for visual diffing
- [ ] **ability to lock changes to UI screens**
- [ ] figure out how to store artifacts somewhere, and pull them down to be analyzed (s3?)
- [ ] ability to set threshold limits
|
1.0
|
[MTI] diffing sprint - TODOs:
- [x] generate HTML report for diffed images
- [ ] include coverage information for each screen? - should navigate to coverage report
- [x] watch mode for visual diffing
- [ ] **ability to lock changes to UI screens**
- [ ] figure out how to store artifacts somewhere, and pull them down to be analyzed (s3?)
- [ ] ability to set threshold limits
|
non_defect
|
diffing sprint todos generate html report for diffed images include coverage information for each screen should navigate to coverage report watch mode for visual diffing ability to lock changes to ui screens figure out how to store artifacts somewhere and pull them down to be analyzed ability to set threshold limits
| 0
|
75,152
| 7,460,860,694
|
IssuesEvent
|
2018-03-30 21:45:29
|
Microsoft/PTVS
|
https://api.github.com/repos/Microsoft/PTVS
|
opened
|
PythonToolsMockTests fail with MEF composition issues
|
Tests
|
```cs
System.ComponentModel.Composition.ImportCardinalityMismatchException:
No exports were found that match the constraint:
ContractName Microsoft.PythonTools.Editor.PythonEditorServices
RequiredTypeIdentity Microsoft.PythonTools.Editor.PythonEditorServices
```
|
1.0
|
PythonToolsMockTests fail with MEF composition issues - ```cs
System.ComponentModel.Composition.ImportCardinalityMismatchException:
No exports were found that match the constraint:
ContractName Microsoft.PythonTools.Editor.PythonEditorServices
RequiredTypeIdentity Microsoft.PythonTools.Editor.PythonEditorServices
```
|
non_defect
|
pythontoolsmocktests fail with mef composition issues cs system componentmodel composition importcardinalitymismatchexception no exports were found that match the constraint contractname microsoft pythontools editor pythoneditorservices requiredtypeidentity microsoft pythontools editor pythoneditorservices
| 0
|
79,333
| 15,586,130,695
|
IssuesEvent
|
2021-03-18 01:14:30
|
revfactory/task-api
|
https://api.github.com/repos/revfactory/task-api
|
opened
|
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.16.jar
|
security vulnerability
|
## CVE-2020-11996 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: task-api/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.3.RELEASE.jar
- :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive.
<p>Publish Date: 2020-06-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p>
<p>Release Date: 2020-06-26</p>
<p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.16.jar - ## CVE-2020-11996 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: task-api/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.3.RELEASE.jar
- :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive.
<p>Publish Date: 2020-06-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p>
<p>Release Date: 2020-06-26</p>
<p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file task api build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details a specially crafted sequence of http requests sent to apache tomcat to to and to could trigger high cpu usage for several seconds if a sufficient number of such requests were made on concurrent http connections the server could become unresponsive publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat coyote org apache tomcat embed org apache tomcat embed step up your open source security game with whitesource
| 0
|
88,755
| 25,497,747,821
|
IssuesEvent
|
2022-11-27 21:55:00
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
reopened
|
@fluentui/e2e: Cypress command produces wrong paths for tests
|
Type: Bug :bug: Area: Build System Resolution: Soft Close
|
### Environment Information
- **Package version(s)**: `master`
- **Browser and OS versions**: Windows
### Please provide a reproduction of the bug in a codepen:
This happens only on Windows as custom command handles only Unix-like paths.
https://github.com/microsoft/fluentui/blob/dab45a2afe609709735b9b6e604a1af40d50e809/packages/fluentui/e2e/cypress/support/commands.js#L3-L9
```
// Cypress passes to a custom command
const testFilePath = tests\carouselClickableContent.spec.ts
```
Then it produces a wrong path that contains `tests-`, for example `tests-carousel-clickable-content` instead of `carousel-clickable-content`. As requested page does not exist, everything fails 💣
#### Actual behavior:
```
1) Popup - on ESC key press
"before each" hook for "is not closed if ESC is handled by children":
Invariant Violation: The following error originated from your application code, not from Cypress.
> Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports.
Check the render method of `E2EExample`.
When Cypress detects uncaught errors originating from your application it will automatically fail the current test.
This behavior is configurable, and you can choose to turn this off by listening to the `uncaught:exception` event.
https://on.cypress.io/uncaught-exception-from-application
```
#### Expected behavior:
It works.
### Priorities and help requested:
Are you willing to submit a PR to fix? No
Requested priority: Normal
|
1.0
|
@fluentui/e2e: Cypress command produces wrong paths for tests - ### Environment Information
- **Package version(s)**: `master`
- **Browser and OS versions**: Windows
### Please provide a reproduction of the bug in a codepen:
This happens only on Windows as custom command handles only Unix-like paths.
https://github.com/microsoft/fluentui/blob/dab45a2afe609709735b9b6e604a1af40d50e809/packages/fluentui/e2e/cypress/support/commands.js#L3-L9
```
// Cypress passes to a custom command
const testFilePath = tests\carouselClickableContent.spec.ts
```
Then it produces a wrong path that contains `tests-`, for example `tests-carousel-clickable-content` instead of `carousel-clickable-content`. As requested page does not exist, everything fails 💣
#### Actual behavior:
```
1) Popup - on ESC key press
"before each" hook for "is not closed if ESC is handled by children":
Invariant Violation: The following error originated from your application code, not from Cypress.
> Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports.
Check the render method of `E2EExample`.
When Cypress detects uncaught errors originating from your application it will automatically fail the current test.
This behavior is configurable, and you can choose to turn this off by listening to the `uncaught:exception` event.
https://on.cypress.io/uncaught-exception-from-application
```
#### Expected behavior:
It works.
### Priorities and help requested:
Are you willing to submit a PR to fix? No
Requested priority: Normal
|
non_defect
|
fluentui cypress command produces wrong paths for tests environment information package version s master browser and os versions windows please provide a reproduction of the bug in a codepen this happens only on windows as custom command handles only unix like paths cypress passes to a custom command const testfilepath tests carouselclickablecontent spec ts then it produces a wrong path that contains tests for example tests carousel clickable content instead of carousel clickable content as requested page does not exist everything fails 💣 actual behavior popup on esc key press before each hook for is not closed if esc is handled by children invariant violation the following error originated from your application code not from cypress element type is invalid expected a string for built in components or a class function for composite components but got undefined you likely forgot to export your component from the file it s defined in or you might have mixed up default and named imports check the render method of when cypress detects uncaught errors originating from your application it will automatically fail the current test this behavior is configurable and you can choose to turn this off by listening to the uncaught exception event expected behavior it works priorities and help requested are you willing to submit a pr to fix no requested priority normal
| 0
|
52,096
| 7,747,842,358
|
IssuesEvent
|
2018-05-30 05:57:41
|
benwasd/ssah
|
https://api.github.com/repos/benwasd/ssah
|
closed
|
Workshops mit Fachpersonen
|
documentation
|
Use Cases besprechen mit und Verbesserungen einpflegen. Feedback dokumentieren.
- Administration der Skischule SnowValley Linden.
- Martina, Administration Skischule Saas-Fee
|
1.0
|
Workshops mit Fachpersonen - Use Cases besprechen mit und Verbesserungen einpflegen. Feedback dokumentieren.
- Administration der Skischule SnowValley Linden.
- Martina, Administration Skischule Saas-Fee
|
non_defect
|
workshops mit fachpersonen use cases besprechen mit und verbesserungen einpflegen feedback dokumentieren administration der skischule snowvalley linden martina administration skischule saas fee
| 0
|
27,391
| 13,239,087,655
|
IssuesEvent
|
2020-08-19 02:23:01
|
radis/radis
|
https://api.github.com/repos/radis/radis
|
closed
|
Improve definition of width-grid resolution (res_L, res_G)
|
performance physics
|
With the manuscript draft being updated, we now have a much better handle on how to precisely define the errors in the width-grid for the "DLM"-approach. It would be good to propagate these updates to the RADIS code, but some executive decisions have to be made on how to implement it.
There are basically three candidates of how the user could specify the desired error:
- **`p`** -> _relative increase between two width-gridpoints_
- **`log_p`** = np.log(p) -> _actual step-size of the logarithmic width-grid_
- **`err_rms`** = RMS[Y_i]*np.log(p)/12 -> _contribution to RMS-error by a single line._
(The factor 1/12 = 1/2 x 1/6, where 1/6 comes from integrating the grid-alignment assuming uniform distribution between 0 and 1)
Each has their own advantage, namely:
- **`p`**: unambiguous what is meant, and makes sense from a human perspective (e.g. `p = 1.2` means 20% increase between gridpoints)
- **`log_p`**: equally unambiguous, but perhaps makes less sense from a human perspective (e.g. what really does `log_p = 0.2` mean?). Advantage is that when `log_p` is closer to zero, it means higher resolution, which perhaps is more intuitive.
- **`err_rms`**: Makes most sense from human point of view, and resolution also gets better as `err_rms` gets closer to zero. Problem is that choosing the RMS-error as the defining error measure is somewhat arbitrary, and depending on the kind of weight used (1,2, or 3), the same value of `err_rms` will lead to different `p` and `log_p` values, which the user may not be aware of (but perhaps equally possibly they do not care about)
Anyway so that's a decision that has to be made. My personal preference is in the presented order (with **`p`** most favorable), but of course I'd like to hear everyone's thoughts.
|
True
|
Improve definition of width-grid resolution (res_L, res_G) - With the manuscript draft being updated, we now have a much better handle on how to precisely define the errors in the width-grid for the "DLM"-approach. It would be good to propagate these updates to the RADIS code, but some executive decisions have to be made on how to implement it.
There are basically three candidates of how the user could specify the desired error:
- **`p`** -> _relative increase between two width-gridpoints_
- **`log_p`** = np.log(p) -> _actual step-size of the logarithmic width-grid_
- **`err_rms`** = RMS[Y_i]*np.log(p)/12 -> _contribution to RMS-error by a single line._
(The factor 1/12 = 1/2 x 1/6, where 1/6 comes from integrating the grid-alignment assuming uniform distribution between 0 and 1)
Each has their own advantage, namely:
- **`p`**: unambiguous what is meant, and makes sense from a human perspective (e.g. `p = 1.2` means 20% increase between gridpoints)
- **`log_p`**: equally unambiguous, but perhaps makes less sense from a human perspective (e.g. what really does `log_p = 0.2` mean?). Advantage is that when `log_p` is closer to zero, it means higher resolution, which perhaps is more intuitive.
- **`err_rms`**: Makes most sense from human point of view, and resolution also gets better as `err_rms` gets closer to zero. Problem is that choosing the RMS-error as the defining error measure is somewhat arbitrary, and depending on the kind of weight used (1,2, or 3), the same value of `err_rms` will lead to different `p` and `log_p` values, which the user may not be aware of (but perhaps equally possibly they do not care about)
Anyway so that's a decision that has to be made. My personal preference is in the presented order (with **`p`** most favorable), but of course I'd like to hear everyone's thoughts.
|
non_defect
|
improve definition of width grid resolution res l res g with the manuscript draft being updated we now have a much better handle on how to precisely define the errors in the width grid for the dlm approach it would be good to propagate these updates to the radis code but some executive decisions have to be made on how to implement it there are basically three candidates of how the user could specify the desired error p relative increase between two width gridpoints log p np log p actual step size of the logarithmic width grid err rms rms np log p contribution to rms error by a single line the factor x where comes from integrating the grid alignment assuming uniform distribution between and each has their own advantage namely p unambiguous what is meant and makes sense from a human perspective e g p means increase between gridpoints log p equally unambiguous but perhaps makes less sense from a human perspective e g what really does log p mean advantage is that when log p is closer to zero it means higher resolution which perhaps is more intuitive err rms makes most sense from human point of view and resolution also gets better as err rms gets closer to zero problem is that choosing the rms error as the defining error measure is somewhat arbitrary and depending on the kind of weight used or the same value of err rms will lead to different p and log p values which the user may not be aware of but perhaps equally possibly they do not care about anyway so that s a decision that has to be made my personal preference is in the presented order with p most favorable but of course i d like to hear everyone s thoughts
| 0
|
29,405
| 5,682,679,605
|
IssuesEvent
|
2017-04-13 10:22:42
|
bancika/diy-layout-creator
|
https://api.github.com/repos/bancika/diy-layout-creator
|
closed
|
Trace mask -> export to png doesn't add .png to filename
|
auto-migrated Priority-Medium Type-Defect
|
```
When you export (at least) trace mask to png and just type the filename part
the program won't add the extension part (.png).
This might be true for all exports but I've only tested traces to png.
```
Original issue reported on code.google.com by `per.tre...@gmail.com` on 25 Jan 2011 at 10:26
|
1.0
|
Trace mask -> export to png doesn't add .png to filename - ```
When you export (at least) trace mask to png and just type the filename part
the program won't add the extension part (.png).
This might be true for all exports but I've only tested traces to png.
```
Original issue reported on code.google.com by `per.tre...@gmail.com` on 25 Jan 2011 at 10:26
|
defect
|
trace mask export to png doesn t add png to filename when you export at least trace mask to png and just type the filename part the program won t add the extension part png this might be true for all exports but i ve only tested traces to png original issue reported on code google com by per tre gmail com on jan at
| 1
|
42,501
| 11,092,169,162
|
IssuesEvent
|
2019-12-15 17:10:21
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
sparse eigs (arpack) shift-invert drops the smallest eigenvalue for some k
|
defect scipy.sparse.linalg
|
scipy.sparse.linalg.eigs shift-invert mode for the SM, smallest-magnitude eigenvalue
`( A=sparse.rand(), k=k, which='LM', sigma=0 )`
drops the smallest one for some k:
k 1 shift: [ 17 +0j ]
k 2 shift: [ 17 +0j 52 +0j ]
k 3 shift: [ 52 +0j -23 -47j -23 +47j ]
k 4 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j ]
k 5 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j ]
k 6 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j ]
k 7 shift: [ 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j 1e+02 -36j 1e+02 +36j ]
It's hard to believe that this is an ARPACK bug; what else could be wrong ?
Versions: numpy 1.17.4 scipy 1.3.3 python 3.7.3 macos 10.10.5
3581068 24 Nov 18:25 .../libopenblasp-r0.3.7.dev.dylib
669500 24 Nov 18:25 .../_arpack.cpython-37m-darwin.so
Notes:
The sometimes-missing smallest eigenvalue from `sparse.rand` with various n, density, seed is, so far, real.
I take `v0=ones`, not random -- this should be the default.
The nice scipy arpack tutorial says "which = 'SM' may lead to slow execution time and/or anomalous results". That's too weak --in fact SM often fails on many `sparse.rand` matrices.
#### Reproducing code example:
<!--
If you place your code between the triple backticks below,
it will be rendered as a code block.
-->
#!/usr/bin/env python
""" test scipy.sparse.linalg.eigs LM shift-invert: some k drop the smallest real eigenvalue ? """
# https://docs.scipy.org/doc/scipy/reference/tutorial/arpack.html X.XT symmetric, easy
# which = 'SM' may lead to slow execution time and/or anomalous results
# https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.eigs.html
# denis 10 Dec 2019
from __future__ import division, print_function
import sys
import numpy as np
import scipy
from scipy import sparse
from scipy.sparse.linalg import eigs, ArpackError, ArpackNoConvergence
try:
import scikits.umfpack # eigs -> splu spsolve ?
print( "umfpack:", scikits.umfpack.__version__ )
except ImportError:
pass
def randomsparse( n, density=0.1, seed=0 ):
""" -> n x n scipy.sparse.rand csc, uniform -1 .. 1 """
random = np.random.RandomState( seed=seed )
A = sparse.rand( n, n, density, format="csc", random_state=random ) # uniform 0 .. 1
# if distrib:
A.data *= 2
A.data -= 1
print( "randomsparse: n %d density %.2g seed %d A %.2g .. %.2g nnz %d " % (
n, density, seed, A.data.min(), A.data.max(), A.nnz ))
return A
def zsort( z ):
""" sort z abs increasing, then imag """
jsort = np.lexsort( [np.imag(z), np.abs(z)] )
return z[jsort]
np.set_printoptions( threshold=20, edgeitems=10, linewidth=140,
formatter = dict( float = lambda x: "%.2g" % x, # float arrays %.2g
complexfloat = " {0.real:.2g} {0.imag:+.2g}j " .format ))
print( "\n" + 80 * "=" )
print( "versions: numpy %s scipy %s python %s " % (
np.__version__, scipy.__version__ , sys.version.split()[0] ))
#...............................................................................
# gen --
gen = randomsparse
n = 10
density = .5
seed = 0 # some -> LinAlgWarning singular, lu_factor
dense = True # test arpack dense first
# eigs --
kmax = 8
sigma = 0
tol = 0 # 0: machine precision, eps ?
v0 = 1 # None: random, 1: ones
SM = False
# to change these params, run this.py a=1 b=None 'c = expr' ... in sh or ipython --
for arg in sys.argv[1:]:
exec( arg )
genparams = "%s n %d density %.2g seed %d dense %s " % (
gen.__name__, n, density, seed, dense )
eigsparams = "sigma %s tol %g v0 %s SM %d " % (
sigma, tol, v0, SM )
print( "gen params:", genparams )
print( "eigs params:", eigsparams )
kmax = min( kmax, n - 2 )
if v0 is not None:
v0 = np.ones( n )
#...............................................................................
# generate a test matrix A --
# may be singular for some seeds: arpack.py sometimes calls lu_factor, SM doesn't
A = randomsparse( n, density=density, seed=seed )
A *= 100
if dense: # test arpack dense first
A = A.A
#...............................................................................
print( "\n# scipy.sparse.linalg.eigs (arpack) shift-invert vs. SM" )
for k in range( 1, kmax+1 ):
evals = eigs( A, k=k, which='LM', sigma=0, v0=v0, maxiter=None, tol=tol,
return_eigenvectors=False )
# eigs(A, k=6, M=None, sigma=None, which='LM', v0=None, maxiter=None, tol=0,
# return_eigenvectors=True, Minv=None, OPinv=None, OPpart=None)
print( "k %d shift: %s" % (k, zsort( evals )))
if SM:
for k in range( 1, kmax+1 ):
try:
evals = eigs( A, k=k, which='SM', v0=v0, maxiter=None, tol=tol,
return_eigenvectors=False )
print( "k %d SM: %s" % (k, zsort( evals )[0] ))
except (ArpackError, ArpackNoConvergence) as msg:
print( "k %d SM: %s" % (k, msg) )
|
1.0
|
sparse eigs (arpack) shift-invert drops the smallest eigenvalue for some k - scipy.sparse.linalg.eigs shift-invert mode for the SM, smallest-magnitude eigenvalue
`( A=sparse.rand(), k=k, which='LM', sigma=0 )`
drops the smallest one for some k:
k 1 shift: [ 17 +0j ]
k 2 shift: [ 17 +0j 52 +0j ]
k 3 shift: [ 52 +0j -23 -47j -23 +47j ]
k 4 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j ]
k 5 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j ]
k 6 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j ]
k 7 shift: [ 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j 1e+02 -36j 1e+02 +36j ]
It's hard to believe that this is an ARPACK bug; what else could be wrong ?
Versions: numpy 1.17.4 scipy 1.3.3 python 3.7.3 macos 10.10.5
3581068 24 Nov 18:25 .../libopenblasp-r0.3.7.dev.dylib
669500 24 Nov 18:25 .../_arpack.cpython-37m-darwin.so
Notes:
The sometimes-missing smallest eigenvalue from `sparse.rand` with various n, density, seed is, so far, real.
I take `v0=ones`, not random -- this should be the default.
The nice scipy arpack tutorial says "which = 'SM' may lead to slow execution time and/or anomalous results". That's too weak --in fact SM often fails on many `sparse.rand` matrices.
#### Reproducing code example:
<!--
If you place your code between the triple backticks below,
it will be rendered as a code block.
-->
#!/usr/bin/env python
""" test scipy.sparse.linalg.eigs LM shift-invert: some k drop the smallest real eigenvalue ? """
# https://docs.scipy.org/doc/scipy/reference/tutorial/arpack.html X.XT symmetric, easy
# which = 'SM' may lead to slow execution time and/or anomalous results
# https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.eigs.html
# denis 10 Dec 2019
from __future__ import division, print_function
import sys
import numpy as np
import scipy
from scipy import sparse
from scipy.sparse.linalg import eigs, ArpackError, ArpackNoConvergence
try:
import scikits.umfpack # eigs -> splu spsolve ?
print( "umfpack:", scikits.umfpack.__version__ )
except ImportError:
pass
def randomsparse( n, density=0.1, seed=0 ):
""" -> n x n scipy.sparse.rand csc, uniform -1 .. 1 """
random = np.random.RandomState( seed=seed )
A = sparse.rand( n, n, density, format="csc", random_state=random ) # uniform 0 .. 1
# if distrib:
A.data *= 2
A.data -= 1
print( "randomsparse: n %d density %.2g seed %d A %.2g .. %.2g nnz %d " % (
n, density, seed, A.data.min(), A.data.max(), A.nnz ))
return A
def zsort( z ):
""" sort z abs increasing, then imag """
jsort = np.lexsort( [np.imag(z), np.abs(z)] )
return z[jsort]
np.set_printoptions( threshold=20, edgeitems=10, linewidth=140,
formatter = dict( float = lambda x: "%.2g" % x, # float arrays %.2g
complexfloat = " {0.real:.2g} {0.imag:+.2g}j " .format ))
print( "\n" + 80 * "=" )
print( "versions: numpy %s scipy %s python %s " % (
np.__version__, scipy.__version__ , sys.version.split()[0] ))
#...............................................................................
# gen --
gen = randomsparse
n = 10
density = .5
seed = 0 # some -> LinAlgWarning singular, lu_factor
dense = True # test arpack dense first
# eigs --
kmax = 8
sigma = 0
tol = 0 # 0: machine precision, eps ?
v0 = 1 # None: random, 1: ones
SM = False
# to change these params, run this.py a=1 b=None 'c = expr' ... in sh or ipython --
for arg in sys.argv[1:]:
exec( arg )
genparams = "%s n %d density %.2g seed %d dense %s " % (
gen.__name__, n, density, seed, dense )
eigsparams = "sigma %s tol %g v0 %s SM %d " % (
sigma, tol, v0, SM )
print( "gen params:", genparams )
print( "eigs params:", eigsparams )
kmax = min( kmax, n - 2 )
if v0 is not None:
v0 = np.ones( n )
#...............................................................................
# generate a test matrix A --
# may be singular for some seeds: arpack.py sometimes calls lu_factor, SM doesn't
A = randomsparse( n, density=density, seed=seed )
A *= 100
if dense: # test arpack dense first
A = A.A
#...............................................................................
print( "\n# scipy.sparse.linalg.eigs (arpack) shift-invert vs. SM" )
for k in range( 1, kmax+1 ):
evals = eigs( A, k=k, which='LM', sigma=0, v0=v0, maxiter=None, tol=tol,
return_eigenvectors=False )
# eigs(A, k=6, M=None, sigma=None, which='LM', v0=None, maxiter=None, tol=0,
# return_eigenvectors=True, Minv=None, OPinv=None, OPpart=None)
print( "k %d shift: %s" % (k, zsort( evals )))
if SM:
for k in range( 1, kmax+1 ):
try:
evals = eigs( A, k=k, which='SM', v0=v0, maxiter=None, tol=tol,
return_eigenvectors=False )
print( "k %d SM: %s" % (k, zsort( evals )[0] ))
except (ArpackError, ArpackNoConvergence) as msg:
print( "k %d SM: %s" % (k, msg) )
|
defect
|
sparse eigs arpack shift invert drops the smallest eigenvalue for some k scipy sparse linalg eigs shift invert mode for the sm smallest magnitude eigenvalue a sparse rand k k which lm sigma drops the smallest one for some k k shift k shift k shift k shift k shift k shift k shift it s hard to believe that this is an arpack bug what else could be wrong versions numpy scipy python macos nov libopenblasp dev dylib nov arpack cpython darwin so notes the sometimes missing smallest eigenvalue from sparse rand with various n density seed is so far real i take ones not random this should be the default the nice scipy arpack tutorial says which sm may lead to slow execution time and or anomalous results that s too weak in fact sm often fails on many sparse rand matrices reproducing code example if you place your code between the triple backticks below it will be rendered as a code block usr bin env python test scipy sparse linalg eigs lm shift invert some k drop the smallest real eigenvalue x xt symmetric easy which sm may lead to slow execution time and or anomalous results denis dec from future import division print function import sys import numpy as np import scipy from scipy import sparse from scipy sparse linalg import eigs arpackerror arpacknoconvergence try import scikits umfpack eigs splu spsolve print umfpack scikits umfpack version except importerror pass def randomsparse n density seed n x n scipy sparse rand csc uniform random np random randomstate seed seed a sparse rand n n density format csc random state random uniform if distrib a data a data print randomsparse n d density seed d a nnz d n density seed a data min a data max a nnz return a def zsort z sort z abs increasing then imag jsort np lexsort return z np set printoptions threshold edgeitems linewidth formatter dict float lambda x x float arrays complexfloat real imag j format print n print versions numpy s scipy s python s np version scipy version sys version split gen gen randomsparse n density seed some linalgwarning singular lu factor dense true test arpack dense first eigs kmax sigma tol machine precision eps none random ones sm false to change these params run this py a b none c expr in sh or ipython for arg in sys argv exec arg genparams s n d density seed d dense s gen name n density seed dense eigsparams sigma s tol g s sm d sigma tol sm print gen params genparams print eigs params eigsparams kmax min kmax n if is not none np ones n generate a test matrix a may be singular for some seeds arpack py sometimes calls lu factor sm doesn t a randomsparse n density density seed seed a if dense test arpack dense first a a a print n scipy sparse linalg eigs arpack shift invert vs sm for k in range kmax evals eigs a k k which lm sigma maxiter none tol tol return eigenvectors false eigs a k m none sigma none which lm none maxiter none tol return eigenvectors true minv none opinv none oppart none print k d shift s k zsort evals if sm for k in range kmax try evals eigs a k k which sm maxiter none tol tol return eigenvectors false print k d sm s k zsort evals except arpackerror arpacknoconvergence as msg print k d sm s k msg
| 1
|
19,513
| 3,218,564,272
|
IssuesEvent
|
2015-10-08 02:32:13
|
darkfeline/python-abx
|
https://api.github.com/repos/darkfeline/python-abx
|
closed
|
ImportError: No module named gst
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Download source zip and extract
2. ./abx.py
What is the expected output?
An ABX comparator
What do you see instead?
Traceback (most recent call last):
File "./abx.py", line 6, in <module>
import gst
ImportError: No module named gst
What version of the product are you using?
2012-05-26
On what operating system?
Mint 17 Cinnamon 64-bit live
Please provide any additional information below.
Python 2.7.6
```
Original issue reported on code.google.com by `southi...@gmail.com` on 1 Mar 2015 at 4:53
|
1.0
|
ImportError: No module named gst - ```
What steps will reproduce the problem?
1. Download source zip and extract
2. ./abx.py
What is the expected output?
An ABX comparator
What do you see instead?
Traceback (most recent call last):
File "./abx.py", line 6, in <module>
import gst
ImportError: No module named gst
What version of the product are you using?
2012-05-26
On what operating system?
Mint 17 Cinnamon 64-bit live
Please provide any additional information below.
Python 2.7.6
```
Original issue reported on code.google.com by `southi...@gmail.com` on 1 Mar 2015 at 4:53
|
defect
|
importerror no module named gst what steps will reproduce the problem download source zip and extract abx py what is the expected output an abx comparator what do you see instead traceback most recent call last file abx py line in import gst importerror no module named gst what version of the product are you using on what operating system mint cinnamon bit live please provide any additional information below python original issue reported on code google com by southi gmail com on mar at
| 1
|
406,269
| 27,557,434,044
|
IssuesEvent
|
2023-03-07 19:04:25
|
pharmaverse/admiral
|
https://api.github.com/repos/pharmaverse/admiral
|
closed
|
Documentation: derive_vars_duration additional example
|
documentation good first issue release Q1-2023
|
derive_vars_duration start_date and end_date inputs can be counter-intuitive when start_date needs to be an end data variable and end_date needs to be a start date variable.
"Last Dose Relative Time" LDRELTM requires start_date = LDOSEDTM, end_date = ASTDTM so that LDRELTM will be positive (unless there are data issues). Suggestion is to include an example for "Derive adverse event start time since last dose in hours" showing missing data examples.
```
data <- tribble(
~USUBJID, ~ASTDTM, ~LDOSEDTM,
"P01", ymd_hms("2019-08-09T04:30:56"), ymd_hms("2019-08-08T10:05:00"),
"P02", ymd_hms("2019-11-11T23:59:59"), ymd_hms("2019-10-11T11:37:00"),
"P03", ymd_hms("2019-11-11T00:00:00"), ymd_hms("2019-11-10T23:59:59"),
"P04", ymd_hms("2019-11-11T12:34:56"),NA,
"P05", NA, ymd_hms("2019-09-28T12:34:56")
)
derive_vars_duration(
data,
new_var = LDRELTM,
new_var_unit = LDRELTMU,
start_date = LDOSEDTM,
end_date = ASTDTM,
in_unit = "hours",
out_unit = "hours",
add_one = FALSE
)
```
|
1.0
|
Documentation: derive_vars_duration additional example - derive_vars_duration start_date and end_date inputs can be counter-intuitive when start_date needs to be an end data variable and end_date needs to be a start date variable.
"Last Dose Relative Time" LDRELTM requires start_date = LDOSEDTM, end_date = ASTDTM so that LDRELTM will be positive (unless there are data issues). Suggestion is to include an example for "Derive adverse event start time since last dose in hours" showing missing data examples.
```
data <- tribble(
~USUBJID, ~ASTDTM, ~LDOSEDTM,
"P01", ymd_hms("2019-08-09T04:30:56"), ymd_hms("2019-08-08T10:05:00"),
"P02", ymd_hms("2019-11-11T23:59:59"), ymd_hms("2019-10-11T11:37:00"),
"P03", ymd_hms("2019-11-11T00:00:00"), ymd_hms("2019-11-10T23:59:59"),
"P04", ymd_hms("2019-11-11T12:34:56"),NA,
"P05", NA, ymd_hms("2019-09-28T12:34:56")
)
derive_vars_duration(
data,
new_var = LDRELTM,
new_var_unit = LDRELTMU,
start_date = LDOSEDTM,
end_date = ASTDTM,
in_unit = "hours",
out_unit = "hours",
add_one = FALSE
)
```
|
non_defect
|
documentation derive vars duration additional example derive vars duration start date and end date inputs can be counter intuitive when start date needs to be an end data variable and end date needs to be a start date variable last dose relative time ldreltm requires start date ldosedtm end date astdtm so that ldreltm will be positive unless there are data issues suggestion is to include an example for derive adverse event start time since last dose in hours showing missing data examples data tribble usubjid astdtm ldosedtm ymd hms ymd hms ymd hms ymd hms ymd hms ymd hms ymd hms na na ymd hms derive vars duration data new var ldreltm new var unit ldreltmu start date ldosedtm end date astdtm in unit hours out unit hours add one false
| 0
|
33,425
| 7,121,468,065
|
IssuesEvent
|
2018-01-19 07:54:08
|
MethodicalAcceleratorDesign/MAD-X
|
https://api.github.com/repos/MethodicalAcceleratorDesign/MAD-X
|
closed
|
AC-dipole in PTC but not in convertor
|
defect
|
*Issue migrated from trac ticket # 108*
**component:** ptc_proper | **priority:** major
___
#### 2012-01-30 13:50:56: frank.schmidt commented
date: 2011.11.19
The AC-dipole is in PTC and lacking in the convertor.
___
#### 2012-01-30 13:51:16: @ldeniau changed title from *AC-dipole is in PTC but not in convertor* to *AC-dipole in PTC but not in convertor*
___
#### 2012-12-07 13:02:38: @ldeniau changed owner from ** to *piotr.skowronski*
___
#### 2012-12-07 13:02:38: @ldeniau changed status from *new* to *assigned*
|
1.0
|
AC-dipole in PTC but not in convertor - *Issue migrated from trac ticket # 108*
**component:** ptc_proper | **priority:** major
___
#### 2012-01-30 13:50:56: frank.schmidt commented
date: 2011.11.19
The AC-dipole is in PTC and lacking in the convertor.
___
#### 2012-01-30 13:51:16: @ldeniau changed title from *AC-dipole is in PTC but not in convertor* to *AC-dipole in PTC but not in convertor*
___
#### 2012-12-07 13:02:38: @ldeniau changed owner from ** to *piotr.skowronski*
___
#### 2012-12-07 13:02:38: @ldeniau changed status from *new* to *assigned*
|
defect
|
ac dipole in ptc but not in convertor issue migrated from trac ticket component ptc proper priority major frank schmidt commented date the ac dipole is in ptc and lacking in the convertor ldeniau changed title from ac dipole is in ptc but not in convertor to ac dipole in ptc but not in convertor ldeniau changed owner from to piotr skowronski ldeniau changed status from new to assigned
| 1
|
347,388
| 10,429,048,757
|
IssuesEvent
|
2019-09-17 01:07:28
|
eveporcello/graphql-fun
|
https://api.github.com/repos/eveporcello/graphql-fun
|
opened
|
Review Notes target Game Prototypes
|
high-priority planning
|
Together we need to review the notes and see what game features we should prototype.
- [ ] Identify Games (easiest - hardest)
- [ ] Create a new task for each game
- [ ] Draft Schemas for each game
- [ ] Plan Steps/Issues for Each Game
- [ ] Each prototype gets a milestone
|
1.0
|
Review Notes target Game Prototypes - Together we need to review the notes and see what game features we should prototype.
- [ ] Identify Games (easiest - hardest)
- [ ] Create a new task for each game
- [ ] Draft Schemas for each game
- [ ] Plan Steps/Issues for Each Game
- [ ] Each prototype gets a milestone
|
non_defect
|
review notes target game prototypes together we need to review the notes and see what game features we should prototype identify games easiest hardest create a new task for each game draft schemas for each game plan steps issues for each game each prototype gets a milestone
| 0
|
379,308
| 11,219,741,738
|
IssuesEvent
|
2020-01-07 14:30:28
|
aiidateam/plumpy
|
https://api.github.com/repos/aiidateam/plumpy
|
closed
|
Make `breadcrumbs` integrated attribute of `PortNamespace`
|
priority/nice to have topic/ports type/requested feature
|
The original design is that a `PortNamespace` should have no knowledge whether it is part of another namespace. This led to problems when trying to raise exceptions with useful messages during validation. If a nested port did not know exactly where it is nested, it could only provide its own name in the error message. For heavily nested namespaces, however, there can be multiple ports with the same name, leaving the user guessing which one threw the error. This was solved ad-hoc by passing an `breadcrumbs` argument along during recursive port validation. However, this concept is starting to become more important in other use cases as well. The original design of not having a bidirectional dependency/knowledge within the nesting should be kept, but we can make the breadcrumbs an attribute of a port when it is created. If it just has the namespace path but no actual object references this should be fine.
|
1.0
|
Make `breadcrumbs` integrated attribute of `PortNamespace` - The original design is that a `PortNamespace` should have no knowledge whether it is part of another namespace. This led to problems when trying to raise exceptions with useful messages during validation. If a nested port did not know exactly where it is nested, it could only provide its own name in the error message. For heavily nested namespaces, however, there can be multiple ports with the same name, leaving the user guessing which one threw the error. This was solved ad-hoc by passing an `breadcrumbs` argument along during recursive port validation. However, this concept is starting to become more important in other use cases as well. The original design of not having a bidirectional dependency/knowledge within the nesting should be kept, but we can make the breadcrumbs an attribute of a port when it is created. If it just has the namespace path but no actual object references this should be fine.
|
non_defect
|
make breadcrumbs integrated attribute of portnamespace the original design is that a portnamespace should have no knowledge whether it is part of another namespace this led to problems when trying to raise exceptions with useful messages during validation if a nested port did not know exactly where it is nested it could only provide its own name in the error message for heavily nested namespaces however there can be multiple ports with the same name leaving the user guessing which one threw the error this was solved ad hoc by passing an breadcrumbs argument along during recursive port validation however this concept is starting to become more important in other use cases as well the original design of not having a bidirectional dependency knowledge within the nesting should be kept but we can make the breadcrumbs an attribute of a port when it is created if it just has the namespace path but no actual object references this should be fine
| 0
|
67,591
| 21,009,060,049
|
IssuesEvent
|
2022-03-30 03:26:29
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: scipy.stats.multiscale_graphcorr p-values are computed differently from literature and other packages
|
defect scipy.stats
|
### Describe your issue.
This bug is as described as here: https://github.com/neurodata/hyppo/issues/124. The issue is reproduced here:
p-values do not appear to be computed correctly. Literature review/review of similar independence testing code seems to suggest most people use the approach suggested in Phipson et al., 2011 described in 6.2 to always include the given ordering as a permutation. Should be updated for all permutation-based approaches to my knowledge.
### Reproducing Code Example
```python
from scipy.stats import multiscale_graphcorr
import numpy as np
X = np.arange(0, 25)
Y = np.arange(0, 25)
stat, pval, _ = multiscale_graphcorr(X, Y, reps=100)
print(pval)
print(stat)
>> .01
>> 1
```
p-value obtained is 1/100.
R example using energy package (contains other nonparametric multivariate independence tests using permutation tests), with energy version 1.7-7 and R version 4.0.2:
```R
require(energy)
X = 0:25; Y = 0:25; result = dcor.test(X, Y, R=100)
print(result$p.value)
>> .0099...
print(result$stat)
>> 1
```
p-value obtained is 1/(100 + 1). We should always use the result from Phipson et. al 2011, rather than only in the case where the p-value would otherwise be 0.
### Error message
```shell
There is no error message for this issue. The proposed change is very small. It would require changing this line (https://github.com/neurodata/hyppo/pull/223):
# calculate p-value and significant permutation map through list
pvalue = (null_dist >= stat).sum() / reps
# correct for a p-value of 0. This is because, with bootstrapping
# permutations, a p-value of 0 is incorrect
if pvalue == 0:
pvalue = 1 / reps
```
to:
```
pvalue = ((null_dist >= stat).sum() + 1) / (1 + reps)
```
### SciPy/NumPy/Python version information
```
1.7.0 1.21.1 sys.version_info(major=3, minor=8, micro=5, releaselevel='final', serial=0)d
```
|
1.0
|
BUG: scipy.stats.multiscale_graphcorr p-values are computed differently from literature and other packages - ### Describe your issue.
This bug is as described as here: https://github.com/neurodata/hyppo/issues/124. The issue is reproduced here:
p-values do not appear to be computed correctly. Literature review/review of similar independence testing code seems to suggest most people use the approach suggested in Phipson et al., 2011 described in 6.2 to always include the given ordering as a permutation. Should be updated for all permutation-based approaches to my knowledge.
### Reproducing Code Example
```python
from scipy.stats import multiscale_graphcorr
import numpy as np
X = np.arange(0, 25)
Y = np.arange(0, 25)
stat, pval, _ = multiscale_graphcorr(X, Y, reps=100)
print(pval)
print(stat)
>> .01
>> 1
```
p-value obtained is 1/100.
R example using energy package (contains other nonparametric multivariate independence tests using permutation tests), with energy version 1.7-7 and R version 4.0.2:
```R
require(energy)
X = 0:25; Y = 0:25; result = dcor.test(X, Y, R=100)
print(result$p.value)
>> .0099...
print(result$stat)
>> 1
```
p-value obtained is 1/(100 + 1). We should always use the result from Phipson et. al 2011, rather than only in the case where the p-value would otherwise be 0.
### Error message
```shell
There is no error message for this issue. The proposed change is very small. It would require changing this line (https://github.com/neurodata/hyppo/pull/223):
# calculate p-value and significant permutation map through list
pvalue = (null_dist >= stat).sum() / reps
# correct for a p-value of 0. This is because, with bootstrapping
# permutations, a p-value of 0 is incorrect
if pvalue == 0:
pvalue = 1 / reps
```
to:
```
pvalue = ((null_dist >= stat).sum() + 1) / (1 + reps)
```
### SciPy/NumPy/Python version information
```
1.7.0 1.21.1 sys.version_info(major=3, minor=8, micro=5, releaselevel='final', serial=0)d
```
|
defect
|
bug scipy stats multiscale graphcorr p values are computed differently from literature and other packages describe your issue this bug is as described as here the issue is reproduced here p values do not appear to be computed correctly literature review review of similar independence testing code seems to suggest most people use the approach suggested in phipson et al described in to always include the given ordering as a permutation should be updated for all permutation based approaches to my knowledge reproducing code example python from scipy stats import multiscale graphcorr import numpy as np x np arange y np arange stat pval multiscale graphcorr x y reps print pval print stat p value obtained is r example using energy package contains other nonparametric multivariate independence tests using permutation tests with energy version and r version r require energy x y result dcor test x y r print result p value print result stat p value obtained is we should always use the result from phipson et al rather than only in the case where the p value would otherwise be error message shell there is no error message for this issue the proposed change is very small it would require changing this line calculate p value and significant permutation map through list pvalue null dist stat sum reps correct for a p value of this is because with bootstrapping permutations a p value of is incorrect if pvalue pvalue reps to pvalue null dist stat sum reps scipy numpy python version information sys version info major minor micro releaselevel final serial d
| 1
|
60,023
| 17,023,314,365
|
IssuesEvent
|
2021-07-03 01:23:04
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
inkscape version 0.44.1 can export large png
|
Component: tilesathome Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 11.05am, Thursday, 23rd October 2008]**
The current check in TahConf.pm just checks for versions of 0.46 vs earlier, while there are other earlier versions that can export the needed 8192x8192px png files.
|
1.0
|
inkscape version 0.44.1 can export large png - **[Submitted to the original trac issue database at 11.05am, Thursday, 23rd October 2008]**
The current check in TahConf.pm just checks for versions of 0.46 vs earlier, while there are other earlier versions that can export the needed 8192x8192px png files.
|
defect
|
inkscape version can export large png the current check in tahconf pm just checks for versions of vs earlier while there are other earlier versions that can export the needed png files
| 1
|
45,694
| 13,023,360,559
|
IssuesEvent
|
2020-07-27 09:52:10
|
pymc-devs/pymc3
|
https://api.github.com/repos/pymc-devs/pymc3
|
closed
|
Shape argument passed to an observed variable is ignored
|
defects shape problem
|
If you specify a `shape` argument for an observed variable, this parameter is silently ignored in favor of the shape of the `observed` value. Specifying `shape` should return an error if specified and does not correspond to the passed data.
|
1.0
|
Shape argument passed to an observed variable is ignored - If you specify a `shape` argument for an observed variable, this parameter is silently ignored in favor of the shape of the `observed` value. Specifying `shape` should return an error if specified and does not correspond to the passed data.
|
defect
|
shape argument passed to an observed variable is ignored if you specify a shape argument for an observed variable this parameter is silently ignored in favor of the shape of the observed value specifying shape should return an error if specified and does not correspond to the passed data
| 1
|
627,143
| 19,896,937,380
|
IssuesEvent
|
2022-01-25 00:52:52
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.att.com - Unable to access the site
|
priority-important priority-normal severity-critical action-needssitepatch browser-fenix engine-gecko QA_triaged
|
<!-- @browser: Firefox Mobile 98.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/98232 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.att.com/
**Browser / Version**: Firefox Mobile 98.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The page says access denied. This may require users to be in the US. Tried reference browser and release Firefox without success.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/09759a9d-26c0-4df0-aaf7-feeab98eade0.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220111093827</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/1/2e1854c7-7ff0-4ffc-b447-a2860c2fbb0c)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
2.0
|
www.att.com - Unable to access the site - <!-- @browser: Firefox Mobile 98.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/98232 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.att.com/
**Browser / Version**: Firefox Mobile 98.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The page says access denied. This may require users to be in the US. Tried reference browser and release Firefox without success.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/09759a9d-26c0-4df0-aaf7-feeab98eade0.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220111093827</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/1/2e1854c7-7ff0-4ffc-b447-a2860c2fbb0c)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
unable to access the site url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce the page says access denied this may require users to be in the us tried reference browser and release firefox without success view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
241,892
| 7,835,397,009
|
IssuesEvent
|
2018-06-17 05:00:29
|
rathena/rathena
|
https://api.github.com/repos/rathena/rathena
|
closed
|
Rand attack in nonfunctional refinement
|
component:core mode:renewal priority:low type:bug
|
Currently I checked the equipment refining system with random bonus ATK is not working.
A weapons LV4 for example, after the fifth refining should add in a weapons attack rand (0 ~ 14), which now appears to be constant. In refine_db.txt have the correct information, but does not appear to work.
In the right situation an armament for example, +12 refining could have difference in the damage other equal also +12.
Link: http://irowiki.org/wiki/Refinement_System **(Effects Gained)**
NOTE: The link follows a model similar to the refine_db the rAthena, but the end result is strange. There is no difference between weapons refinements.
|
1.0
|
Rand attack in nonfunctional refinement - Currently I checked the equipment refining system with random bonus ATK is not working.
A weapons LV4 for example, after the fifth refining should add in a weapons attack rand (0 ~ 14), which now appears to be constant. In refine_db.txt have the correct information, but does not appear to work.
In the right situation an armament for example, +12 refining could have difference in the damage other equal also +12.
Link: http://irowiki.org/wiki/Refinement_System **(Effects Gained)**
NOTE: The link follows a model similar to the refine_db the rAthena, but the end result is strange. There is no difference between weapons refinements.
|
non_defect
|
rand attack in nonfunctional refinement currently i checked the equipment refining system with random bonus atk is not working a weapons for example after the fifth refining should add in a weapons attack rand which now appears to be constant in refine db txt have the correct information but does not appear to work in the right situation an armament for example refining could have difference in the damage other equal also link effects gained note the link follows a model similar to the refine db the rathena but the end result is strange there is no difference between weapons refinements
| 0
|
284,112
| 8,735,807,612
|
IssuesEvent
|
2018-12-11 17:43:59
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
Cracks Clipper is broken in 2.x
|
bug crash likelihood medium priority reviewed severity high wrong results
|
The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 402
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Cracks Clipper is broken in 2.x
Assigned to: Kathleen Biagas
Category:
Target version: 2.1.1
Author: Kathleen Biagas
Start: 09/22/2010
Due date:
% Done: 0
Estimated time:
Created: 09/22/2010 12:54 pm
Updated: 09/29/2010 02:53 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap.
Comments:
Restored functionality of CracksClipper operator.SVN revisions 12594 (2.1 RC) 12596 (trunk).
|
1.0
|
Cracks Clipper is broken in 2.x - The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 402
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Cracks Clipper is broken in 2.x
Assigned to: Kathleen Biagas
Category:
Target version: 2.1.1
Author: Kathleen Biagas
Start: 09/22/2010
Due date:
% Done: 0
Estimated time:
Created: 09/22/2010 12:54 pm
Updated: 09/29/2010 02:53 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap.
Comments:
Restored functionality of CracksClipper operator.SVN revisions 12594 (2.1 RC) 12596 (trunk).
|
non_defect
|
cracks clipper is broken in x the cracksclipper operator no longer works as of greg burton has need of this functionality and would like it fixed asap redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject cracks clipper is broken in x assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description the cracksclipper operator no longer works as of greg burton has need of this functionality and would like it fixed asap comments restored functionality of cracksclipper operator svn revisions rc trunk
| 0
|
17,577
| 3,012,747,511
|
IssuesEvent
|
2015-07-29 02:09:30
|
yawlfoundation/yawl
|
https://api.github.com/repos/yawlfoundation/yawl
|
closed
|
[CLOSED] Unchain/unpile should ge greyed out in certain circumstances
|
auto-migrated Priority-Low Type-Defect
|
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:20 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/25_
----
```
In the Edit Profile screen the unchain/unpile should be greyed out if the
user does not have those priviliges and/or no tasks are selected.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 21 Jul 2008 at 8:05
|
1.0
|
[CLOSED] Unchain/unpile should ge greyed out in certain circumstances - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:20 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/25_
----
```
In the Edit Profile screen the unchain/unpile should be greyed out if the
user does not have those priviliges and/or no tasks are selected.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 21 Jul 2008 at 8:05
|
defect
|
unchain unpile should ge greyed out in certain circumstances issue by monday jul at gmt originally opened as in the edit profile screen the unchain unpile should be greyed out if the user does not have those priviliges and or no tasks are selected original issue reported on code google com by arthurte gmail com on jul at
| 1
|
58,448
| 14,398,713,499
|
IssuesEvent
|
2020-12-03 09:57:29
|
joncampbell123/dosbox-x
|
https://api.github.com/repos/joncampbell123/dosbox-x
|
closed
|
flatpak build problem
|
build issues platform: Linux
|
**Describe the bug**
Ok, that only took one release :-)
For some reason, that I'm not sure about the flatpak build process no longer works. The build seems to get stuck in a continuous loop. It seems it successfully does the autogen.sh and configure steps, then starts the make step and shows the following:
```
...
config.status: executing depfiles commands
Running: make
make: Warning: File 'Makefile.am' has modification time 1275 s in the future
CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing aclocal-1.16
cd . && /bin/sh /run/build/dosbox-x/missing automake-1.16 --foreign
CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing autoconf
/bin/sh ./config.status --recheck
running CONFIG_SHELL=/bin/sh /bin/sh ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 CFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection LDFLAGS=-L/app/lib -Wl,-z,relro,-z,now -Wl,--as-needed CXXFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection --no-create --no-recursion
checking build system type... x86_64-pc-linux-gnu
...
```
And the configure and make just keeps looping.
**To Reproduce**
Steps to reproduce the behavior:
```
git clone --recursive https://github.com/flathub/com.dosbox_x.DOSBox-X.git
cd com.dosbox_x.DOSBox-X
flatpak install flathub org.freedesktop.Sdk//20.08 -y
flatpak-builder --force-clean --install --user -y builddir com.dosbox_x.DOSBox-X.yaml
```
**Additional context**
By default the yaml file will use buildsystem=autotool and that worked with the last version. I tried switching to buildsystem=simple and specifying the build-commands, but that also gets into the same circular configure issue.
e.g.
```
- name: dosbox-x
buildsystem: simple
build-commands:
- ./autogen.sh
- ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2
- make
- make install
sources:
- type: archive
url: https://github.com/joncampbell123/dosbox-x/archive/dosbox-x-v0.83.7.tar.gz
sha256: 9cdfa3267c340a869255d8eb1c4ebf4adde47c22854e1d013da22190350bfbb3
post-install:
- install -Dm644 /app/share/icons/hicolor/scalable/apps/dosbox-x.svg /app/share/icons/hicolor/scalable/apps/${FLATPAK_ID}.svg
- desktop-file-edit --set-key=Icon --set-value=${FLATPAK_ID} /app/share/applications/${FLATPAK_ID}.desktop
```
I also tried replacing the build-commands with just a simple ``./build-debug-sdl2`` and then it crashes out with a:
```
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... configure: error: newly created file is older than distributed files!
Check your system clock
Error: module dosbox-x: Child process exited with code 1
```
To clarify the flatpak build runs in an environment with EPOCH set to zero. This is intentional, and I don't think you can disable it.
Loop during regular build:
```
```
|
1.0
|
flatpak build problem - **Describe the bug**
Ok, that only took one release :-)
For some reason, that I'm not sure about the flatpak build process no longer works. The build seems to get stuck in a continuous loop. It seems it successfully does the autogen.sh and configure steps, then starts the make step and shows the following:
```
...
config.status: executing depfiles commands
Running: make
make: Warning: File 'Makefile.am' has modification time 1275 s in the future
CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing aclocal-1.16
cd . && /bin/sh /run/build/dosbox-x/missing automake-1.16 --foreign
CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing autoconf
/bin/sh ./config.status --recheck
running CONFIG_SHELL=/bin/sh /bin/sh ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 CFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection LDFLAGS=-L/app/lib -Wl,-z,relro,-z,now -Wl,--as-needed CXXFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection --no-create --no-recursion
checking build system type... x86_64-pc-linux-gnu
...
```
And the configure and make just keeps looping.
**To Reproduce**
Steps to reproduce the behavior:
```
git clone --recursive https://github.com/flathub/com.dosbox_x.DOSBox-X.git
cd com.dosbox_x.DOSBox-X
flatpak install flathub org.freedesktop.Sdk//20.08 -y
flatpak-builder --force-clean --install --user -y builddir com.dosbox_x.DOSBox-X.yaml
```
**Additional context**
By default the yaml file will use buildsystem=autotool and that worked with the last version. I tried switching to buildsystem=simple and specifying the build-commands, but that also gets into the same circular configure issue.
e.g.
```
- name: dosbox-x
buildsystem: simple
build-commands:
- ./autogen.sh
- ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2
- make
- make install
sources:
- type: archive
url: https://github.com/joncampbell123/dosbox-x/archive/dosbox-x-v0.83.7.tar.gz
sha256: 9cdfa3267c340a869255d8eb1c4ebf4adde47c22854e1d013da22190350bfbb3
post-install:
- install -Dm644 /app/share/icons/hicolor/scalable/apps/dosbox-x.svg /app/share/icons/hicolor/scalable/apps/${FLATPAK_ID}.svg
- desktop-file-edit --set-key=Icon --set-value=${FLATPAK_ID} /app/share/applications/${FLATPAK_ID}.desktop
```
I also tried replacing the build-commands with just a simple ``./build-debug-sdl2`` and then it crashes out with a:
```
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... configure: error: newly created file is older than distributed files!
Check your system clock
Error: module dosbox-x: Child process exited with code 1
```
To clarify the flatpak build runs in an environment with EPOCH set to zero. This is intentional, and I don't think you can disable it.
Loop during regular build:
```
```
|
non_defect
|
flatpak build problem describe the bug ok that only took one release for some reason that i m not sure about the flatpak build process no longer works the build seems to get stuck in a continuous loop it seems it successfully does the autogen sh and configure steps then starts the make step and shows the following config status executing depfiles commands running make make warning file makefile am has modification time s in the future cdpath zsh version cd bin sh run build dosbox x missing aclocal cd bin sh run build dosbox x missing automake foreign cdpath zsh version cd bin sh run build dosbox x missing autoconf bin sh config status recheck running config shell bin sh bin sh configure enable core inline enable debug heavy enable cflags g pipe wp d fortify source wp d glibcxx assertions fexceptions fstack protector strong grecord gcc switches fasynchronous unwind tables fstack clash protection fcf protection ldflags l app lib wl z relro z now wl as needed cxxflags g pipe wp d fortify source wp d glibcxx assertions fexceptions fstack protector strong grecord gcc switches fasynchronous unwind tables fstack clash protection fcf protection no create no recursion checking build system type pc linux gnu and the configure and make just keeps looping to reproduce steps to reproduce the behavior git clone recursive cd com dosbox x dosbox x flatpak install flathub org freedesktop sdk y flatpak builder force clean install user y builddir com dosbox x dosbox x yaml additional context by default the yaml file will use buildsystem autotool and that worked with the last version i tried switching to buildsystem simple and specifying the build commands but that also gets into the same circular configure issue e g name dosbox x buildsystem simple build commands autogen sh configure enable core inline enable debug heavy enable make make install sources type archive url post install install app share icons hicolor scalable apps dosbox x svg app share icons hicolor scalable apps flatpak id svg desktop file edit set key icon set value flatpak id app share applications flatpak id desktop i also tried replacing the build commands with just a simple build debug and then it crashes out with a checking for a bsd compatible install usr bin install c checking whether build environment is sane configure error newly created file is older than distributed files check your system clock error module dosbox x child process exited with code to clarify the flatpak build runs in an environment with epoch set to zero this is intentional and i don t think you can disable it loop during regular build
| 0
|
54,826
| 23,341,619,081
|
IssuesEvent
|
2022-08-09 14:27:04
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Ability to create and use alias for index pattern
|
enhancement loe:hours Feature:Data Views Team:AppServicesSv impact:low
|
**Describe the feature:**
Would like to have the ability to add a meaningful displayname for an indexpattern. This name should be shown in the locations where a user can choose an indexpattern for his/her search. Not all users are tech savvy and indexnames often are quite technical. For those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use.
**Describe a specific use case for the feature:**
On the discover tab there is this dropdown to choose an index pattern, right below the 'add filter' link. I would like to see different names here, so instead of lnd-clust:idx-ms3-sw-ds,lp-clust:idx-ms3-sw-ds and lnd-clust::idx-ms3-sw-ds , Users could see the names London-and-liverpool-subway-stats and London-subway-stats
Above is just an example, but it would be helpful, especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences.
I can image this is just an alias one could set when creating the index pattern and if it is not set, the indexpattern itselves is shown like it is now.
|
1.0
|
Ability to create and use alias for index pattern - **Describe the feature:**
Would like to have the ability to add a meaningful displayname for an indexpattern. This name should be shown in the locations where a user can choose an indexpattern for his/her search. Not all users are tech savvy and indexnames often are quite technical. For those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use.
**Describe a specific use case for the feature:**
On the discover tab there is this dropdown to choose an index pattern, right below the 'add filter' link. I would like to see different names here, so instead of lnd-clust:idx-ms3-sw-ds,lp-clust:idx-ms3-sw-ds and lnd-clust::idx-ms3-sw-ds , Users could see the names London-and-liverpool-subway-stats and London-subway-stats
Above is just an example, but it would be helpful, especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences.
I can image this is just an alias one could set when creating the index pattern and if it is not set, the indexpattern itselves is shown like it is now.
|
non_defect
|
ability to create and use alias for index pattern describe the feature would like to have the ability to add a meaningful displayname for an indexpattern this name should be shown in the locations where a user can choose an indexpattern for his her search not all users are tech savvy and indexnames often are quite technical for those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use describe a specific use case for the feature on the discover tab there is this dropdown to choose an index pattern right below the add filter link i would like to see different names here so instead of lnd clust idx sw ds lp clust idx sw ds and lnd clust idx sw ds users could see the names london and liverpool subway stats and london subway stats above is just an example but it would be helpful especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences i can image this is just an alias one could set when creating the index pattern and if it is not set the indexpattern itselves is shown like it is now
| 0
|
71,708
| 18,848,499,757
|
IssuesEvent
|
2021-11-11 17:36:07
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
closed
|
Fenics fails to run after installation with spack
|
build-error
|
Fenics fails to run after installation with spack
### Expected Result
To be able to execute simple fenics scripts using the spack-built fenics.
### Actual Result
Simply importing the fenics module into the spack-provided python fails with:
```console
ImportError: No module named _common
```
### Steps to reproduce the issue
```
bridges@wheeler-sn[500]> module load fenics-2017.1.0-intel-17.0.4-impi-mkl-python2-zre3f5b
Autoloading eigen-3.3.3-intel-17.0.4-impi-mkl-lnbosvu
Autoloading metis-5.1.0-intel-17.0.4-xyonw4o
Autoloading fftw-3.3.6-pl2-intel-17.0.4-impi-bwlmvhs
Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp
Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc
Autoloading intel-mkl-17.0.4-intel-17.0.4-z2nms2r
Autoloading gmp-6.1.2-intel-17.0.4-asgihpm
Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk
Autoloading mpfr-3.1.5-intel-17.0.4-e2wn26r
Autoloading py-mpmath-0.19-intel-17.0.4-python2-bub4nhr
Autoloading python-2.7.13-intel-17.0.4-ufhcwdr
Autoloading bzip2-1.0.6-intel-17.0.4-zjtu25w
Autoloading ncurses-6.0-intel-17.0.4-j3ftwhg
Autoloading openssl-1.0.2k-intel-17.0.4-6n2jfwi
Autoloading sqlite-3.18.0-intel-17.0.4-cyhl22m
Autoloading readline-7.0-intel-17.0.4-psasazv
Autoloading readline-7.0-intel-17.0.4-psasazv
Autoloading py-ply-3.8-intel-17.0.4-python2-ntmnh7h
Autoloading py-instant-2017.1.0-intel-17.0.4-mkl-python2-ebd5pyo
Autoloading swig-3.0.12-intel-17.0.4-bbmqam2
Autoloading pcre-8.40-intel-17.0.4-72ed5lu
Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx
Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx
Autoloading py-appdirs-1.4.3-intel-17.0.4-python2-mupwe3g
Autoloading py-packaging-16.8-intel-17.0.4-python2-js2s5wi
Autoloading py-pyparsing-2.2.0-intel-17.0.4-python2-vh7hq5e
Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc
Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl
Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl
Autoloading py-slepc4py-3.7.0-intel-17.0.4-impi-mkl-python2-iahwthv
Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6
Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-3x6zx5l
Autoloading hypre-2.11.2-intel-17.0.4-impi-mkl-g5srp2y
Autoloading superlu-dist-5.1.3-intel-17.0.4-impi-mkl-ymsttjs
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading py-mpi4py-2.0.0-intel-17.0.4-impi-python2-ozw6g4g
Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-fy5wf4h
Autoloading arpack-ng-3.5.0-intel-17.0.4-impi-mkl-l7ewls2
Autoloading py-ufl-2017.1.0-intel-17.0.4-mkl-python2-2xalq65
Autoloading cmake-3.8.1-intel-17.0.4-m63t5kd
Autoloading py-sympy-1.0-intel-17.0.4-python2-tap7mf2
Autoloading py-dijitso-2017.1.0-intel-17.0.4-mkl-python2-ljegoot
Autoloading boost-1.63.0-intel-17.0.4-bjbzbtc
Autoloading trilinos-12.10.1-intel-17.0.4-impi-mkl-l67gizf
Autoloading glm-0.9.7.1-intel-17.0.4-wmre6yo
Autoloading netcdf-4.4.1.1-intel-17.0.4-impi-bdktit6
Autoloading mumps-5.1.1-intel-17.0.4-impi-mkl-6enszkq
Autoloading matio-1.5.9-intel-17.0.4-impi-sjmlxnf
Autoloading python-2.7.13-intel-17.0.4-ufhcwdr
Autoloading swig-3.0.12-intel-17.0.4-bbmqam2
Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx
Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc
Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2
Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-python2-fy5wf4h
Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc
Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-python2-3x6zx5l
Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6
Autoloading py-fiat-2017.1.0-intel-17.0.4-mkl-python2-6krzj5q
Autoloading py-ffc-2017.1.0-intel-17.0.4-mkl-python2-mleunix
Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk
Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp
bridges@wheeler-sn[501]> which python
python is /opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/bin/python
bridges@wheeler-sn[502]> python -c 'import fenics'
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/fenics/__init__.py", line 7, in <module>
from dolfin import *
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/__init__.py", line 17, in <module>
from . import cpp
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/__init__.py", line 43, in <module>
exec("from . import %s" % module_name)
File "<string>", line 1, in <module>
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 21, in <module>
_common = swig_import_helper()
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 20, in swig_import_helper
return importlib.import_module('_common')
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/lib/python2.7/importlib/__init__.py", line 37, in import_module
__import__(name)
ImportError: No module named _common
bridges@wheeler-sn[503]>
```
### Information on your system
This includes:
1. OpenHPC Running on CentOS 7.3
|
1.0
|
Fenics fails to run after installation with spack - Fenics fails to run after installation with spack
### Expected Result
To be able to execute simple fenics scripts using the spack-built fenics.
### Actual Result
Simply importing the fenics module into the spack-provided python fails with:
```console
ImportError: No module named _common
```
### Steps to reproduce the issue
```
bridges@wheeler-sn[500]> module load fenics-2017.1.0-intel-17.0.4-impi-mkl-python2-zre3f5b
Autoloading eigen-3.3.3-intel-17.0.4-impi-mkl-lnbosvu
Autoloading metis-5.1.0-intel-17.0.4-xyonw4o
Autoloading fftw-3.3.6-pl2-intel-17.0.4-impi-bwlmvhs
Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp
Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc
Autoloading intel-mkl-17.0.4-intel-17.0.4-z2nms2r
Autoloading gmp-6.1.2-intel-17.0.4-asgihpm
Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk
Autoloading mpfr-3.1.5-intel-17.0.4-e2wn26r
Autoloading py-mpmath-0.19-intel-17.0.4-python2-bub4nhr
Autoloading python-2.7.13-intel-17.0.4-ufhcwdr
Autoloading bzip2-1.0.6-intel-17.0.4-zjtu25w
Autoloading ncurses-6.0-intel-17.0.4-j3ftwhg
Autoloading openssl-1.0.2k-intel-17.0.4-6n2jfwi
Autoloading sqlite-3.18.0-intel-17.0.4-cyhl22m
Autoloading readline-7.0-intel-17.0.4-psasazv
Autoloading readline-7.0-intel-17.0.4-psasazv
Autoloading py-ply-3.8-intel-17.0.4-python2-ntmnh7h
Autoloading py-instant-2017.1.0-intel-17.0.4-mkl-python2-ebd5pyo
Autoloading swig-3.0.12-intel-17.0.4-bbmqam2
Autoloading pcre-8.40-intel-17.0.4-72ed5lu
Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx
Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx
Autoloading py-appdirs-1.4.3-intel-17.0.4-python2-mupwe3g
Autoloading py-packaging-16.8-intel-17.0.4-python2-js2s5wi
Autoloading py-pyparsing-2.2.0-intel-17.0.4-python2-vh7hq5e
Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc
Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl
Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl
Autoloading py-slepc4py-3.7.0-intel-17.0.4-impi-mkl-python2-iahwthv
Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6
Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-3x6zx5l
Autoloading hypre-2.11.2-intel-17.0.4-impi-mkl-g5srp2y
Autoloading superlu-dist-5.1.3-intel-17.0.4-impi-mkl-ymsttjs
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading py-mpi4py-2.0.0-intel-17.0.4-impi-python2-ozw6g4g
Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-fy5wf4h
Autoloading arpack-ng-3.5.0-intel-17.0.4-impi-mkl-l7ewls2
Autoloading py-ufl-2017.1.0-intel-17.0.4-mkl-python2-2xalq65
Autoloading cmake-3.8.1-intel-17.0.4-m63t5kd
Autoloading py-sympy-1.0-intel-17.0.4-python2-tap7mf2
Autoloading py-dijitso-2017.1.0-intel-17.0.4-mkl-python2-ljegoot
Autoloading boost-1.63.0-intel-17.0.4-bjbzbtc
Autoloading trilinos-12.10.1-intel-17.0.4-impi-mkl-l67gizf
Autoloading glm-0.9.7.1-intel-17.0.4-wmre6yo
Autoloading netcdf-4.4.1.1-intel-17.0.4-impi-bdktit6
Autoloading mumps-5.1.1-intel-17.0.4-impi-mkl-6enszkq
Autoloading matio-1.5.9-intel-17.0.4-impi-sjmlxnf
Autoloading python-2.7.13-intel-17.0.4-ufhcwdr
Autoloading swig-3.0.12-intel-17.0.4-bbmqam2
Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx
Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc
Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx
Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v
Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2
Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-python2-fy5wf4h
Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc
Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-python2-3x6zx5l
Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6
Autoloading py-fiat-2017.1.0-intel-17.0.4-mkl-python2-6krzj5q
Autoloading py-ffc-2017.1.0-intel-17.0.4-mkl-python2-mleunix
Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk
Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp
bridges@wheeler-sn[501]> which python
python is /opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/bin/python
bridges@wheeler-sn[502]> python -c 'import fenics'
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/fenics/__init__.py", line 7, in <module>
from dolfin import *
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/__init__.py", line 17, in <module>
from . import cpp
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/__init__.py", line 43, in <module>
exec("from . import %s" % module_name)
File "<string>", line 1, in <module>
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 21, in <module>
_common = swig_import_helper()
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 20, in swig_import_helper
return importlib.import_module('_common')
File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/lib/python2.7/importlib/__init__.py", line 37, in import_module
__import__(name)
ImportError: No module named _common
bridges@wheeler-sn[503]>
```
### Information on your system
This includes:
1. OpenHPC Running on CentOS 7.3
|
non_defect
|
fenics fails to run after installation with spack fenics fails to run after installation with spack expected result to be able to execute simple fenics scripts using the spack built fenics actual result simply importing the fenics module into the spack provided python fails with console importerror no module named common steps to reproduce the issue bridges wheeler sn module load fenics intel impi mkl autoloading eigen intel impi mkl lnbosvu autoloading metis intel autoloading fftw intel impi bwlmvhs autoloading intel mpi intel ataurnp autoloading suite sparse intel mkl autoloading intel mkl intel autoloading gmp intel asgihpm autoloading scotch intel impi dxrzrfk autoloading mpfr intel autoloading py mpmath intel autoloading python intel ufhcwdr autoloading intel autoloading ncurses intel autoloading openssl intel autoloading sqlite intel autoloading readline intel psasazv autoloading readline intel psasazv autoloading py ply intel autoloading py instant intel mkl autoloading swig intel autoloading pcre intel autoloading py setuptools intel autoloading py six intel autoloading py appdirs intel autoloading py packaging intel autoloading py pyparsing intel autoloading py numpy intel mkl autoloading py intel autoloading py intel autoloading py intel impi mkl iahwthv autoloading py intel impi mkl autoloading petsc intel impi mkl autoloading hypre intel impi mkl autoloading superlu dist intel impi mkl ymsttjs autoloading parmetis intel impi autoloading intel impi autoloading parmetis intel impi autoloading py intel impi autoloading slepc intel impi mkl autoloading arpack ng intel impi mkl autoloading py ufl intel mkl autoloading cmake intel autoloading py sympy intel autoloading py dijitso intel mkl ljegoot autoloading boost intel bjbzbtc autoloading trilinos intel impi mkl autoloading glm intel autoloading netcdf intel impi autoloading mumps intel impi mkl autoloading matio intel impi sjmlxnf autoloading python intel ufhcwdr autoloading swig intel autoloading py six intel autoloading suite sparse intel mkl autoloading py setuptools intel autoloading parmetis intel impi autoloading intel impi autoloading slepc intel impi mkl autoloading py numpy intel mkl autoloading petsc intel impi mkl autoloading py intel impi mkl autoloading py fiat intel mkl autoloading py ffc intel mkl mleunix autoloading scotch intel impi dxrzrfk autoloading intel mpi intel ataurnp bridges wheeler sn which python python is opt spack opt spack linux intel python bin python bridges wheeler sn python c import fenics traceback most recent call last file line in file opt spack opt spack linux intel fenics lib site packages fenics init py line in from dolfin import file opt spack opt spack linux intel fenics lib site packages dolfin init py line in from import cpp file opt spack opt spack linux intel fenics lib site packages dolfin cpp init py line in exec from import s module name file line in file opt spack opt spack linux intel fenics lib site packages dolfin cpp common py line in common swig import helper file opt spack opt spack linux intel fenics lib site packages dolfin cpp common py line in swig import helper return importlib import module common file opt spack opt spack linux intel python lib importlib init py line in import module import name importerror no module named common bridges wheeler sn information on your system this includes openhpc running on centos
| 0
|
72,769
| 24,280,758,850
|
IssuesEvent
|
2022-09-28 17:10:51
|
jccastillo0007/eFacturaT
|
https://api.github.com/repos/jccastillo0007/eFacturaT
|
closed
|
PARA LO DE PEPSI, NO ESTÁ RESPETANDO EL USO DE CFDI ELEGIDO
|
bug resolved defect
|
Te mandé un correo donde el uso de CFDI no está coincidiendo el uso de CFDI en el PDF con el XML.
Ellos eligen I08 y lo incluye en el PDF, pero en el XML lo pone distinto.
|
1.0
|
PARA LO DE PEPSI, NO ESTÁ RESPETANDO EL USO DE CFDI ELEGIDO - Te mandé un correo donde el uso de CFDI no está coincidiendo el uso de CFDI en el PDF con el XML.
Ellos eligen I08 y lo incluye en el PDF, pero en el XML lo pone distinto.
|
defect
|
para lo de pepsi no está respetando el uso de cfdi elegido te mandé un correo donde el uso de cfdi no está coincidiendo el uso de cfdi en el pdf con el xml ellos eligen y lo incluye en el pdf pero en el xml lo pone distinto
| 1
|
399,064
| 11,742,661,888
|
IssuesEvent
|
2020-03-12 01:37:31
|
thaliawww/concrexit
|
https://api.github.com/repos/thaliawww/concrexit
|
closed
|
Revisit urls schemes
|
priority: low refactor technical change
|
In GitLab by @se-bastiaan on Apr 28, 2019, 21:30
### One-sentence description
Revisit urls schemes
### Why?
It's currently a bit messy and not consistent.
### Current implementation
We have one central urls.py for our application in thaliawebsite which decides the namespacing of all the modules.
### Suggested implementation
The central urls.py only includes the urls.py files of all the packages. These packages are then allowed to fill in their url schemes how they see fit. This allows us to re-use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page.
All urls used in the django admin and implemented via the central urls.py (like push notifications and newsletters) should be implemented using the ModelAdmin's `get_urls` method like it was done in registrations and payments.
|
1.0
|
Revisit urls schemes - In GitLab by @se-bastiaan on Apr 28, 2019, 21:30
### One-sentence description
Revisit urls schemes
### Why?
It's currently a bit messy and not consistent.
### Current implementation
We have one central urls.py for our application in thaliawebsite which decides the namespacing of all the modules.
### Suggested implementation
The central urls.py only includes the urls.py files of all the packages. These packages are then allowed to fill in their url schemes how they see fit. This allows us to re-use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page.
All urls used in the django admin and implemented via the central urls.py (like push notifications and newsletters) should be implemented using the ModelAdmin's `get_urls` method like it was done in registrations and payments.
|
non_defect
|
revisit urls schemes in gitlab by se bastiaan on apr one sentence description revisit urls schemes why it s currently a bit messy and not consistent current implementation we have one central urls py for our application in thaliawebsite which decides the namespacing of all the modules suggested implementation the central urls py only includes the urls py files of all the packages these packages are then allowed to fill in their url schemes how they see fit this allows us to re use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page all urls used in the django admin and implemented via the central urls py like push notifications and newsletters should be implemented using the modeladmin s get urls method like it was done in registrations and payments
| 0
|
61,952
| 17,023,817,712
|
IssuesEvent
|
2021-07-03 04:00:52
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Query like "75 Gegužių g., Šiauliai" does not find address
|
Component: nominatim Priority: minor Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 8.55am, Thursday, 23rd August 2012]**
This building can be found at http://www.openstreetmap.org/edit?lat=55.91334&lon=23.25451&zoom=16 , but nominatim query results show only streets, most of which are located outside the city.
|
1.0
|
Query like "75 Gegužių g., Šiauliai" does not find address - **[Submitted to the original trac issue database at 8.55am, Thursday, 23rd August 2012]**
This building can be found at http://www.openstreetmap.org/edit?lat=55.91334&lon=23.25451&zoom=16 , but nominatim query results show only streets, most of which are located outside the city.
|
defect
|
query like geguå¾iå³ g å iauliai does not find address this building can be found at but nominatim query results show only streets most of which are located outside the city
| 1
|
333,089
| 29,508,088,184
|
IssuesEvent
|
2023-06-03 15:13:09
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix decompositions.test_numpy_qr
|
NumPy Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix decompositions.test_numpy_qr - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
non_defect
|
fix decompositions test numpy qr tensorflow img src torch img src numpy img src jax img src paddle img src
| 0
|
24,270
| 3,947,090,343
|
IssuesEvent
|
2016-04-28 08:28:03
|
buildo/github-workflow-pal
|
https://api.github.com/repos/buildo/github-workflow-pal
|
opened
|
[subissue] should filter out "customers" label
|
defect
|
## description
"customers" label should be filtered out like "macro" to avoid adding cards to prisma by mistake

## how to reproduce
- add subissue to customer issue
## specs
filter out "customers" too here: https://github.com/buildo/github-workflow-pal/blob/master/src/plugin/addSubIssueButton.js#L32
|
1.0
|
[subissue] should filter out "customers" label - ## description
"customers" label should be filtered out like "macro" to avoid adding cards to prisma by mistake

## how to reproduce
- add subissue to customer issue
## specs
filter out "customers" too here: https://github.com/buildo/github-workflow-pal/blob/master/src/plugin/addSubIssueButton.js#L32
|
defect
|
should filter out customers label description customers label should be filtered out like macro to avoid adding cards to prisma by mistake how to reproduce add subissue to customer issue specs filter out customers too here
| 1
|
140,173
| 11,303,578,150
|
IssuesEvent
|
2020-01-17 20:28:50
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
opened
|
CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898
|
CWE-134 bug flawfinder flawfinder_level_4 static analysis testing triage
|
```yaml
{
"severity": {
"severity-description": "",
"cvss-score": 0,
"rvss-vector": "",
"cvss-vector": "",
"rvss-score": 0
},
"id": 1,
"mitigation": {
"description": "Use a constant for the format specification",
"date-mitigation": "",
"pull-request": ""
},
"flaw": {
"date-detected": "2020-01-17 (20:28)",
"package": "N/A",
"issue": "",
"detected-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"detected-by-method": "testing static",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/407292189/artifacts/download",
"date-reported": "2020-01-17 (20:28)",
"reproduction": "See artifacts below (if available)",
"specificity": "subject-specific",
"subsystem": "N/A",
"languages": "None",
"trace": "(context) int vsnprintf(char *s, size_t sz, const char *fmt, va_list args)",
"reproducibility": "always",
"application": "N/A",
"reported-by": "Alias Robotics",
"phase": "testing",
"architectural-location": "application-specific"
},
"vendor": null,
"cve": "None",
"system": "./install/motoman_driver/share/motoman_driver/MotoPlus/Controller.c:898:5",
"links": "",
"keywords": [
"flawfinder",
"flawfinder_level_4",
"static analysis",
"testing",
"triage",
"CWE-134",
"bug"
],
"description": "If format strings can be influenced by an attacker, they can be exploited, and note that sprintf variations do not always \\\\0-terminate (CWE-134). Use a constant for the format specification. . Happening @ ...ver/MotoPlus/Controller.c:898",
"title": "CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898",
"exploitation": {
"description": "",
"exploitation-vector": "",
"exploitation-image": ""
},
"cwe": [
"CWE-134"
],
"type": "bug"
}
```
|
1.0
|
CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898 - ```yaml
{
"severity": {
"severity-description": "",
"cvss-score": 0,
"rvss-vector": "",
"cvss-vector": "",
"rvss-score": 0
},
"id": 1,
"mitigation": {
"description": "Use a constant for the format specification",
"date-mitigation": "",
"pull-request": ""
},
"flaw": {
"date-detected": "2020-01-17 (20:28)",
"package": "N/A",
"issue": "",
"detected-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"detected-by-method": "testing static",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/407292189/artifacts/download",
"date-reported": "2020-01-17 (20:28)",
"reproduction": "See artifacts below (if available)",
"specificity": "subject-specific",
"subsystem": "N/A",
"languages": "None",
"trace": "(context) int vsnprintf(char *s, size_t sz, const char *fmt, va_list args)",
"reproducibility": "always",
"application": "N/A",
"reported-by": "Alias Robotics",
"phase": "testing",
"architectural-location": "application-specific"
},
"vendor": null,
"cve": "None",
"system": "./install/motoman_driver/share/motoman_driver/MotoPlus/Controller.c:898:5",
"links": "",
"keywords": [
"flawfinder",
"flawfinder_level_4",
"static analysis",
"testing",
"triage",
"CWE-134",
"bug"
],
"description": "If format strings can be influenced by an attacker, they can be exploited, and note that sprintf variations do not always \\\\0-terminate (CWE-134). Use a constant for the format specification. . Happening @ ...ver/MotoPlus/Controller.c:898",
"title": "CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898",
"exploitation": {
"description": "",
"exploitation-vector": "",
"exploitation-image": ""
},
"cwe": [
"CWE-134"
],
"type": "bug"
}
```
|
non_defect
|
cwe format if format strings can be influenced by an attacker they can be exploi ver motoplus controller c yaml severity severity description cvss score rvss vector cvss vector rvss score id mitigation description use a constant for the format specification date mitigation pull request flaw date detected package n a issue detected by alias robotics reported by relationship automatic detected by method testing static reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros industrial jobs artifacts download date reported reproduction see artifacts below if available specificity subject specific subsystem n a languages none trace context int vsnprintf char s size t sz const char fmt va list args reproducibility always application n a reported by alias robotics phase testing architectural location application specific vendor null cve none system install motoman driver share motoman driver motoplus controller c links keywords flawfinder flawfinder level static analysis testing triage cwe bug description if format strings can be influenced by an attacker they can be exploited and note that sprintf variations do not always terminate cwe use a constant for the format specification happening ver motoplus controller c title cwe format if format strings can be influenced by an attacker they can be exploi ver motoplus controller c exploitation description exploitation vector exploitation image cwe cwe type bug
| 0
|
52,696
| 13,224,940,356
|
IssuesEvent
|
2020-08-17 20:09:41
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
ACER Failed To Converge for all Events (Trac #170)
|
Migrated from Trac combo reconstruction defect
|
AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/170">https://code.icecube.wisc.edu/projects/icecube/ticket/170</a>, reported by pankninand owned by panknin</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:08:15",
"_ts": "1547240895681654",
"description": "AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.",
"reporter": "panknin",
"cc": "",
"resolution": "fixed",
"time": "2009-10-27T10:10:23",
"component": "combo reconstruction",
"summary": "ACER Failed To Converge for all Events",
"priority": "critical",
"keywords": "",
"milestone": "",
"owner": "panknin",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
ACER Failed To Converge for all Events (Trac #170) - AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/170">https://code.icecube.wisc.edu/projects/icecube/ticket/170</a>, reported by pankninand owned by panknin</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:08:15",
"_ts": "1547240895681654",
"description": "AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.",
"reporter": "panknin",
"cc": "",
"resolution": "fixed",
"time": "2009-10-27T10:10:23",
"component": "combo reconstruction",
"summary": "ACER Failed To Converge for all Events",
"priority": "critical",
"keywords": "",
"milestone": "",
"owner": "panknin",
"type": "defect"
}
```
</p>
</details>
|
defect
|
acer failed to converge for all events trac atmcscdeneryreco fails on all event even though they passed in earlier versions i likey induced some bug if not it has to be documented migrated from json status closed changetime ts description atmcscdeneryreco fails on all event even though they passed in earlier versions i likey induced some bug if not it has to be documented reporter panknin cc resolution fixed time component combo reconstruction summary acer failed to converge for all events priority critical keywords milestone owner panknin type defect
| 1
|
52,379
| 13,224,707,379
|
IssuesEvent
|
2020-08-17 19:40:59
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
Inconsistent segment parameter requirements (Trac #2135)
|
Incomplete Migration Migrated from Trac combo core defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2135">https://code.icecube.wisc.edu/projects/icecube/ticket/2135</a>, reported by cweaverand owned by cweaver</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:03:41",
"_ts": "1547240621058514",
"description": "`I3Tray.AddSegment` assumes that segments take two parameters (the tray and a name), but `icetray.traysegment` enforces only that they must have one parameter (the tray):\n\n{{{\n@icetray.traysegment\ndef mySegment(tray):\n\tpass\ntray=I3Tray()\ntray.AddSegment(mySegment)\n}}}\n\nresults in:\n\n{{{\n ...\n File \"/Users/cweaver/Work/IceCube/combo/build/lib/I3Tray.py\", line 224, in AddSegment\n return _segment(self, _name, **kwargs)\nTypeError: mySegment() takes exactly 1 argument (2 given)\n}}}\n\nThis is not caught by [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/python/traysegment.py#L23 the relevant check in icetray.traysegment]. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"time": "2018-02-12T23:26:32",
"component": "combo core",
"summary": "Inconsistent segment parameter requirements",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cweaver",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Inconsistent segment parameter requirements (Trac #2135) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2135">https://code.icecube.wisc.edu/projects/icecube/ticket/2135</a>, reported by cweaverand owned by cweaver</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:03:41",
"_ts": "1547240621058514",
"description": "`I3Tray.AddSegment` assumes that segments take two parameters (the tray and a name), but `icetray.traysegment` enforces only that they must have one parameter (the tray):\n\n{{{\n@icetray.traysegment\ndef mySegment(tray):\n\tpass\ntray=I3Tray()\ntray.AddSegment(mySegment)\n}}}\n\nresults in:\n\n{{{\n ...\n File \"/Users/cweaver/Work/IceCube/combo/build/lib/I3Tray.py\", line 224, in AddSegment\n return _segment(self, _name, **kwargs)\nTypeError: mySegment() takes exactly 1 argument (2 given)\n}}}\n\nThis is not caught by [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/python/traysegment.py#L23 the relevant check in icetray.traysegment]. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"time": "2018-02-12T23:26:32",
"component": "combo core",
"summary": "Inconsistent segment parameter requirements",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cweaver",
"type": "defect"
}
```
</p>
</details>
|
defect
|
inconsistent segment parameter requirements trac migrated from json status closed changetime ts description addsegment assumes that segments take two parameters the tray and a name but icetray traysegment enforces only that they must have one parameter the tray n n n icetray traysegment ndef mysegment tray n tpass ntray ntray addsegment mysegment n n nresults in n n n n file users cweaver work icecube combo build lib py line in addsegment n return segment self name kwargs ntypeerror mysegment takes exactly argument given n n nthis is not caught by reporter cweaver cc resolution fixed time component combo core summary inconsistent segment parameter requirements priority normal keywords milestone owner cweaver type defect
| 1
|
199,297
| 6,987,939,885
|
IssuesEvent
|
2017-12-14 10:59:15
|
canonical-websites/www.ubuntu.com
|
https://api.github.com/repos/canonical-websites/www.ubuntu.com
|
opened
|
All vertical pages should have a condensed version of their url
|
Priority: Medium
|
All vertical pages should have a condensed version of their url. For example, https://www.ubuntu.com/digital-signage should reach https://www.ubuntu.com/internet-of-things/digital-signage. Same for all the other vertical pages - eg; /financial-services, /telecommunications, /security, /digital-signage, /robotics, /gateways, /digital-signage, /education, /government, /enterprise
---
*Reported from: https://www.ubuntu.com/digital-signage*
|
1.0
|
All vertical pages should have a condensed version of their url - All vertical pages should have a condensed version of their url. For example, https://www.ubuntu.com/digital-signage should reach https://www.ubuntu.com/internet-of-things/digital-signage. Same for all the other vertical pages - eg; /financial-services, /telecommunications, /security, /digital-signage, /robotics, /gateways, /digital-signage, /education, /government, /enterprise
---
*Reported from: https://www.ubuntu.com/digital-signage*
|
non_defect
|
all vertical pages should have a condensed version of their url all vertical pages should have a condensed version of their url for example should reach same for all the other vertical pages eg financial services telecommunications security digital signage robotics gateways digital signage education government enterprise reported from
| 0
|
377,100
| 26,232,288,761
|
IssuesEvent
|
2023-01-05 02:04:09
|
firebase/flutterfire
|
https://api.github.com/repos/firebase/flutterfire
|
closed
|
[📚] Cloud Messaging - Notifications: More Android code sample broken
|
type: documentation good first issue blocked: customer-response Stale in triage
|
Please see #10128 for a related issue.
The ["Foreground Notifications" section](https://firebase.flutter.dev/docs/messaging/notifications#foreground-notifications) shows another Android code sample:
```dart
FirebaseMessaging.onMessage.listen((RemoteMessage message) {
RemoteNotification notification = message.notification;
AndroidNotification android = message.notification?.android;
// If `onMessage` is triggered with a notification, construct our own
// local notification to show to users using the created channel.
if (notification != null && android != null) {
flutterLocalNotificationsPlugin.show(
notification.hashCode,
notification.title,
notification.body,
NotificationDetails(
android: AndroidNotificationDetails(
channel.id,
channel.name,
channel.description,
icon: android?.smallIcon,
// other properties...
),
));
}
});
```
With `flutter_local_notifications: 12.0.4`, this doesn't compile anymore:
- `notification` and `android` are nullable.
- The channel description is now a named parameter.

Here's a version that does compile:
```dart
FirebaseMessaging.onMessage.listen((RemoteMessage message) {
RemoteNotification? notification = message.notification;
AndroidNotification? android = message.notification?.android;
// If `onMessage` is triggered with a notification, construct our own
// local notification to show to users using the created channel.
if (notification != null && android != null) {
flutterLocalNotificationsPlugin.show(
notification.hashCode,
notification.title,
notification.body,
NotificationDetails(
android: AndroidNotificationDetails(
channel.id,
channel.name,
channelDescription: channel.description,
icon: android.smallIcon,
// other properties...
),
));
}
});
```
|
1.0
|
[📚] Cloud Messaging - Notifications: More Android code sample broken - Please see #10128 for a related issue.
The ["Foreground Notifications" section](https://firebase.flutter.dev/docs/messaging/notifications#foreground-notifications) shows another Android code sample:
```dart
FirebaseMessaging.onMessage.listen((RemoteMessage message) {
RemoteNotification notification = message.notification;
AndroidNotification android = message.notification?.android;
// If `onMessage` is triggered with a notification, construct our own
// local notification to show to users using the created channel.
if (notification != null && android != null) {
flutterLocalNotificationsPlugin.show(
notification.hashCode,
notification.title,
notification.body,
NotificationDetails(
android: AndroidNotificationDetails(
channel.id,
channel.name,
channel.description,
icon: android?.smallIcon,
// other properties...
),
));
}
});
```
With `flutter_local_notifications: 12.0.4`, this doesn't compile anymore:
- `notification` and `android` are nullable.
- The channel description is now a named parameter.

Here's a version that does compile:
```dart
FirebaseMessaging.onMessage.listen((RemoteMessage message) {
RemoteNotification? notification = message.notification;
AndroidNotification? android = message.notification?.android;
// If `onMessage` is triggered with a notification, construct our own
// local notification to show to users using the created channel.
if (notification != null && android != null) {
flutterLocalNotificationsPlugin.show(
notification.hashCode,
notification.title,
notification.body,
NotificationDetails(
android: AndroidNotificationDetails(
channel.id,
channel.name,
channelDescription: channel.description,
icon: android.smallIcon,
// other properties...
),
));
}
});
```
|
non_defect
|
cloud messaging notifications more android code sample broken please see for a related issue the shows another android code sample dart firebasemessaging onmessage listen remotemessage message remotenotification notification message notification androidnotification android message notification android if onmessage is triggered with a notification construct our own local notification to show to users using the created channel if notification null android null flutterlocalnotificationsplugin show notification hashcode notification title notification body notificationdetails android androidnotificationdetails channel id channel name channel description icon android smallicon other properties with flutter local notifications this doesn t compile anymore notification and android are nullable the channel description is now a named parameter here s a version that does compile dart firebasemessaging onmessage listen remotemessage message remotenotification notification message notification androidnotification android message notification android if onmessage is triggered with a notification construct our own local notification to show to users using the created channel if notification null android null flutterlocalnotificationsplugin show notification hashcode notification title notification body notificationdetails android androidnotificationdetails channel id channel name channeldescription channel description icon android smallicon other properties
| 0
|
1,505
| 16,620,343,934
|
IssuesEvent
|
2021-06-02 23:21:17
|
timberio/vector
|
https://api.github.com/repos/timberio/vector
|
opened
|
BatchNotifier should support partial failures.
|
domain: performance domain: reliability type: enhancement
|
Currently, `BatchNotifier` can be attached to multiple events, but it does not report partial failures. This means that even if only one event in a batch fails, the entire batch could be marked as a failure.
In order to not only unlock higher-performance acknowledgement via batching, but to provide maximally correct processing, `BatchNotifier` should be modified/extended to support tracking the status of individual events that have been attached to it.
|
True
|
BatchNotifier should support partial failures. - Currently, `BatchNotifier` can be attached to multiple events, but it does not report partial failures. This means that even if only one event in a batch fails, the entire batch could be marked as a failure.
In order to not only unlock higher-performance acknowledgement via batching, but to provide maximally correct processing, `BatchNotifier` should be modified/extended to support tracking the status of individual events that have been attached to it.
|
non_defect
|
batchnotifier should support partial failures currently batchnotifier can be attached to multiple events but it does not report partial failures this means that even if only one event in a batch fails the entire batch could be marked as a failure in order to not only unlock higher performance acknowledgement via batching but to provide maximally correct processing batchnotifier should be modified extended to support tracking the status of individual events that have been attached to it
| 0
|
19,910
| 3,274,680,101
|
IssuesEvent
|
2015-10-26 12:18:22
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
Query inconsistency
|
Team: Core Team: QuSP Type: Critical Type: Defect
|
The behavior of retrieving data, deserializing it and setting it on the record is racy since an update could have happened in the same time. This can lead to permanently inconsistent query results.
|
1.0
|
Query inconsistency - The behavior of retrieving data, deserializing it and setting it on the record is racy since an update could have happened in the same time. This can lead to permanently inconsistent query results.
|
defect
|
query inconsistency the behavior of retrieving data deserializing it and setting it on the record is racy since an update could have happened in the same time this can lead to permanently inconsistent query results
| 1
|
143,777
| 22,154,480,401
|
IssuesEvent
|
2022-06-03 20:42:30
|
dotnet/aspnetcore
|
https://api.github.com/repos/dotnet/aspnetcore
|
opened
|
Use bool.TrueString and bool.FalseString instead of normal strings in HostingAbstractionsWebHostBuilderExtensions
|
design-proposal
|
## Summary
Hello there folks.
I think we use bool.TrueString and bool.FalseString instead of actual strings of boolean in this class
https://github.com/dotnet/aspnetcore/blob/main/src/Hosting/Abstractions/src/HostingAbstractionsWebHostBuilderExtensions.cs
## Motivation and goals
We can reduce magic strings.
We can Find references to those static fields easily.
We can reuse those fields.
## In scope
-
## Out of scope
-
## Risks / unknowns
I think there isn't any risks about this change.
## Examples
-
Please let me know and I will create a PR. Thanks
|
1.0
|
Use bool.TrueString and bool.FalseString instead of normal strings in HostingAbstractionsWebHostBuilderExtensions - ## Summary
Hello there folks.
I think we use bool.TrueString and bool.FalseString instead of actual strings of boolean in this class
https://github.com/dotnet/aspnetcore/blob/main/src/Hosting/Abstractions/src/HostingAbstractionsWebHostBuilderExtensions.cs
## Motivation and goals
We can reduce magic strings.
We can Find references to those static fields easily.
We can reuse those fields.
## In scope
-
## Out of scope
-
## Risks / unknowns
I think there isn't any risks about this change.
## Examples
-
Please let me know and I will create a PR. Thanks
|
non_defect
|
use bool truestring and bool falsestring instead of normal strings in hostingabstractionswebhostbuilderextensions summary hello there folks i think we use bool truestring and bool falsestring instead of actual strings of boolean in this class motivation and goals we can reduce magic strings we can find references to those static fields easily we can reuse those fields in scope out of scope risks unknowns i think there isn t any risks about this change examples please let me know and i will create a pr thanks
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.