Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
61,375
17,023,679,167
IssuesEvent
2021-07-03 03:15:43
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
[landcover] Car park area does not render on top of man_made=pier
Component: mapnik Priority: major Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 8.33pm, Sunday, 6th February 2011]** If you look at this car park [1] it has not rendered with a yellow area. I imagine this is because of the man_made=pier that it is on top of. See also [2] [1] http://www.openstreetmap.org/browse/way/10519366 [2] http://help.openstreetmap.org/questions/2733/why-are-these-car-parks-refusing-to-render
1.0
[landcover] Car park area does not render on top of man_made=pier - **[Submitted to the original trac issue database at 8.33pm, Sunday, 6th February 2011]** If you look at this car park [1] it has not rendered with a yellow area. I imagine this is because of the man_made=pier that it is on top of. See also [2] [1] http://www.openstreetmap.org/browse/way/10519366 [2] http://help.openstreetmap.org/questions/2733/why-are-these-car-parks-refusing-to-render
defect
car park area does not render on top of man made pier if you look at this car park it has not rendered with a yellow area i imagine this is because of the man made pier that it is on top of see also
1
181,332
14,860,165,997
IssuesEvent
2021-01-18 19:57:16
wando-advanced-robotics/infinite-recharge-robot
https://api.github.com/repos/wando-advanced-robotics/infinite-recharge-robot
opened
Enumerate minimum auto modes
documentation task
How many different auto modes do we think we'll need initially? How many for competition? Let's document those here.
1.0
Enumerate minimum auto modes - How many different auto modes do we think we'll need initially? How many for competition? Let's document those here.
non_defect
enumerate minimum auto modes how many different auto modes do we think we ll need initially how many for competition let s document those here
0
73,422
24,619,736,479
IssuesEvent
2022-10-15 19:21:42
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG:
defect
### Describe your issue. cython commit https://github.com/cython/cython/commit/77918c57b5f36bee708f625e2499bf05b23a87f1 in https://github.com/cython/cython/pull/4670 broke scipy installation. This is currently only on cython's master branch and not in an alpha release yet. I'm not sure if this is best fixed on the scipy or cython side. I reported this via back-channels to @rgommers. ### Reproducing Code Example ```python pip install -v --no-build-isolation . ``` with the development version of cython installed. ``` ### Error message ```shell The relevant error from compilation is : Error compiling Cython file: ------------------------------------------------------------ ... if name == NULL: name_copy = name else: name_copy = strdup(name) capsule = PyCapsule_New(func, name_copy, &raw_capsule_destructor) ^ ------------------------------------------------------------ /home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx:80:45: Cannot assign type 'void (*)(object) except *' to 'PyCapsule_Destructor' Traceback (most recent call last): File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 28, in <module> main() File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 20, in main sbp.run(['cython', '-3', '--fast-fail', File "/home/tcaswell/.pybuild/bleeding/lib/python3.12/subprocess.py", line 571, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['cython', '-3', '--fast-fail', '--output-file', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build/scipy/_lib/_ccallback_c.cpython-312-x86_64-linux-gnu.so.p/_ccallback_c.c', '--include-dir', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build', '/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx']' returned non-zero exit status 1. ``` ``` ### SciPy/NumPy/Python version information numpy main, cpython main, scipy main
1.0
BUG: - ### Describe your issue. cython commit https://github.com/cython/cython/commit/77918c57b5f36bee708f625e2499bf05b23a87f1 in https://github.com/cython/cython/pull/4670 broke scipy installation. This is currently only on cython's master branch and not in an alpha release yet. I'm not sure if this is best fixed on the scipy or cython side. I reported this via back-channels to @rgommers. ### Reproducing Code Example ```python pip install -v --no-build-isolation . ``` with the development version of cython installed. ``` ### Error message ```shell The relevant error from compilation is : Error compiling Cython file: ------------------------------------------------------------ ... if name == NULL: name_copy = name else: name_copy = strdup(name) capsule = PyCapsule_New(func, name_copy, &raw_capsule_destructor) ^ ------------------------------------------------------------ /home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx:80:45: Cannot assign type 'void (*)(object) except *' to 'PyCapsule_Destructor' Traceback (most recent call last): File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 28, in <module> main() File "/home/tcaswell/source/p/scipy/scipy/scipy/_build_utils/cythoner.py", line 20, in main sbp.run(['cython', '-3', '--fast-fail', File "/home/tcaswell/.pybuild/bleeding/lib/python3.12/subprocess.py", line 571, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['cython', '-3', '--fast-fail', '--output-file', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build/scipy/_lib/_ccallback_c.cpython-312-x86_64-linux-gnu.so.p/_ccallback_c.c', '--include-dir', '/home/tcaswell/source/p/scipy/scipy/.mesonpy-jnf3i747/build', '/home/tcaswell/source/p/scipy/scipy/scipy/_lib/_ccallback_c.pyx']' returned non-zero exit status 1. ``` ``` ### SciPy/NumPy/Python version information numpy main, cpython main, scipy main
defect
bug describe your issue cython commit in broke scipy installation this is currently only on cython s master branch and not in an alpha release yet i m not sure if this is best fixed on the scipy or cython side i reported this via back channels to rgommers reproducing code example python pip install v no build isolation with the development version of cython installed error message shell the relevant error from compilation is error compiling cython file if name null name copy name else name copy strdup name capsule pycapsule new func name copy raw capsule destructor home tcaswell source p scipy scipy scipy lib ccallback c pyx cannot assign type void object except to pycapsule destructor traceback most recent call last file home tcaswell source p scipy scipy scipy build utils cythoner py line in main file home tcaswell source p scipy scipy scipy build utils cythoner py line in main sbp run cython fast fail file home tcaswell pybuild bleeding lib subprocess py line in run raise calledprocesserror retcode process args subprocess calledprocesserror command returned non zero exit status scipy numpy python version information numpy main cpython main scipy main
1
20,201
3,314,991,510
IssuesEvent
2015-11-06 09:31:29
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
closed
PeakPicker algorithms don't use Factory pattern [194]
defect Good Ideas for upcoming releases minor OpenMS - library
Submitted by hendrikweisser on 2010-09-25 17:04:34 In contrast to all (?) other tools offering different algorithms (FeatureFinder, MapAligner, FeatureLinker etc.), the PeakPicker does not use the Factory design pattern.
1.0
PeakPicker algorithms don't use Factory pattern [194] - Submitted by hendrikweisser on 2010-09-25 17:04:34 In contrast to all (?) other tools offering different algorithms (FeatureFinder, MapAligner, FeatureLinker etc.), the PeakPicker does not use the Factory design pattern.
defect
peakpicker algorithms don t use factory pattern submitted by hendrikweisser on in contrast to all other tools offering different algorithms featurefinder mapaligner featurelinker etc the peakpicker does not use the factory design pattern
1
49,799
13,187,273,843
IssuesEvent
2020-08-13 02:53:40
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[hdfwriter] gets fatal RunTimeError, file not properly closed, message: "This is a BUG" (Trac #2104)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2104">https://code.icecube.wisc.edu/ticket/2104</a>, reported by lisa and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2019-03-27T16:30:49", "description": "I'm not sure if this is problem of the hdfwriter of the processing before. I didn't get any errors when I produced the i3 files. I can open them without problem and use them as i3 files without issues, but when I try to convert them to hdf5 I get this error.\n\nHere is the complete traceback:\n\nTraceback (most recent call last):\n File \"s0040-m_divide_data_years_in_LE_HE_low_high_samples_i3_to_hdf5_s0233.py\", line 128, in <module>\n tray.Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 209, in Physics\n return self.handle_frame(frame)\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 205, in handle_frame\n self.writer.convert(frame)\nRuntimeError: trying to get the address of unknown field OFUFilter_14 (in T* I3TableRow::GetPointer(const std::string&, size_t) [with T = bool])\nFATAL (I3TableService): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (I3TableService.cxx:219 in virtual I3TableService::~I3TableService())\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (in virtual I3TableService::~I3TableService())\n", "reporter": "lisa", "cc": "", "resolution": "fixed", "_ts": "1553704249811581", "component": "combo reconstruction", "summary": "[hdfwriter] gets fatal RunTimeError, file not properly closed, message: \"This is a BUG\"", "priority": "critical", "keywords": "", "time": "2017-10-22T10:57:04", "milestone": "Vernal Equinox 2019", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[hdfwriter] gets fatal RunTimeError, file not properly closed, message: "This is a BUG" (Trac #2104) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2104">https://code.icecube.wisc.edu/ticket/2104</a>, reported by lisa and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2019-03-27T16:30:49", "description": "I'm not sure if this is problem of the hdfwriter of the processing before. I didn't get any errors when I produced the i3 files. I can open them without problem and use them as i3 files without issues, but when I try to convert them to hdf5 I get this error.\n\nHere is the complete traceback:\n\nTraceback (most recent call last):\n File \"s0040-m_divide_data_years_in_LE_HE_low_high_samples_i3_to_hdf5_s0233.py\", line 128, in <module>\n tray.Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 209, in Physics\n return self.handle_frame(frame)\n File \"/home/lunger/data/meta-projects/processing/release/V05-00-00/trunk/build/lib/icecube/tableio/I3TableWriterModule.py\", line 205, in handle_frame\n self.writer.convert(frame)\nRuntimeError: trying to get the address of unknown field OFUFilter_14 (in T* I3TableRow::GetPointer(const std::string&, size_t) [with T = bool])\nFATAL (I3TableService): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (I3TableService.cxx:219 in virtual I3TableService::~I3TableService())\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Table 'AtmCscdEnergyReco_L2' is still connected, which means that the output file was never properly closed. This is a BUG!\n (in virtual I3TableService::~I3TableService())\n", "reporter": "lisa", "cc": "", "resolution": "fixed", "_ts": "1553704249811581", "component": "combo reconstruction", "summary": "[hdfwriter] gets fatal RunTimeError, file not properly closed, message: \"This is a BUG\"", "priority": "critical", "keywords": "", "time": "2017-10-22T10:57:04", "milestone": "Vernal Equinox 2019", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
gets fatal runtimeerror file not properly closed message this is a bug trac migrated from json status closed changetime description i m not sure if this is problem of the hdfwriter of the processing before i didn t get any errors when i produced the files i can open them without problem and use them as files without issues but when i try to convert them to i get this error n nhere is the complete traceback n ntraceback most recent call last n file m divide data years in le he low high samples to py line in n tray execute n file home lunger data meta projects processing release trunk build lib py line in execute n super self execute n file home lunger data meta projects processing release trunk build lib icecube tableio py line in physics n return self handle frame frame n file home lunger data meta projects processing release trunk build lib icecube tableio py line in handle frame n self writer convert frame nruntimeerror trying to get the address of unknown field ofufilter in t getpointer const std string size t nfatal table atmcscdenergyreco is still connected which means that the output file was never properly closed this is a bug n cxx in virtual nterminate called after throwing an instance of std runtime error n what table atmcscdenergyreco is still connected which means that the output file was never properly closed this is a bug n in virtual n reporter lisa cc resolution fixed ts component combo reconstruction summary gets fatal runtimeerror file not properly closed message this is a bug priority critical keywords time milestone vernal equinox owner jvansanten type defect
1
57,641
15,889,780,805
IssuesEvent
2021-04-10 12:58:07
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
Highlight new users who join the voice/web
design: user experience module: client priority: normal type: defect
Originally reported on Google Code with ID 757 ``` What steps will reproduce the problem? 1. Have 20 users in the Listener's window 2. A new user joins 3. It's really hard to see who was the 21st user who joined. If BigBlueButton could highlight the new user for a few moments, such as a different color that fades after a few moments, it would be much easier to pick out the new users in the list. Same goes for the Users window. ``` Reported by `ffdixon` on 2010-11-23 23:47:42
1.0
Highlight new users who join the voice/web - Originally reported on Google Code with ID 757 ``` What steps will reproduce the problem? 1. Have 20 users in the Listener's window 2. A new user joins 3. It's really hard to see who was the 21st user who joined. If BigBlueButton could highlight the new user for a few moments, such as a different color that fades after a few moments, it would be much easier to pick out the new users in the list. Same goes for the Users window. ``` Reported by `ffdixon` on 2010-11-23 23:47:42
defect
highlight new users who join the voice web originally reported on google code with id what steps will reproduce the problem have users in the listener s window a new user joins it s really hard to see who was the user who joined if bigbluebutton could highlight the new user for a few moments such as a different color that fades after a few moments it would be much easier to pick out the new users in the list same goes for the users window reported by ffdixon on
1
437,309
12,576,527,270
IssuesEvent
2020-06-09 08:02:56
kubeflow/manifests
https://api.github.com/repos/kubeflow/manifests
closed
Jupyter web app config not compatible with latest jupyter web app image
area/jupyter kind/bug lifecycle/stale priority/p2
Got error when click `New Server` `KeyError: 'notebook-name'` `notebook-name` is not in client side request. Related: https://github.com/kubeflow/manifests/pull/163
1.0
Jupyter web app config not compatible with latest jupyter web app image - Got error when click `New Server` `KeyError: 'notebook-name'` `notebook-name` is not in client side request. Related: https://github.com/kubeflow/manifests/pull/163
non_defect
jupyter web app config not compatible with latest jupyter web app image got error when click new server keyerror notebook name notebook name is not in client side request related
0
10,067
2,618,934,329
IssuesEvent
2015-03-03 00:01:28
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
Impossible to create new fleet
auto-migrated Priority-Medium Type-Defect
``` Game version: 0.95.130 Operating System: (e.g., Windows 7 x86, Windows XP 64-bit) Fedora 19 x86_64, but I suspect this is unrelated to the operating system. Java runtime version: (run java -version) 1.6 and 1.7 openJDK, but I suspect this has nothing to do with the JVM Installed using the Launcher? (yes, no) Yes Game language (en, hu, de): en (I suspect all) What steps will reproduce the problem? 1. Install the game 2. Run the game 3. Try to find some way to create a fleet without an already existing fleet 4. If 3 fails, give up on the game. What is the expected output? What do you see instead? There should be some -easy to find- button which will allow me to create a fleet. No such button is present Please provide any additional information below. Please upload any save before and/or after the problem happened. Please attach the open-ig.log file found in the application's directory. ``` Original issue reported on code.google.com by `frederik...@gmail.com` on 24 Jul 2013 at 7:40
1.0
Impossible to create new fleet - ``` Game version: 0.95.130 Operating System: (e.g., Windows 7 x86, Windows XP 64-bit) Fedora 19 x86_64, but I suspect this is unrelated to the operating system. Java runtime version: (run java -version) 1.6 and 1.7 openJDK, but I suspect this has nothing to do with the JVM Installed using the Launcher? (yes, no) Yes Game language (en, hu, de): en (I suspect all) What steps will reproduce the problem? 1. Install the game 2. Run the game 3. Try to find some way to create a fleet without an already existing fleet 4. If 3 fails, give up on the game. What is the expected output? What do you see instead? There should be some -easy to find- button which will allow me to create a fleet. No such button is present Please provide any additional information below. Please upload any save before and/or after the problem happened. Please attach the open-ig.log file found in the application's directory. ``` Original issue reported on code.google.com by `frederik...@gmail.com` on 24 Jul 2013 at 7:40
defect
impossible to create new fleet game version operating system e g windows windows xp bit fedora but i suspect this is unrelated to the operating system java runtime version run java version and openjdk but i suspect this has nothing to do with the jvm installed using the launcher yes no yes game language en hu de en i suspect all what steps will reproduce the problem install the game run the game try to find some way to create a fleet without an already existing fleet if fails give up on the game what is the expected output what do you see instead there should be some easy to find button which will allow me to create a fleet no such button is present please provide any additional information below please upload any save before and or after the problem happened please attach the open ig log file found in the application s directory original issue reported on code google com by frederik gmail com on jul at
1
363,893
25,472,705,770
IssuesEvent
2022-11-25 11:37:26
UnBArqDsw2022-2/2022.2_G4_IDotPet
https://api.github.com/repos/UnBArqDsw2022-2/2022.2_G4_IDotPet
opened
Correção do Léxico
documentation
<!-- Certifique-se de ser uma tarefa bem contida. Certifique-se da possibilidade de desmembrar a issue em issues melhores e, caso seja possível, o faça --> # Descrição: <!-- Dê os objetivos da issue, para quê serve, qual issue afeta, que artefato é criado/alterado. Ao citar outras issues, utilize o código delas: #Número_da_issue --> Essa issue tem como objetivo corrigir o Léxico conforme recomendações da apresentação base. # Tarefas: <!-- Não economize na listagem de tarefas --> - [ ] Revisar e corrigir Léxico (Nicolas e Herick) # Critério de aceitação: <!-- O revisor da issue deve marcar esses critérios antes de fechá-la --> - [ ] Documento de Léxico revisado e corrigido.
1.0
Correção do Léxico - <!-- Certifique-se de ser uma tarefa bem contida. Certifique-se da possibilidade de desmembrar a issue em issues melhores e, caso seja possível, o faça --> # Descrição: <!-- Dê os objetivos da issue, para quê serve, qual issue afeta, que artefato é criado/alterado. Ao citar outras issues, utilize o código delas: #Número_da_issue --> Essa issue tem como objetivo corrigir o Léxico conforme recomendações da apresentação base. # Tarefas: <!-- Não economize na listagem de tarefas --> - [ ] Revisar e corrigir Léxico (Nicolas e Herick) # Critério de aceitação: <!-- O revisor da issue deve marcar esses critérios antes de fechá-la --> - [ ] Documento de Léxico revisado e corrigido.
non_defect
correção do léxico descrição essa issue tem como objetivo corrigir o léxico conforme recomendações da apresentação base tarefas revisar e corrigir léxico nicolas e herick critério de aceitação documento de léxico revisado e corrigido
0
653,463
21,582,845,927
IssuesEvent
2022-05-02 20:44:25
RAF-SI-2021/Banka-Back
https://api.github.com/repos/RAF-SI-2021/Banka-Back
closed
Promena naziva klasa, metoda i atributa da koriste camelCase
area/backend priority/critical
Java generalno koristi camelCase za nazive klasa, metoda, atributa. Mi trenutno koristimo snake_case, što prouzrokuje dosta problema pogotova sa Hibernate-om. Potrebno je promeniti sve nazive u camelCase (npr. `hartija_od_vrednosti_id` u `hartijaOdVrednostiId`).
1.0
Promena naziva klasa, metoda i atributa da koriste camelCase - Java generalno koristi camelCase za nazive klasa, metoda, atributa. Mi trenutno koristimo snake_case, što prouzrokuje dosta problema pogotova sa Hibernate-om. Potrebno je promeniti sve nazive u camelCase (npr. `hartija_od_vrednosti_id` u `hartijaOdVrednostiId`).
non_defect
promena naziva klasa metoda i atributa da koriste camelcase java generalno koristi camelcase za nazive klasa metoda atributa mi trenutno koristimo snake case što prouzrokuje dosta problema pogotova sa hibernate om potrebno je promeniti sve nazive u camelcase npr hartija od vrednosti id u hartijaodvrednostiid
0
27,885
5,117,890,683
IssuesEvent
2017-01-07 21:46:35
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
restart self video window again after closing fails
Accepted Defect Low Priority Usability Video
Originally reported on Google Code with ID 1554 ``` closing self video, (use X on video window,) and immediately trying to start video again, will cause the video button to change icon but video will not start, and button functionality stops. ``` Reported by `yossia6` on 2013-06-12 07:09:21
1.0
restart self video window again after closing fails - Originally reported on Google Code with ID 1554 ``` closing self video, (use X on video window,) and immediately trying to start video again, will cause the video button to change icon but video will not start, and button functionality stops. ``` Reported by `yossia6` on 2013-06-12 07:09:21
defect
restart self video window again after closing fails originally reported on google code with id closing self video use x on video window and immediately trying to start video again will cause the video button to change icon but video will not start and button functionality stops reported by on
1
2,272
2,603,992,125
IssuesEvent
2015-02-24 19:06:53
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳好的疱疹专科医院
auto-migrated Priority-Medium Type-Defect
``` 沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:36
1.0
沈阳好的疱疹专科医院 - ``` 沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:36
defect
沈阳好的疱疹专科医院 沈阳好的疱疹专科医院〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
1
30,990
6,388,361,896
IssuesEvent
2017-08-03 15:25:41
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
Firefox: synthetic keyboard events do not set KeyboardEvent.key property properly
D-firefox E-less easy I-defect
Selenium 2.45.0 uses synthetic keyboard events for Firefox 36.0.4 on all(?) platforms. These synthetic keyboard events set the javascript [KeyboardEvent.key](https://developer.mozilla.org/en-US/docs/Web/API/KeyboardEvent/key) property always to 'Unidentified' instead of the expected character. Using native events with Firefox <= 34.0.5 on Windows 7 the following phpunit test run successfully: ``` php <?php class KeyTest extends PHPUnit_Extensions_Selenium2TestCase { protected function setUp() { $this->setBrowser('firefox'); $this->setHost('localhost'); $this->setBrowserUrl('http://unixpapa.com/js/testkey.html'); } public function testKeys() { $this->url('http://unixpapa.com/js/testkey.html'); $this->byName('classic')->click(); $this->byName('dom3')->click(); $this->keys('z'); $this->assertContains('key=z', $this->byTag('textarea')->value()); } } ``` As far as I can see, it's related to the use of `sendKeyEvent` in favour to `sendNativeKeyEvent` in https://github.com/SeleniumHQ/selenium/blob/cf26e94543b7c838996a21995fcb8bbe11a800bd/javascript/firefox-driver/js/utils.js#L671. While testing `sendNativeKeyEvent` in Firefox's Scratchpad I had mixed results: - Windows: working - Linux: Exception - MacOS: obviously working (key is send to scratchpad window instead of webpage)
1.0
Firefox: synthetic keyboard events do not set KeyboardEvent.key property properly - Selenium 2.45.0 uses synthetic keyboard events for Firefox 36.0.4 on all(?) platforms. These synthetic keyboard events set the javascript [KeyboardEvent.key](https://developer.mozilla.org/en-US/docs/Web/API/KeyboardEvent/key) property always to 'Unidentified' instead of the expected character. Using native events with Firefox <= 34.0.5 on Windows 7 the following phpunit test run successfully: ``` php <?php class KeyTest extends PHPUnit_Extensions_Selenium2TestCase { protected function setUp() { $this->setBrowser('firefox'); $this->setHost('localhost'); $this->setBrowserUrl('http://unixpapa.com/js/testkey.html'); } public function testKeys() { $this->url('http://unixpapa.com/js/testkey.html'); $this->byName('classic')->click(); $this->byName('dom3')->click(); $this->keys('z'); $this->assertContains('key=z', $this->byTag('textarea')->value()); } } ``` As far as I can see, it's related to the use of `sendKeyEvent` in favour to `sendNativeKeyEvent` in https://github.com/SeleniumHQ/selenium/blob/cf26e94543b7c838996a21995fcb8bbe11a800bd/javascript/firefox-driver/js/utils.js#L671. While testing `sendNativeKeyEvent` in Firefox's Scratchpad I had mixed results: - Windows: working - Linux: Exception - MacOS: obviously working (key is send to scratchpad window instead of webpage)
defect
firefox synthetic keyboard events do not set keyboardevent key property properly selenium uses synthetic keyboard events for firefox on all platforms these synthetic keyboard events set the javascript property always to unidentified instead of the expected character using native events with firefox on windows the following phpunit test run successfully php php class keytest extends phpunit extensions protected function setup this setbrowser firefox this sethost localhost this setbrowserurl public function testkeys this url this byname classic click this byname click this keys z this assertcontains key z this bytag textarea value as far as i can see it s related to the use of sendkeyevent in favour to sendnativekeyevent in while testing sendnativekeyevent in firefox s scratchpad i had mixed results windows working linux exception macos obviously working key is send to scratchpad window instead of webpage
1
225,707
24,881,192,684
IssuesEvent
2022-10-28 01:21:39
tomdgl397/juice-shop
https://api.github.com/repos/tomdgl397/juice-shop
opened
CVE-2022-40304 (High) detected in reactos0.4.13-dev
security vulnerability
## CVE-2022-40304 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>reactos0.4.13-dev</b></p></summary> <p> <p>A free Windows-compatible Operating System</p> <p>Library home page: <a href=https://github.com/reactos/reactos.git>https://github.com/reactos/reactos.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/libxmljs2/vendor/libxml/entities.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> CVE-2022-40304 libxml2: dict corruption caused by entity reference cycles <p>Publish Date: 2022-09-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40304>CVE-2022-40304</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-09</p> <p>Fix Resolution: v2.10.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-40304 (High) detected in reactos0.4.13-dev - ## CVE-2022-40304 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>reactos0.4.13-dev</b></p></summary> <p> <p>A free Windows-compatible Operating System</p> <p>Library home page: <a href=https://github.com/reactos/reactos.git>https://github.com/reactos/reactos.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/libxmljs2/vendor/libxml/entities.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> CVE-2022-40304 libxml2: dict corruption caused by entity reference cycles <p>Publish Date: 2022-09-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40304>CVE-2022-40304</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-09</p> <p>Fix Resolution: v2.10.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in dev cve high severity vulnerability vulnerable library dev a free windows compatible operating system library home page a href found in head commit a href vulnerable source files node modules vendor libxml entities c vulnerability details cve dict corruption caused by entity reference cycles publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend
0
58,381
16,515,154,407
IssuesEvent
2021-05-26 09:06:59
galasa-dev/projectmanagement
https://api.github.com/repos/galasa-dev/projectmanagement
opened
Find a way to have the pom name/derscription set correctly using gradle build convention plugins
Framework defect
For the 16 release, needs to change the gradle builds so that the name and description is set correctly for every project, even when using the new build convention plugions. Assigning to James so that he learns more about the gradle plugins and the build convention method.
1.0
Find a way to have the pom name/derscription set correctly using gradle build convention plugins - For the 16 release, needs to change the gradle builds so that the name and description is set correctly for every project, even when using the new build convention plugions. Assigning to James so that he learns more about the gradle plugins and the build convention method.
defect
find a way to have the pom name derscription set correctly using gradle build convention plugins for the release needs to change the gradle builds so that the name and description is set correctly for every project even when using the new build convention plugions assigning to james so that he learns more about the gradle plugins and the build convention method
1
74,578
25,188,478,689
IssuesEvent
2022-11-11 20:48:44
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Design | Profile | Accessibility | Mobile #
design 508/Accessibility authenticated-experience profile 508-defect-4
## Background Review design changes for disability rating and full name being added to the main sections of profile. We will need to discuss these changes with the content and IA folks and get their feedback, then make adjustments to the design as necessary. Also, determine if we should only show this if we know user is a Veteran or always show. ## Tasks - [ ] Review designs with Content/IA - [ ] Document feedback - [ ] Adjust designs as necessary ## Acceptance Criteria - [ ] Designs for full name/disability rating are reviewed by content and IA ## Associated Tickets - #48264
1.0
Design | Profile | Accessibility | Mobile # - ## Background Review design changes for disability rating and full name being added to the main sections of profile. We will need to discuss these changes with the content and IA folks and get their feedback, then make adjustments to the design as necessary. Also, determine if we should only show this if we know user is a Veteran or always show. ## Tasks - [ ] Review designs with Content/IA - [ ] Document feedback - [ ] Adjust designs as necessary ## Acceptance Criteria - [ ] Designs for full name/disability rating are reviewed by content and IA ## Associated Tickets - #48264
defect
design profile accessibility mobile background review design changes for disability rating and full name being added to the main sections of profile we will need to discuss these changes with the content and ia folks and get their feedback then make adjustments to the design as necessary also determine if we should only show this if we know user is a veteran or always show tasks review designs with content ia document feedback adjust designs as necessary acceptance criteria designs for full name disability rating are reviewed by content and ia associated tickets
1
73,932
15,286,689,075
IssuesEvent
2021-02-23 14:58:35
Baneeishaque/Android-Common-Utils4
https://api.github.com/repos/Baneeishaque/Android-Common-Utils4
opened
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar
security vulnerability
## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: Android-Common-Utils4/common4/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.10/4b95f4897fa13f2cd904aee711aeafc0c5295cd8/commons-codec-1.10.jar</p> <p> Dependency Hierarchy: - lint-gradle-27.1.2.jar (Root Library) - sdk-common-27.1.2.jar - sdklib-27.1.2.jar - httpmime-4.5.6.jar - httpclient-4.5.6.jar - :x: **commons-codec-1.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils4/commit/e8328a1ad0239aaf16726916e083bb49de9642ec">e8328a1ad0239aaf16726916e083bb49de9642ec</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-20</p> <p>Fix Resolution: commons-codec:commons-codec:1.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar - ## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: Android-Common-Utils4/common4/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.10/4b95f4897fa13f2cd904aee711aeafc0c5295cd8/commons-codec-1.10.jar</p> <p> Dependency Hierarchy: - lint-gradle-27.1.2.jar (Root Library) - sdk-common-27.1.2.jar - sdklib-27.1.2.jar - httpmime-4.5.6.jar - httpclient-4.5.6.jar - :x: **commons-codec-1.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils4/commit/e8328a1ad0239aaf16726916e083bb49de9642ec">e8328a1ad0239aaf16726916e083bb49de9642ec</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-20</p> <p>Fix Resolution: commons-codec:commons-codec:1.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file android common build gradle path to vulnerable library home wss scanner gradle caches modules files commons codec commons codec commons codec jar dependency hierarchy lint gradle jar root library sdk common jar sdklib jar httpmime jar httpclient jar x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons codec commons codec step up your open source security game with whitesource
0
53,853
13,262,390,222
IssuesEvent
2020-08-20 21:41:53
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__' (Trac #2197)
Migrated from Trac defect infrastructure
Hi, I saw this error while experimenting with the trunk of filterscripts: ```text Traceback (most recent call last): File "./IceCube_BaseProc_only_MM_trunk.py", line 309, in <module> tray.Execute() File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py", line 256, in Execute super(I3Tray, self).Execute() File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None) File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__ self.threadName = threading.current_thread().name File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread return _active[_get_ident()] File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None) File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__ self.threadName = threading.current_thread().name File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread return _active[_get_ident()] File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py", line 26, in MaskMaker g = omgeo[omkey] TypeError: 'I3Geometry' object has no attribute '__getitem__' ``` I think the problem is line 26 in http://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py and should be fixed by replacing ``` g = omgeo[omkey] ``` with ```text g = omgeo.omgeo[omkey] ``` (thanks to alex olivias and jvs for pointing this out to me) Cheers, Frederik <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2197">https://code.icecube.wisc.edu/projects/icecube/ticket/2197</a>, reported by flauberand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-12T03:53:01", "_ts": "1547265181974434", "description": "Hi,\n\nI saw this error while experimenting with the trunk of filterscripts:\n\n{{{\nTraceback (most recent call last):\n File \"./IceCube_BaseProc_only_MM_trunk.py\", line 309, in <module>\n tray.Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py\", line 26, in MaskMaker\n g = omgeo[omkey]\nTypeError: 'I3Geometry' object has no attribute '__getitem__'\n}}}\n\nI think the problem is line 26 in \nhttp://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py\n\nand should be fixed by replacing\n{{{ \ng = omgeo[omkey]\n}}}\nwith \n{{{\ng = omgeo.omgeo[omkey]\n}}}\n(thanks to alex olivias and jvs for pointing this out to me)\nCheers,\nFrederik\n\n\n", "reporter": "flauber", "cc": "", "resolution": "fixed", "time": "2018-10-09T14:06:46", "component": "infrastructure", "summary": "filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__'", "priority": "normal", "keywords": "filterscripts", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__' (Trac #2197) - Hi, I saw this error while experimenting with the trunk of filterscripts: ```text Traceback (most recent call last): File "./IceCube_BaseProc_only_MM_trunk.py", line 309, in <module> tray.Execute() File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py", line 256, in Execute super(I3Tray, self).Execute() File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None) File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__ self.threadName = threading.current_thread().name File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread return _active[_get_ident()] File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py", line 28, in log record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None) File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py", line 285, in __init__ self.threadName = threading.current_thread().name File "/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py", line 1151, in currentThread return _active[_get_ident()] File "/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py", line 26, in MaskMaker g = omgeo[omkey] TypeError: 'I3Geometry' object has no attribute '__getitem__' ``` I think the problem is line 26 in http://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py and should be fixed by replacing ``` g = omgeo[omkey] ``` with ```text g = omgeo.omgeo[omkey] ``` (thanks to alex olivias and jvs for pointing this out to me) Cheers, Frederik <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2197">https://code.icecube.wisc.edu/projects/icecube/ticket/2197</a>, reported by flauberand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-12T03:53:01", "_ts": "1547265181974434", "description": "Hi,\n\nI saw this error while experimenting with the trunk of filterscripts:\n\n{{{\nTraceback (most recent call last):\n File \"./IceCube_BaseProc_only_MM_trunk.py\", line 309, in <module>\n tray.Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/icetray/i3logging.py\", line 28, in log\n record = logging.LogRecord(logger.name, self.pylevels[level], file, line, msg, tuple(), None, None)\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/logging/__init__.py\", line 285, in __init__\n self.threadName = threading.current_thread().name\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/RHEL_7_x86_64/lib/python2.7/threading.py\", line 1151, in currentThread\n return _active[_get_ident()]\n File \"/data/user/flauber/software/py2-v3.0.1/icerec_trunk/build/lib/icecube/filterscripts/baseproc_superdst.py\", line 26, in MaskMaker\n g = omgeo[omkey]\nTypeError: 'I3Geometry' object has no attribute '__getitem__'\n}}}\n\nI think the problem is line 26 in \nhttp://code.icecube.wisc.edu/svn/projects/filterscripts/trunk/python/baseproc_superdst.py\n\nand should be fixed by replacing\n{{{ \ng = omgeo[omkey]\n}}}\nwith \n{{{\ng = omgeo.omgeo[omkey]\n}}}\n(thanks to alex olivias and jvs for pointing this out to me)\nCheers,\nFrederik\n\n\n", "reporter": "flauber", "cc": "", "resolution": "fixed", "time": "2018-10-09T14:06:46", "component": "infrastructure", "summary": "filterscripts: TypeError: 'I3Geometry' object has no attribute '__getitem__'", "priority": "normal", "keywords": "filterscripts", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
filterscripts typeerror object has no attribute getitem trac hi i saw this error while experimenting with the trunk of filterscripts text traceback most recent call last file icecube baseproc only mm trunk py line in tray execute file data user flauber software icerec trunk build lib py line in execute super self execute file data user flauber software icerec trunk build lib icecube icetray py line in log record logging logrecord logger name self pylevels file line msg tuple none none file cvmfs icecube opensciencegrid org rhel lib logging init py line in init self threadname threading current thread name file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread return active file data user flauber software icerec trunk build lib icecube icetray py line in log record logging logrecord logger name self pylevels file line msg tuple none none file cvmfs icecube opensciencegrid org rhel lib logging init py line in init self threadname threading current thread name file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread return active file data user flauber software icerec trunk build lib icecube filterscripts baseproc superdst py line in maskmaker g omgeo typeerror object has no attribute getitem i think the problem is line in and should be fixed by replacing g omgeo with text g omgeo omgeo thanks to alex olivias and jvs for pointing this out to me cheers frederik migrated from json status closed changetime ts description hi n ni saw this error while experimenting with the trunk of filterscripts n n ntraceback most recent call last n file icecube baseproc only mm trunk py line in n tray execute n file data user flauber software icerec trunk build lib py line in execute n super self execute n file data user flauber software icerec trunk build lib icecube icetray py line in log n record logging logrecord logger name self pylevels file line msg tuple none none n file cvmfs icecube opensciencegrid org rhel lib logging init py line in init n self threadname threading current thread name n file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread n return active n file data user flauber software icerec trunk build lib icecube icetray py line in log n record logging logrecord logger name self pylevels file line msg tuple none none n file cvmfs icecube opensciencegrid org rhel lib logging init py line in init n self threadname threading current thread name n file cvmfs icecube opensciencegrid org rhel lib threading py line in currentthread n return active n file data user flauber software icerec trunk build lib icecube filterscripts baseproc superdst py line in maskmaker n g omgeo ntypeerror object has no attribute getitem n n ni think the problem is line in n should be fixed by replacing n ng omgeo n nwith n ng omgeo omgeo n n thanks to alex olivias and jvs for pointing this out to me ncheers nfrederik n n n reporter flauber cc resolution fixed time component infrastructure summary filterscripts typeerror object has no attribute getitem priority normal keywords filterscripts milestone owner olivas type defect
1
56,332
15,023,514,986
IssuesEvent
2021-02-01 18:20:09
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
508-defect-3 [COGNITION]: COVID Vaccine Trials - Checkboxes should not be separated from additional information inputs
508-defect-3 508-issue-cognition 508/Accessibility covid-vaccine-trials
# [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## User Story or Problem Statement <!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. --> As a user, I want to make sure all of my answers are recorded. When additional information fields appear away from the question I am answering, I might miss them. ## Details <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> The "What is your gender?" question gives users an option to self-describe. If users check this option, an additional text input appears after the next checkbox. The input should come directly after this checkbox. An option might be to put the "Prefer to self-describe" checkbox last, after Prefer not to answer. Screenshot attached below. ## Acceptance Criteria - [ ] Checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input ## Environment * https://staging.va.gov/coronavirus-research/volunteer/sign-up ## WCAG or Vendor Guidance (optional) * [Meaningful Sequence: Understanding SC 1.3.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-sequence.html) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2021-01-07 at 4 48 44 PM](https://user-images.githubusercontent.com/934879/103954285-7e45ef00-5109-11eb-82ce-156b68823c3f.png)
1.0
508-defect-3 [COGNITION]: COVID Vaccine Trials - Checkboxes should not be separated from additional information inputs - # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## User Story or Problem Statement <!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. --> As a user, I want to make sure all of my answers are recorded. When additional information fields appear away from the question I am answering, I might miss them. ## Details <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> The "What is your gender?" question gives users an option to self-describe. If users check this option, an additional text input appears after the next checkbox. The input should come directly after this checkbox. An option might be to put the "Prefer to self-describe" checkbox last, after Prefer not to answer. Screenshot attached below. ## Acceptance Criteria - [ ] Checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input ## Environment * https://staging.va.gov/coronavirus-research/volunteer/sign-up ## WCAG or Vendor Guidance (optional) * [Meaningful Sequence: Understanding SC 1.3.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-sequence.html) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2021-01-07 at 4 48 44 PM](https://user-images.githubusercontent.com/934879/103954285-7e45ef00-5109-11eb-82ce-156b68823c3f.png)
defect
defect covid vaccine trials checkboxes should not be separated from additional information inputs enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact trevor user story or problem statement as a user i want to make sure all of my answers are recorded when additional information fields appear away from the question i am answering i might miss them details the what is your gender question gives users an option to self describe if users check this option an additional text input appears after the next checkbox the input should come directly after this checkbox an option might be to put the prefer to self describe checkbox last after prefer not to answer screenshot attached below acceptance criteria checkboxes are reordered or the team decides on another course of action that makes the relationship clear between the checkbox and its additional text input environment wcag or vendor guidance optional screenshots or trace logs
1
38,722
8,527,328,041
IssuesEvent
2018-11-02 19:07:13
ccharlier/orionbanque
https://api.github.com/repos/ccharlier/orionbanque
closed
Fix 1 Maintainability, 6 Style issues in OrionBanque\Classe\Configuration.cs
codefactor
[CodeFactor](https://www.codefactor.io/repository/github/ccharlier/orionbanque/overview/master) found multiple issues: #### A C# document may only contain a single class at the root level unless all of the classes are partial and are of the same type. [OrionBanque\Classe\Configuration.cs:19 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L19) #### A closing curly bracket must not be preceded by a blank line. [OrionBanque\Classe\Configuration.cs:287 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L287) #### The code must not contain multiple blank lines in a row. [OrionBanque\Classe\Configuration.cs:5 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L5)[OrionBanque\Classe\Configuration.cs:199 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L199) #### An opening curly bracket must not be followed by a blank line. [OrionBanque\Classe\Configuration.cs:219 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L219) #### Invalid spacing around the semicolon. [OrionBanque\Classe\Configuration.cs:195 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)[OrionBanque\Classe\Configuration.cs:195 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)
1.0
Fix 1 Maintainability, 6 Style issues in OrionBanque\Classe\Configuration.cs - [CodeFactor](https://www.codefactor.io/repository/github/ccharlier/orionbanque/overview/master) found multiple issues: #### A C# document may only contain a single class at the root level unless all of the classes are partial and are of the same type. [OrionBanque\Classe\Configuration.cs:19 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L19) #### A closing curly bracket must not be preceded by a blank line. [OrionBanque\Classe\Configuration.cs:287 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L287) #### The code must not contain multiple blank lines in a row. [OrionBanque\Classe\Configuration.cs:5 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L5)[OrionBanque\Classe\Configuration.cs:199 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L199) #### An opening curly bracket must not be followed by a blank line. [OrionBanque\Classe\Configuration.cs:219 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L219) #### Invalid spacing around the semicolon. [OrionBanque\Classe\Configuration.cs:195 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)[OrionBanque\Classe\Configuration.cs:195 ](https://www.codefactor.io/repository/github/ccharlier/orionbanque/source/master/OrionBanque/Classe/Configuration.cs#L195)
non_defect
fix maintainability style issues in orionbanque classe configuration cs found multiple issues a c document may only contain a single class at the root level unless all of the classes are partial and are of the same type orionbanque classe configuration cs a closing curly bracket must not be preceded by a blank line orionbanque classe configuration cs the code must not contain multiple blank lines in a row orionbanque classe configuration cs an opening curly bracket must not be followed by a blank line orionbanque classe configuration cs invalid spacing around the semicolon orionbanque classe configuration cs
0
78,730
15,586,068,790
IssuesEvent
2021-03-18 01:06:01
Pr0methean/s5
https://api.github.com/repos/Pr0methean/s5
opened
CVE-2019-3875 (Medium) detected in keycloak-core-4.7.0.Final.jar
security vulnerability
## CVE-2019-3875 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>keycloak-core-4.7.0.Final.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://keycloak.org/keycloak-core">http://keycloak.org/keycloak-core</a></p> <p>Path to dependency file: /s5/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/keycloak/keycloak-core/4.7.0.Final/keycloak-core-4.7.0.Final.jar</p> <p> Dependency Hierarchy: - keycloak-spring-boot-starter-4.7.0.Final.jar (Root Library) - keycloak-spring-boot-2-adapter-4.7.0.Final.jar - :x: **keycloak-core-4.7.0.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in keycloak before 6.0.2. The X.509 authenticator supports the verification of client certificates through the CRL, where the CRL list can be obtained from the URL provided in the certificate itself (CDP) or through the separately configured path. The CRL are often available over the network through unsecured protocols ('http' or 'ldap') and hence the caller should verify the signature and possibly the certification path. Keycloak currently doesn't validate signatures on CRL, which can result in a possibility of various attacks like man-in-the-middle. <p>Publish Date: 2019-06-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3875>CVE-2019-3875</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875</a></p> <p>Release Date: 2019-06-12</p> <p>Fix Resolution: 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-3875 (Medium) detected in keycloak-core-4.7.0.Final.jar - ## CVE-2019-3875 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>keycloak-core-4.7.0.Final.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://keycloak.org/keycloak-core">http://keycloak.org/keycloak-core</a></p> <p>Path to dependency file: /s5/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/keycloak/keycloak-core/4.7.0.Final/keycloak-core-4.7.0.Final.jar</p> <p> Dependency Hierarchy: - keycloak-spring-boot-starter-4.7.0.Final.jar (Root Library) - keycloak-spring-boot-2-adapter-4.7.0.Final.jar - :x: **keycloak-core-4.7.0.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in keycloak before 6.0.2. The X.509 authenticator supports the verification of client certificates through the CRL, where the CRL list can be obtained from the URL provided in the certificate itself (CDP) or through the separately configured path. The CRL are often available over the network through unsecured protocols ('http' or 'ldap') and hence the caller should verify the signature and possibly the certification path. Keycloak currently doesn't validate signatures on CRL, which can result in a possibility of various attacks like man-in-the-middle. <p>Publish Date: 2019-06-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3875>CVE-2019-3875</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3875</a></p> <p>Release Date: 2019-06-12</p> <p>Fix Resolution: 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in keycloak core final jar cve medium severity vulnerability vulnerable library keycloak core final jar library home page a href path to dependency file pom xml path to vulnerable library root repository org keycloak keycloak core final keycloak core final jar dependency hierarchy keycloak spring boot starter final jar root library keycloak spring boot adapter final jar x keycloak core final jar vulnerable library vulnerability details a vulnerability was found in keycloak before the x authenticator supports the verification of client certificates through the crl where the crl list can be obtained from the url provided in the certificate itself cdp or through the separately configured path the crl are often available over the network through unsecured protocols http or ldap and hence the caller should verify the signature and possibly the certification path keycloak currently doesn t validate signatures on crl which can result in a possibility of various attacks like man in the middle publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
775,983
27,242,614,973
IssuesEvent
2023-02-21 21:57:21
o3de/o3de.org
https://api.github.com/repos/o3de/o3de.org
closed
Error code index
kind/reference kind/feature priority/critical
It would be useful to have several different error code indexes - things like editor errors, command-line tool errors, core engine or atom errors - and provide an index that maps them to human-readable information so that it's possible to do a quick lookup online. This also acts as a **great** search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something. This is also a very good candidate for automation, if there's a way to extract error codes reliably from the O3DE source (such as if they're contained in, say, a filemap somewhere.)
1.0
Error code index - It would be useful to have several different error code indexes - things like editor errors, command-line tool errors, core engine or atom errors - and provide an index that maps them to human-readable information so that it's possible to do a quick lookup online. This also acts as a **great** search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something. This is also a very good candidate for automation, if there's a way to extract error codes reliably from the O3DE source (such as if they're contained in, say, a filemap somewhere.)
non_defect
error code index it would be useful to have several different error code indexes things like editor errors command line tool errors core engine or atom errors and provide an index that maps them to human readable information so that it s possible to do a quick lookup online this also acts as a great search sink for getting people encountering common error codes into a troubleshooting article instead of having to crawl stackoverflow or something this is also a very good candidate for automation if there s a way to extract error codes reliably from the source such as if they re contained in say a filemap somewhere
0
80,766
30,523,453,788
IssuesEvent
2023-07-19 09:36:47
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectCheckboxMenu: unable to partial search for an option without following value order
:lady_beetle: defect :bangbang: needs-triage
### Describe the bug This also doesn't work for the showcase; When using selectCheckboxMenu with options like "London City", "Leeds City", "Manchester City" "Manchester United" . when using the search feature of selectcheckboxmenu, I can't just type City and filter the options that way, I currently have to start at the beginning of the value/string like "Manchester" which would show Manchester city and Manchester United. ### Reproducer can be checked using the primefaces showcase, if you wanted to look for Miami but typed iami you dont get any result ### Expected behavior the search should filter the list based on any matched letters within the values ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation All ### JSF version 2.1 ### Java version 19 ### Browser(s) _No response_
1.0
SelectCheckboxMenu: unable to partial search for an option without following value order - ### Describe the bug This also doesn't work for the showcase; When using selectCheckboxMenu with options like "London City", "Leeds City", "Manchester City" "Manchester United" . when using the search feature of selectcheckboxmenu, I can't just type City and filter the options that way, I currently have to start at the beginning of the value/string like "Manchester" which would show Manchester city and Manchester United. ### Reproducer can be checked using the primefaces showcase, if you wanted to look for Miami but typed iami you dont get any result ### Expected behavior the search should filter the list based on any matched letters within the values ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation All ### JSF version 2.1 ### Java version 19 ### Browser(s) _No response_
defect
selectcheckboxmenu unable to partial search for an option without following value order describe the bug this also doesn t work for the showcase when using selectcheckboxmenu with options like london city leeds city manchester city manchester united when using the search feature of selectcheckboxmenu i can t just type city and filter the options that way i currently have to start at the beginning of the value string like manchester which would show manchester city and manchester united reproducer can be checked using the primefaces showcase if you wanted to look for miami but typed iami you dont get any result expected behavior the search should filter the list based on any matched letters within the values primefaces edition community primefaces version theme no response jsf implementation all jsf version java version browser s no response
1
38,218
8,701,643,514
IssuesEvent
2018-12-05 12:12:00
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Reflection on MethodHandles$Lookup fails in JDK 9
C: Functionality P: Medium T: Defect
This is the same issue as encountered in jOOR: https://github.com/jOOQ/jOOR/issues/49. The `DefaultRecordMapper.ProxyMapper` cannot call default methods on such proxies, starting from JDK 9
1.0
Reflection on MethodHandles$Lookup fails in JDK 9 - This is the same issue as encountered in jOOR: https://github.com/jOOQ/jOOR/issues/49. The `DefaultRecordMapper.ProxyMapper` cannot call default methods on such proxies, starting from JDK 9
defect
reflection on methodhandles lookup fails in jdk this is the same issue as encountered in joor the defaultrecordmapper proxymapper cannot call default methods on such proxies starting from jdk
1
315,566
23,587,115,878
IssuesEvent
2022-08-23 12:33:33
r-lib/cli
https://api.github.com/repos/r-lib/cli
closed
"Building a Semantic CLI" vignette has misplaced asciicast svgs.
documentation
For some reason, the code present at the `asciicast` block in [line 207](https://github.com/r-lib/cli/blob/HEAD/vignettes/semantic-cli.Rmd#L207) isn't rendered, and what's rendered instead is the next `asciicast` block. All remaning `asciicast` blocks after this point render before they were supposed to, resulting in a confusing vignette, like in the following screenshot: ![image](https://user-images.githubusercontent.com/95953/158028744-395797d6-9abe-49ed-8371-7288ed76e276.png) I have tried this in multiple browsers, it does not seem to be a local issue.
1.0
"Building a Semantic CLI" vignette has misplaced asciicast svgs. - For some reason, the code present at the `asciicast` block in [line 207](https://github.com/r-lib/cli/blob/HEAD/vignettes/semantic-cli.Rmd#L207) isn't rendered, and what's rendered instead is the next `asciicast` block. All remaning `asciicast` blocks after this point render before they were supposed to, resulting in a confusing vignette, like in the following screenshot: ![image](https://user-images.githubusercontent.com/95953/158028744-395797d6-9abe-49ed-8371-7288ed76e276.png) I have tried this in multiple browsers, it does not seem to be a local issue.
non_defect
building a semantic cli vignette has misplaced asciicast svgs for some reason the code present at the asciicast block in isn t rendered and what s rendered instead is the next asciicast block all remaning asciicast blocks after this point render before they were supposed to resulting in a confusing vignette like in the following screenshot i have tried this in multiple browsers it does not seem to be a local issue
0
40,040
9,810,129,174
IssuesEvent
2019-06-12 19:42:49
IBM/CAST
https://api.github.com/repos/IBM/CAST
closed
standalone_ib_and_switch_collection error after upgraded to 1.6.0-2583
Comp: CSM.api Comp: CSM.db PhaseFound: IST Sev: 2 Type: Defect
**Describe the bug** Running the command `standalone_ib_and_switch_collection` (1.6.0-2583) produced csmapi and database errors at the end. **To Reproduce** Steps to reproduce the behavior: 1. Go to `/opt/ibm/csm/sbin/` 2. Run `./standalone_ib_and_switch_collection` 3. See errors: ``` [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" (null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" (null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" ``` **Expected behavior** The command should not display error messages. **Screenshots** ``` [root@f4n05 sbin]# ./standalone_ib_and_switch_collection UFM reported 190 IB records. This report from UFM can be found in 'ufm_ib_cable_output_file.json' located at '/var/log/ibm/csm/inv' WARNING: 9 IB cables were discovered, but are missing serial numbers and have been removed from CSM inventory collection data. These records copied into 'bad_ib_cable_records.txt' located at '/var/log/ibm/csm/inv' UFM reported 9 switch records. This report from UFM can be found in 'ufm_switch_output_file.json' located at '/var/log/ibm/csm/inv' WARNING: 2 Switches found with 'N/A' serial numbers and have been removed from CSM inventory collection data. These records copied into 'bad_switch_records.txt' located at '/var/log/ibm/csm/inv' --- # ib inventory successful ... # total ib inventory collected: 181 # new ib records inserted into database: 0 # old ib records updated in database: 181 # old ib records removed from the database: 0 [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" (null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" (null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" ``` **Environment (please complete the following information):** - Machine [CSM master: f4n05] - Version [CSM 1.6.0-2583] **Additional context** The command worked fine with 1.6.0-2575 before upgrading to 1.6.0-2583. **Issue Source:** IST CSM upgrade on alt. cluster.
1.0
standalone_ib_and_switch_collection error after upgraded to 1.6.0-2583 - **Describe the bug** Running the command `standalone_ib_and_switch_collection` (1.6.0-2583) produced csmapi and database errors at the end. **To Reproduce** Steps to reproduce the behavior: 1. Go to `/opt/ibm/csm/sbin/` 2. Run `./standalone_ib_and_switch_collection` 3. See errors: ``` [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" (null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" (null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" ``` **Expected behavior** The command should not display error messages. **Screenshots** ``` [root@f4n05 sbin]# ./standalone_ib_and_switch_collection UFM reported 190 IB records. This report from UFM can be found in 'ufm_ib_cable_output_file.json' located at '/var/log/ibm/csm/inv' WARNING: 9 IB cables were discovered, but are missing serial numbers and have been removed from CSM inventory collection data. These records copied into 'bad_ib_cable_records.txt' located at '/var/log/ibm/csm/inv' UFM reported 9 switch records. This report from UFM can be found in 'ufm_switch_output_file.json' located at '/var/log/ibm/csm/inv' WARNING: 2 Switches found with 'N/A' serial numbers and have been removed from CSM inventory collection data. These records copied into 'bad_switch_records.txt' located at '/var/log/ibm/csm/inv' --- # ib inventory successful ... # total ib inventory collected: 181 # new ib records inserted into database: 0 # old ib records updated in database: 181 # old ib records removed from the database: 0 [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" (null) FAILED: errcode: 25 errmsg: csm_switch_inventory_collection[431235260]; Database Error Message: update or delete on table "csm_switch" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" on table "csm_switch_inventory" [csmapi][warning] /home/ppsbld/workspace/PUBLIC_CAST_V1.0.0_ppc64LE_RH7.5_ProdBuild/csmi/src/common/src/csmi_common_utils.c-147: the Error Flag Set [csmapi][error] csmi_sendrecv_cmd failed: 25 - csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" (null) FAILED: errcode: 25 errmsg: csm_switch_children_inventory_collection[431235261]; Database Error Message: insert or update on table "csm_switch_inventory" violates foreign key constraint "csm_switch_inventory_host_system_guid_fkey" ``` **Environment (please complete the following information):** - Machine [CSM master: f4n05] - Version [CSM 1.6.0-2583] **Additional context** The command worked fine with 1.6.0-2575 before upgrading to 1.6.0-2583. **Issue Source:** IST CSM upgrade on alt. cluster.
defect
standalone ib and switch collection error after upgraded to describe the bug running the command standalone ib and switch collection produced csmapi and database errors at the end to reproduce steps to reproduce the behavior go to opt ibm csm sbin run standalone ib and switch collection see errors home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory null failed errcode errmsg csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey null failed errcode errmsg csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey expected behavior the command should not display error messages screenshots standalone ib and switch collection ufm reported ib records this report from ufm can be found in ufm ib cable output file json located at var log ibm csm inv warning ib cables were discovered but are missing serial numbers and have been removed from csm inventory collection data these records copied into bad ib cable records txt located at var log ibm csm inv ufm reported switch records this report from ufm can be found in ufm switch output file json located at var log ibm csm inv warning switches found with n a serial numbers and have been removed from csm inventory collection data these records copied into bad switch records txt located at var log ibm csm inv ib inventory successful total ib inventory collected new ib records inserted into database old ib records updated in database old ib records removed from the database home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory null failed errcode errmsg csm switch inventory collection database error message update or delete on table csm switch violates foreign key constraint csm switch inventory host system guid fkey on table csm switch inventory home ppsbld workspace public cast prodbuild csmi src common src csmi common utils c the error flag set csmi sendrecv cmd failed csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey null failed errcode errmsg csm switch children inventory collection database error message insert or update on table csm switch inventory violates foreign key constraint csm switch inventory host system guid fkey environment please complete the following information machine version additional context the command worked fine with before upgrading to issue source ist csm upgrade on alt cluster
1
665,001
22,295,559,721
IssuesEvent
2022-06-13 00:47:15
idom-team/idom
https://api.github.com/repos/idom-team/idom
opened
Fix Flaky Tests
priority: 2 (moderate)
### Current Situation There are a number of tests in the suite that are very flaky (usually requiring 2-3 runs to pass). For new contributors this is likely to be confusing since they'll be unsure if those failures are related to their changes. ### Proposed Actions It's likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order (or at least that's been my experience in the past). Unfortunately the problems are hard to reproduce since running the local tests rarely, if ever, fails at the rate seen in CI/CD.
1.0
Fix Flaky Tests - ### Current Situation There are a number of tests in the suite that are very flaky (usually requiring 2-3 runs to pass). For new contributors this is likely to be confusing since they'll be unsure if those failures are related to their changes. ### Proposed Actions It's likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order (or at least that's been my experience in the past). Unfortunately the problems are hard to reproduce since running the local tests rarely, if ever, fails at the rate seen in CI/CD.
non_defect
fix flaky tests current situation there are a number of tests in the suite that are very flaky usually requiring runs to pass for new contributors this is likely to be confusing since they ll be unsure if those failures are related to their changes proposed actions it s likely that these flaky tests are missing some sort of locking mechanism to ensure that actions occur in the correct order or at least that s been my experience in the past unfortunately the problems are hard to reproduce since running the local tests rarely if ever fails at the rate seen in ci cd
0
63,843
18,015,927,521
IssuesEvent
2021-09-16 13:55:13
hazelcast/hazelcast-go-client
https://api.github.com/repos/hazelcast/hazelcast-go-client
opened
User callbacks for events should not block internals of the client
Type: Defect Source: Internal
Currently since events are subscribed as sync if a user blocks an event, all the client events are blocked.
1.0
User callbacks for events should not block internals of the client - Currently since events are subscribed as sync if a user blocks an event, all the client events are blocked.
defect
user callbacks for events should not block internals of the client currently since events are subscribed as sync if a user blocks an event all the client events are blocked
1
101,022
8,767,448,643
IssuesEvent
2018-12-17 19:47:30
nuxdie/frontendweekend
https://api.github.com/repos/nuxdie/frontendweekend
closed
add dev chat
monitoring testing tracking
something like slack for dev notifications and web hooks (maybe send to telegram)
1.0
add dev chat - something like slack for dev notifications and web hooks (maybe send to telegram)
non_defect
add dev chat something like slack for dev notifications and web hooks maybe send to telegram
0
62,420
17,023,920,047
IssuesEvent
2021-07-03 04:33:38
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
HTTPS support for osm replication updates
Component: admin Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 7.54am, Tuesday, 7th April 2015]** https://planet.openstreetmap.org/replication/day/000/000/ Since the OSM updates diff support only http, I cannot update the OSM database in my production environment due to security issue. It is vulnerable to MITM(Man-In-The-Middle) attack. When I tried https url, it redirects to http url. Is there any specific reason for not having https support ? It would be great, if OSM gives support https. And also OSM may add md5check sum details in the xx_state.txt files. I am ready to take this task and contribute. Can you guide me, where to start ?
1.0
HTTPS support for osm replication updates - **[Submitted to the original trac issue database at 7.54am, Tuesday, 7th April 2015]** https://planet.openstreetmap.org/replication/day/000/000/ Since the OSM updates diff support only http, I cannot update the OSM database in my production environment due to security issue. It is vulnerable to MITM(Man-In-The-Middle) attack. When I tried https url, it redirects to http url. Is there any specific reason for not having https support ? It would be great, if OSM gives support https. And also OSM may add md5check sum details in the xx_state.txt files. I am ready to take this task and contribute. Can you guide me, where to start ?
defect
https support for osm replication updates since the osm updates diff support only http i cannot update the osm database in my production environment due to security issue it is vulnerable to mitm man in the middle attack when i tried https url it redirects to http url is there any specific reason for not having https support it would be great if osm gives support https and also osm may add sum details in the xx state txt files i am ready to take this task and contribute can you guide me where to start
1
41,122
10,313,853,205
IssuesEvent
2019-08-30 00:45:54
techo/voluntariado-eventual
https://api.github.com/repos/techo/voluntariado-eventual
closed
Búsqueda por varios términos en los inscriptos a una actividad
Defecto
**El pedido está relacionado con un problema, describí** Cómo Coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones, si se pone más de un término, no trae resultados. **Describí la solución que te gustaría** Qué traiga los resultados que coincidan para con el nombre, apellido, email, dni de esos términos **Describí alternativas que hayas considerado** **Contexto adicional** Lo mismo pasaba sobre las tablas de buscar actividad o persona. Se puede copiar de ahí prácticamente. ![image](https://user-images.githubusercontent.com/94343/63716861-5bb97b80-c81d-11e9-8a81-ff6cb860b31c.png)
1.0
Búsqueda por varios términos en los inscriptos a una actividad - **El pedido está relacionado con un problema, describí** Cómo Coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones, si se pone más de un término, no trae resultados. **Describí la solución que te gustaría** Qué traiga los resultados que coincidan para con el nombre, apellido, email, dni de esos términos **Describí alternativas que hayas considerado** **Contexto adicional** Lo mismo pasaba sobre las tablas de buscar actividad o persona. Se puede copiar de ahí prácticamente. ![image](https://user-images.githubusercontent.com/94343/63716861-5bb97b80-c81d-11e9-8a81-ff6cb860b31c.png)
defect
búsqueda por varios términos en los inscriptos a una actividad el pedido está relacionado con un problema describí cómo coordinador al usar el campo de búsqueda rápida sobre la tabla de inscripciones si se pone más de un término no trae resultados describí la solución que te gustaría qué traiga los resultados que coincidan para con el nombre apellido email dni de esos términos describí alternativas que hayas considerado contexto adicional lo mismo pasaba sobre las tablas de buscar actividad o persona se puede copiar de ahí prácticamente
1
65,365
19,431,758,862
IssuesEvent
2021-12-21 12:50:03
catmaid/CATMAID
https://api.github.com/repos/catmaid/CATMAID
closed
Use of localStorage and cookies
type: defect status: done
Currently some semi-persistent client settings (e.g., selected stack mirror) use either/both of localStorage and cookies. Two issues: 1. Cookies are not appropriate for this purpose, as we've decided this is per-browser-client-specific data (hence not being in a `CATMAID.Settings` persistent store). Additionally the duplication/interference of cookies and localStorage is a needless complication now that all supported browsers have localStorage. 2. Our current use of localStorage does not use instance-specific keys, which can cause collisions on servers with multiple instances.
1.0
Use of localStorage and cookies - Currently some semi-persistent client settings (e.g., selected stack mirror) use either/both of localStorage and cookies. Two issues: 1. Cookies are not appropriate for this purpose, as we've decided this is per-browser-client-specific data (hence not being in a `CATMAID.Settings` persistent store). Additionally the duplication/interference of cookies and localStorage is a needless complication now that all supported browsers have localStorage. 2. Our current use of localStorage does not use instance-specific keys, which can cause collisions on servers with multiple instances.
defect
use of localstorage and cookies currently some semi persistent client settings e g selected stack mirror use either both of localstorage and cookies two issues cookies are not appropriate for this purpose as we ve decided this is per browser client specific data hence not being in a catmaid settings persistent store additionally the duplication interference of cookies and localstorage is a needless complication now that all supported browsers have localstorage our current use of localstorage does not use instance specific keys which can cause collisions on servers with multiple instances
1
58,267
16,457,461,514
IssuesEvent
2021-05-21 14:21:57
Questie/Questie
https://api.github.com/repos/Questie/Questie
opened
Game stutters everytime it updates a quest
Type - Defect
<!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- --> My game stutters everytime I loot a quest item, kill a quest objective etc. Issue disappears if I disable questie. ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version 6.3.11
1.0
Game stutters everytime it updates a quest - <!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- --> My game stutters everytime I loot a quest item, kill a quest objective etc. Issue disappears if I disable questie. ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version 6.3.11
defect
game stutters everytime it updates a quest read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description my game stutters everytime i loot a quest item kill a quest objective etc issue disappears if i disable questie screenshots questie version
1
65,393
19,479,717,572
IssuesEvent
2021-12-25 01:45:45
unascribed/Fabrication
https://api.github.com/repos/unascribed/Fabrication
closed
Game crashes when i click create new world or load a world when i have extra.encroaching_emeralds enabled
k: Defect n: Fabric s: New
https://gist.github.com/gitunown/bce59b7db3a389f8881caf1afd3c4bef
1.0
Game crashes when i click create new world or load a world when i have extra.encroaching_emeralds enabled - https://gist.github.com/gitunown/bce59b7db3a389f8881caf1afd3c4bef
defect
game crashes when i click create new world or load a world when i have extra encroaching emeralds enabled
1
355,973
10,586,911,304
IssuesEvent
2019-10-08 20:48:09
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Annotation Problem *.pyi
high priority module: typing triaged
## 🚀 Feature It is difficult to control the validity of function arguments when they are in different files. Very often the information in *.pyi does not match *.py. https://github.com/pytorch/pytorch/blob/fcd13549f923342a7474993928d86181b75734d3/torch/utils/data/dataloader.pyi#L35 I suggest using Python 2 Annotations: ```python def optional_unwrap (self, x, y):    # type: (Optional [int], Optional [int]) -> int ``` You can expand this format to indicate in which file to place annotations. After that, all *.pyi can be completely assembled in one script. Information will always be up to date. cc @ezyang @gchanan @zou3519
1.0
Annotation Problem *.pyi - ## 🚀 Feature It is difficult to control the validity of function arguments when they are in different files. Very often the information in *.pyi does not match *.py. https://github.com/pytorch/pytorch/blob/fcd13549f923342a7474993928d86181b75734d3/torch/utils/data/dataloader.pyi#L35 I suggest using Python 2 Annotations: ```python def optional_unwrap (self, x, y):    # type: (Optional [int], Optional [int]) -> int ``` You can expand this format to indicate in which file to place annotations. After that, all *.pyi can be completely assembled in one script. Information will always be up to date. cc @ezyang @gchanan @zou3519
non_defect
annotation problem pyi 🚀 feature it is difficult to control the validity of function arguments when they are in different files very often the information in pyi does not match py i suggest using python annotations python def optional unwrap self x y    type optional optional int you can expand this format to indicate in which file to place annotations after that all pyi can be completely assembled in one script information will always be up to date cc ezyang gchanan
0
58,443
3,089,416,162
IssuesEvent
2015-08-25 21:23:17
phetsims/joist
https://api.github.com/repos/phetsims/joist
opened
Add additional Google Analytics tracker
enhancement high-priority Summer 2015 redeploy
See https://github.com/phetsims/yotta/issues/12 for details. Will include sending data to another tracker conditionally.
1.0
Add additional Google Analytics tracker - See https://github.com/phetsims/yotta/issues/12 for details. Will include sending data to another tracker conditionally.
non_defect
add additional google analytics tracker see for details will include sending data to another tracker conditionally
0
62,189
17,023,869,201
IssuesEvent
2021-07-03 04:16:40
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
[roads] end of oneway unclear
Component: mapnik Priority: minor Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 8.33pm, Monday, 22nd July 2013]** An [http://www.openstreetmap.org/?lat=50.620504&lon=5.660834&zoom=18&layers=M inhabitant of this place] draws my attention (aren't some people nice?) on the oneway street Jean Leclercq. He says that the map makes it unclear that the oneway restriction stops half way, at the crossing with street Jean Jaurs. He says that there should be an arrow there to show that end.[[BR]] I'm taking the idea one step further. An additional arrow is a hint and not a certainty. Could it be easily possible, when a oneway stops in the middle of a street, to draw a clear double headed arrow at the beginning of the unrestricted segment?[[BR]] This is similar to Ticket #4338, but different. They should be considered in parallel. Thanks.
1.0
[roads] end of oneway unclear - **[Submitted to the original trac issue database at 8.33pm, Monday, 22nd July 2013]** An [http://www.openstreetmap.org/?lat=50.620504&lon=5.660834&zoom=18&layers=M inhabitant of this place] draws my attention (aren't some people nice?) on the oneway street Jean Leclercq. He says that the map makes it unclear that the oneway restriction stops half way, at the crossing with street Jean Jaurs. He says that there should be an arrow there to show that end.[[BR]] I'm taking the idea one step further. An additional arrow is a hint and not a certainty. Could it be easily possible, when a oneway stops in the middle of a street, to draw a clear double headed arrow at the beginning of the unrestricted segment?[[BR]] This is similar to Ticket #4338, but different. They should be considered in parallel. Thanks.
defect
end of oneway unclear an draws my attention aren t some people nice on the oneway street jean leclercq he says that the map makes it unclear that the oneway restriction stops half way at the crossing with street jean jaurs he says that there should be an arrow there to show that end i m taking the idea one step further an additional arrow is a hint and not a certainty could it be easily possible when a oneway stops in the middle of a street to draw a clear double headed arrow at the beginning of the unrestricted segment this is similar to ticket but different they should be considered in parallel thanks
1
68,120
21,483,720,850
IssuesEvent
2022-04-26 20:30:09
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
Enabling threads hides the voice call button
T-Defect X-Needs-Design Z-UI UX A-VoIP S-Minor O-Occasional
### Steps to reproduce 1. Voice button is now hidden behind the hamburger icon 2. Except it’s not labelled “voice call” it’s “call”, which makes it even harder to find 3. Instead, perhaps we need a single call button which then gives you voice or video as a choice ### Outcome #### What did you expect? Intuitive calling #### What happened instead? I kept getting video calls because folks can no longer find the voice call button. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Enabling threads hides the voice call button - ### Steps to reproduce 1. Voice button is now hidden behind the hamburger icon 2. Except it’s not labelled “voice call” it’s “call”, which makes it even harder to find 3. Instead, perhaps we need a single call button which then gives you voice or video as a choice ### Outcome #### What did you expect? Intuitive calling #### What happened instead? I kept getting video calls because folks can no longer find the voice call button. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
enabling threads hides the voice call button steps to reproduce voice button is now hidden behind the hamburger icon except it’s not labelled “voice call” it’s “call” which makes it even harder to find instead perhaps we need a single call button which then gives you voice or video as a choice outcome what did you expect intuitive calling what happened instead i kept getting video calls because folks can no longer find the voice call button your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no
1
18,502
2,615,172,357
IssuesEvent
2015-03-01 06:54:25
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
opened
Streaming audio on the web
auto-migrated Milestone-X New Priority-P2 Tutorial Type-Feature
``` - media source api - binary web sockets ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 31 Jan 2012 at 8:26
1.0
Streaming audio on the web - ``` - media source api - binary web sockets ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 31 Jan 2012 at 8:26
non_defect
streaming audio on the web media source api binary web sockets original issue reported on code google com by ericbide com on jan at
0
36,822
5,088,285,280
IssuesEvent
2016-12-31 17:41:09
NMGRL/pychron
https://api.github.com/repos/NMGRL/pychron
closed
Furnace loader OK
TestingRequired
Felix and Jan running well. Furnace sample dumped perfectly with electromagnets. Perhaps Matt did have the valves shut when he tried it. All controls working well except "Dump" button, which is seemingly not doing anything. All other button elements fine. Should dump button actuate valves and shutters, then dump? Or does it actuate the magnets and mark the position as dumped? Perhaps the button should be labelled "Load" no dump, and the "Magnets" button should be labelled "Dump". I will have a look at whether all buckets will now dump. Could funnel position be encoded into the status word? These would allow easy animation of the funnel graphic.
1.0
Furnace loader OK - Felix and Jan running well. Furnace sample dumped perfectly with electromagnets. Perhaps Matt did have the valves shut when he tried it. All controls working well except "Dump" button, which is seemingly not doing anything. All other button elements fine. Should dump button actuate valves and shutters, then dump? Or does it actuate the magnets and mark the position as dumped? Perhaps the button should be labelled "Load" no dump, and the "Magnets" button should be labelled "Dump". I will have a look at whether all buckets will now dump. Could funnel position be encoded into the status word? These would allow easy animation of the funnel graphic.
non_defect
furnace loader ok felix and jan running well furnace sample dumped perfectly with electromagnets perhaps matt did have the valves shut when he tried it all controls working well except dump button which is seemingly not doing anything all other button elements fine should dump button actuate valves and shutters then dump or does it actuate the magnets and mark the position as dumped perhaps the button should be labelled load no dump and the magnets button should be labelled dump i will have a look at whether all buckets will now dump could funnel position be encoded into the status word these would allow easy animation of the funnel graphic
0
83,181
10,329,866,973
IssuesEvent
2019-09-02 13:18:14
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
opened
Trading: Updating resized components + color changes
Design Roadmap: Trading
After changes made to the Trading page, resizing and refining components. We need to apply that to components on: - Account summary - Portfolio - Reporting We're also updating the UI color scheme Need to be applied across desktop/mobile etc
1.0
Trading: Updating resized components + color changes - After changes made to the Trading page, resizing and refining components. We need to apply that to components on: - Account summary - Portfolio - Reporting We're also updating the UI color scheme Need to be applied across desktop/mobile etc
non_defect
trading updating resized components color changes after changes made to the trading page resizing and refining components we need to apply that to components on account summary portfolio reporting we re also updating the ui color scheme need to be applied across desktop mobile etc
0
13,693
3,176,000,096
IssuesEvent
2015-09-24 05:28:54
sulu-io/sulu
https://api.github.com/repos/sulu-io/sulu
closed
Admin Frontend Issues
Design
The following issues have been observed Contact/Account in ContactExtensionBundle * [x] Wrong space distribution when sidebar is activated and browser width is between 900px and 1090px datagrid: <img width="1082" alt="bildschirmfoto 2015-09-17 um 11 32 40" src="https://cloud.githubusercontent.com/assets/5152633/9929873/dec9c294-5d2f-11e5-819c-aaffaa157757.png"> * [x] Styling of content, when sidebar overlaps content <img width="941" alt="bildschirmfoto 2015-09-17 um 10 55 14" src="https://cloud.githubusercontent.com/assets/5152633/9929068/c1ae8dde-5d2a-11e5-9004-5865419d0c26.png"> * [x] nested table view: error when resizing browser ![image](https://cloud.githubusercontent.com/assets/5152633/9901465/393d3a62-5c68-11e5-9e84-c6cdaef1b111.png) * [x] Resolution 1440px does not comply with sidebar. In fact, the navigation should be collapsed: <img width="1402" alt="bildschirmfoto 2015-09-17 um 11 26 22" src="https://cloud.githubusercontent.com/assets/5152633/9929812/8038b226-5d2f-11e5-8563-9be0e1b1b70c.png">
1.0
Admin Frontend Issues - The following issues have been observed Contact/Account in ContactExtensionBundle * [x] Wrong space distribution when sidebar is activated and browser width is between 900px and 1090px datagrid: <img width="1082" alt="bildschirmfoto 2015-09-17 um 11 32 40" src="https://cloud.githubusercontent.com/assets/5152633/9929873/dec9c294-5d2f-11e5-819c-aaffaa157757.png"> * [x] Styling of content, when sidebar overlaps content <img width="941" alt="bildschirmfoto 2015-09-17 um 10 55 14" src="https://cloud.githubusercontent.com/assets/5152633/9929068/c1ae8dde-5d2a-11e5-9004-5865419d0c26.png"> * [x] nested table view: error when resizing browser ![image](https://cloud.githubusercontent.com/assets/5152633/9901465/393d3a62-5c68-11e5-9e84-c6cdaef1b111.png) * [x] Resolution 1440px does not comply with sidebar. In fact, the navigation should be collapsed: <img width="1402" alt="bildschirmfoto 2015-09-17 um 11 26 22" src="https://cloud.githubusercontent.com/assets/5152633/9929812/8038b226-5d2f-11e5-8563-9be0e1b1b70c.png">
non_defect
admin frontend issues the following issues have been observed contact account in contactextensionbundle wrong space distribution when sidebar is activated and browser width is between and datagrid img width alt bildschirmfoto um src styling of content when sidebar overlaps content img width alt bildschirmfoto um src nested table view error when resizing browser resolution does not comply with sidebar in fact the navigation should be collapsed img width alt bildschirmfoto um src
0
212,390
23,887,922,153
IssuesEvent
2022-09-08 09:09:40
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Security Solution] Any number after decimal is not allowed except zero for the timer
bug impact:medium Team: SecuritySolution Team:Onboarding and Lifecycle Mgt v8.5.0
**Description:** Any number after decimal is not allowed except zero for the timer **Build Details:** ``` VERSION: 8.4.0 BC1 BUILD: 54999 COMMIT: 58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d ARTIFACTS: https://staging.elastic.co/8.4.0-91109cf0/summary-8.4.0.html ``` **Browser Details:** All **Preconditions:** 1. Kibana user should be logged in **Steps to Reproduce:** 1. Navigate to the Endpoints tab under the Manage section of the Security from the left-hand side navigation bar 2. Click on the three dots for an endpoint 3. Click on the launch responder option 4. Click on the Actions log button 5. Click on the timer button 6. Try adding a value after decimal other than zero. - Example: 5.6 or 8.3 7. Observe when we press any number to add, the decimal is removed and no value is added to the timer **Actual Result:** Any number after decimal is not allowed except zero for the timer **Screen Recording:** https://user-images.githubusercontent.com/60252716/189079409-eee117d4-aec2-4b0e-b5d4-2fb88a5a39b2.mp4 **Logs:** N/A
True
[Security Solution] Any number after decimal is not allowed except zero for the timer - **Description:** Any number after decimal is not allowed except zero for the timer **Build Details:** ``` VERSION: 8.4.0 BC1 BUILD: 54999 COMMIT: 58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d ARTIFACTS: https://staging.elastic.co/8.4.0-91109cf0/summary-8.4.0.html ``` **Browser Details:** All **Preconditions:** 1. Kibana user should be logged in **Steps to Reproduce:** 1. Navigate to the Endpoints tab under the Manage section of the Security from the left-hand side navigation bar 2. Click on the three dots for an endpoint 3. Click on the launch responder option 4. Click on the Actions log button 5. Click on the timer button 6. Try adding a value after decimal other than zero. - Example: 5.6 or 8.3 7. Observe when we press any number to add, the decimal is removed and no value is added to the timer **Actual Result:** Any number after decimal is not allowed except zero for the timer **Screen Recording:** https://user-images.githubusercontent.com/60252716/189079409-eee117d4-aec2-4b0e-b5d4-2fb88a5a39b2.mp4 **Logs:** N/A
non_defect
any number after decimal is not allowed except zero for the timer description any number after decimal is not allowed except zero for the timer build details version build commit artifacts browser details all preconditions kibana user should be logged in steps to reproduce navigate to the endpoints tab under the manage section of the security from the left hand side navigation bar click on the three dots for an endpoint click on the launch responder option click on the actions log button click on the timer button try adding a value after decimal other than zero example or observe when we press any number to add the decimal is removed and no value is added to the timer actual result any number after decimal is not allowed except zero for the timer screen recording logs n a
0
70,346
9,403,723,450
IssuesEvent
2019-04-09 02:42:51
pyouroboros/ouroboros
https://api.github.com/repos/pyouroboros/ouroboros
reopened
Document swarm configuration
documentation
The documentation doesn't mention anything about Swarm, but there are configurations relevant to it: I think `SWARM=true` needs to be passed. It seems that then `LABEL_ENABLE` defaults to `true`, so `MONITOR` is ignored. I haven't tried setting it to false.
1.0
Document swarm configuration - The documentation doesn't mention anything about Swarm, but there are configurations relevant to it: I think `SWARM=true` needs to be passed. It seems that then `LABEL_ENABLE` defaults to `true`, so `MONITOR` is ignored. I haven't tried setting it to false.
non_defect
document swarm configuration the documentation doesn t mention anything about swarm but there are configurations relevant to it i think swarm true needs to be passed it seems that then label enable defaults to true so monitor is ignored i haven t tried setting it to false
0
16,834
2,948,322,953
IssuesEvent
2015-07-06 01:31:15
Winetricks/winetricks
https://api.github.com/repos/Winetricks/winetricks
closed
MinGW should install mingw32-make
auto-migrated Priority-Medium Type-Defect
``` MinGW is useless without make. ``` Original issue reported on code.google.com by `disin...@gmail.com` on 12 Feb 2012 at 6:28
1.0
MinGW should install mingw32-make - ``` MinGW is useless without make. ``` Original issue reported on code.google.com by `disin...@gmail.com` on 12 Feb 2012 at 6:28
defect
mingw should install make mingw is useless without make original issue reported on code google com by disin gmail com on feb at
1
267,592
28,509,121,450
IssuesEvent
2023-04-19 01:37:04
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
closed
CVE-2011-3191 (High) detected in linux-yocto-4.12v3.1.10 - autoclosed
Mend: dependency security vulnerability
## CVE-2011-3191 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary> <p> <p>Linux 4.12 Embedded Kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/cifs/cifssmb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Integer signedness error in the CIFSFindNext function in fs/cifs/cifssmb.c in the Linux kernel before 3.1 allows remote CIFS servers to cause a denial of service (memory corruption) or possibly have unspecified other impact via a large length value in a response to a read request for a directory. <p>Publish Date: 2012-05-24 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-3191>CVE-2011-3191</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-3191">https://nvd.nist.gov/vuln/detail/CVE-2011-3191</a></p> <p>Release Date: 2012-05-24</p> <p>Fix Resolution: 3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2011-3191 (High) detected in linux-yocto-4.12v3.1.10 - autoclosed - ## CVE-2011-3191 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary> <p> <p>Linux 4.12 Embedded Kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/cifs/cifssmb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Integer signedness error in the CIFSFindNext function in fs/cifs/cifssmb.c in the Linux kernel before 3.1 allows remote CIFS servers to cause a denial of service (memory corruption) or possibly have unspecified other impact via a large length value in a response to a read request for a directory. <p>Publish Date: 2012-05-24 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-3191>CVE-2011-3191</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-3191">https://nvd.nist.gov/vuln/detail/CVE-2011-3191</a></p> <p>Release Date: 2012-05-24</p> <p>Fix Resolution: 3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in linux yocto autoclosed cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files fs cifs cifssmb c vulnerability details integer signedness error in the cifsfindnext function in fs cifs cifssmb c in the linux kernel before allows remote cifs servers to cause a denial of service memory corruption or possibly have unspecified other impact via a large length value in a response to a read request for a directory publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
53,151
13,261,029,907
IssuesEvent
2020-08-20 19:13:44
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[icetray] Adding services directly to the context bypasses I3TrayInfo (Trac #826)
Migrated from Trac combo core defect
With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this: ```text from I3Tray import I3Tray from icecube import icetray, dataclasses, dataio, phys_services tray = I3Tray() randomService = phys_services.I3GSLRandomService(seed=12345) tray.context["I3RandomService"] = randomService tray.AddModule("I3InfiniteSource") tray.AddModule("I3Writer","writer",filename="test.i3") tray.Execute(1) ``` The seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. There are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/826">https://code.icecube.wisc.edu/projects/icecube/ticket/826</a>, reported by cweaverand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T13:54:32", "_ts": "1550066072250335", "description": "With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:\n\n{{{\nfrom I3Tray import I3Tray\nfrom icecube import icetray, dataclasses, dataio, phys_services\ntray = I3Tray()\nrandomService = phys_services.I3GSLRandomService(seed=12345)\ntray.context[\"I3RandomService\"] = randomService\ntray.AddModule(\"I3InfiniteSource\")\ntray.AddModule(\"I3Writer\",\"writer\",filename=\"test.i3\")\ntray.Execute(1)\n}}}\n\nThe seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. \n\nThere are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. ", "reporter": "cweaver", "cc": "", "resolution": "insufficient resources", "time": "2014-12-04T22:37:06", "component": "combo core", "summary": "[icetray] Adding services directly to the context bypasses I3TrayInfo", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[icetray] Adding services directly to the context bypasses I3TrayInfo (Trac #826) - With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this: ```text from I3Tray import I3Tray from icecube import icetray, dataclasses, dataio, phys_services tray = I3Tray() randomService = phys_services.I3GSLRandomService(seed=12345) tray.context["I3RandomService"] = randomService tray.AddModule("I3InfiniteSource") tray.AddModule("I3Writer","writer",filename="test.i3") tray.Execute(1) ``` The seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. There are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/826">https://code.icecube.wisc.edu/projects/icecube/ticket/826</a>, reported by cweaverand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T13:54:32", "_ts": "1550066072250335", "description": "With the reworking of the context, it has become possible to add services without the aid of a service factory, which is convenient in many cases. Unfortunately, it also does an end run around the TrayInfo bookkeeping system, since this discovers services and their parameters by iterating over the list of factories contained in the tray. A relevant example is code like this:\n\n{{{\nfrom I3Tray import I3Tray\nfrom icecube import icetray, dataclasses, dataio, phys_services\ntray = I3Tray()\nrandomService = phys_services.I3GSLRandomService(seed=12345)\ntray.context[\"I3RandomService\"] = randomService\ntray.AddModule(\"I3InfiniteSource\")\ntray.AddModule(\"I3Writer\",\"writer\",filename=\"test.i3\")\ntray.Execute(1)\n}}}\n\nThe seed used by the RNG is a very important piece of information which the user may want to inspect after-the-fact, but using code like the above hides it from the TrayInfoService, as can be seen by examining the 'I' frame in the file generated by running this code. \n\nThere are a few difficulties associated with trying to correct this: First, arbitrary objects may be constructed and placed in the context, which may not conform to any standard interface to extract parameter values. Second, the TrayInfo may be queried at times after the service factories have run, so it would have to somehow avoid double counting the services installed by the factories, but these services' names may be unrelated to the names of their corresponding factories. ", "reporter": "cweaver", "cc": "", "resolution": "insufficient resources", "time": "2014-12-04T22:37:06", "component": "combo core", "summary": "[icetray] Adding services directly to the context bypasses I3TrayInfo", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
adding services directly to the context bypasses trac with the reworking of the context it has become possible to add services without the aid of a service factory which is convenient in many cases unfortunately it also does an end run around the trayinfo bookkeeping system since this discovers services and their parameters by iterating over the list of factories contained in the tray a relevant example is code like this text from import from icecube import icetray dataclasses dataio phys services tray randomservice phys services seed tray context randomservice tray addmodule tray addmodule writer filename test tray execute the seed used by the rng is a very important piece of information which the user may want to inspect after the fact but using code like the above hides it from the trayinfoservice as can be seen by examining the i frame in the file generated by running this code there are a few difficulties associated with trying to correct this first arbitrary objects may be constructed and placed in the context which may not conform to any standard interface to extract parameter values second the trayinfo may be queried at times after the service factories have run so it would have to somehow avoid double counting the services installed by the factories but these services names may be unrelated to the names of their corresponding factories migrated from json status closed changetime ts description with the reworking of the context it has become possible to add services without the aid of a service factory which is convenient in many cases unfortunately it also does an end run around the trayinfo bookkeeping system since this discovers services and their parameters by iterating over the list of factories contained in the tray a relevant example is code like this n n nfrom import nfrom icecube import icetray dataclasses dataio phys services ntray nrandomservice phys services seed ntray context randomservice ntray addmodule ntray addmodule writer filename test ntray execute n n nthe seed used by the rng is a very important piece of information which the user may want to inspect after the fact but using code like the above hides it from the trayinfoservice as can be seen by examining the i frame in the file generated by running this code n nthere are a few difficulties associated with trying to correct this first arbitrary objects may be constructed and placed in the context which may not conform to any standard interface to extract parameter values second the trayinfo may be queried at times after the service factories have run so it would have to somehow avoid double counting the services installed by the factories but these services names may be unrelated to the names of their corresponding factories reporter cweaver cc resolution insufficient resources time component combo core summary adding services directly to the context bypasses priority normal keywords milestone long term future owner olivas type defect
1
354,814
25,174,967,526
IssuesEvent
2022-11-11 08:24:28
mariavallejo20/proyecto_mvo
https://api.github.com/repos/mariavallejo20/proyecto_mvo
closed
[M0- Dev] Elección del gestor de dependencias
documentation
# Gestor de dependencias Un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades. Para ello, será necesario documentarse sobre distintas posibilidades y, finalmente, realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio. ## Criterios a seguir para la elección Para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta: 1. Estándares del lenguaje: * Fichero "package.json", no ha sido establecido como estándar para NodeJS, sin embargo es la práctica más extendida para la gestión de dependencias. 2. Buenas prácticas: * Debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores. * Criterios específicos de mi proyecto: * Se trata de un proyecto de pequeño tamaño * Se debe tener en cuenta la velocidad del gestor. * Emplear un único fichero para la gestión de las dependencias, en este caso "package.json", lo que permitirá no aumentar la deuda técnica en el futuro. * Se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias. 3. Actualización del gestor: Se deberá tener en cuenta la comunidad de la que dispone el gestor, ya que se busca tener un buen mantenimiento y actualización. Un buen soporte desembocará en una mayor seguridad, criterio que debemos tener en cuenta ya que es importante. ## Posibles gestores de dependencias * **NPM (Node Package Manager)**: - Es un gestor aconsejado por una gran comunidad de desarrolladores, ya que viene empaquetado con NodeJS. - Cuenta una gran comunidad y con un buen soporte y mantenimiento, con varias actualizaciones al año y siendo la última hace 5 meses. - Para la gestión de dependencias utiliza el archivo "package.json". - La instalación de paquetes no es demasiado rápida, sin embargo, teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá, este no será un gran inconveniente. - NPM realiza una evaluación de seguridad a partir de la versión 6, lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible. - NPM habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio. * **Bower**: - Para la gestión de dependencias emplea el archivo "boxer.json", lo que no cumple con la práctica extendida por la mayoría de los desarrolladores. Esto puede provocar deuda técnica en un futuro, por lo que lo descartaré como una opción. * **Yarn**: - Junto con NPM es el gestor de dependencias más utiizado por los desarrolladores de NodeJS. Sin embargo, este no viene instalado con la instalación del lenguaje. - Cuenta con un buen soporte y mantenimiento, con varias actualizaciones anuales; la última hace 5 meses. - Emplea para la gestión el archivo "package.json" lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro. - Yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente, lo que aumenta la velocidad de instalación de las dependencias. Sin embargo, teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa, aunque si habría que tenerlo en cuenta para proyectos mayores. - Mientra descarga paquetes, Yarn realiza una verificación de seguridad en segundo plano, aumentando la seguridad. - Yarn habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio. * **pnpm**: - PNPM es más rápido que los anteriores a la hora de las instalación de dependencias, sobre todo para paquetes de mayor tamaño. - En la seguridad NPM es similar a Yarn, ya que también emplea sumas de verificación. Además, pnpm verifica la integridad de su código antes de ejecutarlo. - PNPM cuenta con una gran comunidad de desarrolladores y un buen soporte, con actualizaciones frecuentes. - Para la gestión de dependencias utiliza el archivo "package.json". ## Elección Tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos, he decidido lo siguiente: NPM, Yarn y PNPM son buenas opciones para usar en mi proyecto. He decidido descartar NPM ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras. Por otro lado, entre PNPM y Yarn, **he decidido escoger PNPM**, ya que ambas tienes características similares que cumplen los criterios establecidos, buen soporte y documentación, pero PNPM presenta una mayor velicidad de instalación de los paquetes. Enlazado con la HU3[#6]
1.0
[M0- Dev] Elección del gestor de dependencias - # Gestor de dependencias Un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades. Para ello, será necesario documentarse sobre distintas posibilidades y, finalmente, realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio. ## Criterios a seguir para la elección Para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta: 1. Estándares del lenguaje: * Fichero "package.json", no ha sido establecido como estándar para NodeJS, sin embargo es la práctica más extendida para la gestión de dependencias. 2. Buenas prácticas: * Debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores. * Criterios específicos de mi proyecto: * Se trata de un proyecto de pequeño tamaño * Se debe tener en cuenta la velocidad del gestor. * Emplear un único fichero para la gestión de las dependencias, en este caso "package.json", lo que permitirá no aumentar la deuda técnica en el futuro. * Se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias. 3. Actualización del gestor: Se deberá tener en cuenta la comunidad de la que dispone el gestor, ya que se busca tener un buen mantenimiento y actualización. Un buen soporte desembocará en una mayor seguridad, criterio que debemos tener en cuenta ya que es importante. ## Posibles gestores de dependencias * **NPM (Node Package Manager)**: - Es un gestor aconsejado por una gran comunidad de desarrolladores, ya que viene empaquetado con NodeJS. - Cuenta una gran comunidad y con un buen soporte y mantenimiento, con varias actualizaciones al año y siendo la última hace 5 meses. - Para la gestión de dependencias utiliza el archivo "package.json". - La instalación de paquetes no es demasiado rápida, sin embargo, teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá, este no será un gran inconveniente. - NPM realiza una evaluación de seguridad a partir de la versión 6, lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible. - NPM habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio. * **Bower**: - Para la gestión de dependencias emplea el archivo "boxer.json", lo que no cumple con la práctica extendida por la mayoría de los desarrolladores. Esto puede provocar deuda técnica en un futuro, por lo que lo descartaré como una opción. * **Yarn**: - Junto con NPM es el gestor de dependencias más utiizado por los desarrolladores de NodeJS. Sin embargo, este no viene instalado con la instalación del lenguaje. - Cuenta con un buen soporte y mantenimiento, con varias actualizaciones anuales; la última hace 5 meses. - Emplea para la gestión el archivo "package.json" lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro. - Yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente, lo que aumenta la velocidad de instalación de las dependencias. Sin embargo, teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa, aunque si habría que tenerlo en cuenta para proyectos mayores. - Mientra descarga paquetes, Yarn realiza una verificación de seguridad en segundo plano, aumentando la seguridad. - Yarn habilita espacios de trabajo, lo que le permite administrar dependencias para varios proyectos desde un único repositorio. * **pnpm**: - PNPM es más rápido que los anteriores a la hora de las instalación de dependencias, sobre todo para paquetes de mayor tamaño. - En la seguridad NPM es similar a Yarn, ya que también emplea sumas de verificación. Además, pnpm verifica la integridad de su código antes de ejecutarlo. - PNPM cuenta con una gran comunidad de desarrolladores y un buen soporte, con actualizaciones frecuentes. - Para la gestión de dependencias utiliza el archivo "package.json". ## Elección Tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos, he decidido lo siguiente: NPM, Yarn y PNPM son buenas opciones para usar en mi proyecto. He decidido descartar NPM ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras. Por otro lado, entre PNPM y Yarn, **he decidido escoger PNPM**, ya que ambas tienes características similares que cumplen los criterios establecidos, buen soporte y documentación, pero PNPM presenta una mayor velicidad de instalación de los paquetes. Enlazado con la HU3[#6]
non_defect
elección del gestor de dependencias gestor de dependencias un problema principal para continuar con el desarrollo de mi servicio será la elección de un gestor de dependencias que se adecúe a mis necesidades para ello será necesario documentarse sobre distintas posibilidades y finalmente realizar una elección atendiendo a los criterios que se adapten a las buenas prácticas de mi servicio criterios a seguir para la elección para poder elegir un gestor de dependencias que se adecue a las buenas prácticas del lenguaje de programación de mi servicio tendré en cuenta estándares del lenguaje fichero package json no ha sido establecido como estándar para nodejs sin embargo es la práctica más extendida para la gestión de dependencias buenas prácticas debemos tener en cuenta que se aconseja con mayor frecuencia por las desarrolladores criterios específicos de mi proyecto se trata de un proyecto de pequeño tamaño se debe tener en cuenta la velocidad del gestor emplear un único fichero para la gestión de las dependencias en este caso package json lo que permitirá no aumentar la deuda técnica en el futuro se debe tener en cuenta que el gestor emplee un único repositorio para la gestión de dependencias actualización del gestor se deberá tener en cuenta la comunidad de la que dispone el gestor ya que se busca tener un buen mantenimiento y actualización un buen soporte desembocará en una mayor seguridad criterio que debemos tener en cuenta ya que es importante posibles gestores de dependencias npm node package manager es un gestor aconsejado por una gran comunidad de desarrolladores ya que viene empaquetado con nodejs cuenta una gran comunidad y con un buen soporte y mantenimiento con varias actualizaciones al año y siendo la última hace meses para la gestión de dependencias utiliza el archivo package json la instalación de paquetes no es demasiado rápida sin embargo teniendo en cuenta el tamaño de mi proyecto y la cantidad de dependencias que tendrá este no será un gran inconveniente npm realiza una evaluación de seguridad a partir de la versión lo que ayuda a evitar vulnerabilidades y asegura que ninguna dependencia sea incompatible npm habilita espacios de trabajo lo que le permite administrar dependencias para varios proyectos desde un único repositorio bower para la gestión de dependencias emplea el archivo boxer json lo que no cumple con la práctica extendida por la mayoría de los desarrolladores esto puede provocar deuda técnica en un futuro por lo que lo descartaré como una opción yarn junto con npm es el gestor de dependencias más utiizado por los desarrolladores de nodejs sin embargo este no viene instalado con la instalación del lenguaje cuenta con un buen soporte y mantenimiento con varias actualizaciones anuales la última hace meses emplea para la gestión el archivo package json lo que nos permite disminuir la posibilidad de crear deuda técnica en el futuro yarn instala las dependencias en paralelo permitiéndote agregar múltiples archivos simultáneamente lo que aumenta la velocidad de instalación de las dependencias sin embargo teniendo en cuenta el tamaño de nuestro proyecto está no es una ventaja significativa aunque si habría que tenerlo en cuenta para proyectos mayores mientra descarga paquetes yarn realiza una verificación de seguridad en segundo plano aumentando la seguridad yarn habilita espacios de trabajo lo que le permite administrar dependencias para varios proyectos desde un único repositorio pnpm pnpm es más rápido que los anteriores a la hora de las instalación de dependencias sobre todo para paquetes de mayor tamaño en la seguridad npm es similar a yarn ya que también emplea sumas de verificación además pnpm verifica la integridad de su código antes de ejecutarlo pnpm cuenta con una gran comunidad de desarrolladores y un buen soporte con actualizaciones frecuentes para la gestión de dependencias utiliza el archivo package json elección tras documentarme sobre los diferentes gestores de dependencias y una vez establecidos los criterios que cumplen o no cada uno de ellos he decidido lo siguiente npm yarn y pnpm son buenas opciones para usar en mi proyecto he decidido descartar npm ya que aunque presenta una características similares a las demás pero tiene un rendimiento y velocidad de instalación menor a las otras por otro lado entre pnpm y yarn he decidido escoger pnpm ya que ambas tienes características similares que cumplen los criterios establecidos buen soporte y documentación pero pnpm presenta una mayor velicidad de instalación de los paquetes enlazado con la
0
7,219
2,610,358,338
IssuesEvent
2015-02-26 19:56:01
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Blogspot post, Title and URL problem
auto-migrated Priority-Medium Type-Defect
``` What's the problem? When we posted to blogspot, url blogspot and title almost similar but today, when i post an article to blogspot, the url post using first paragraph from article? Are scribefire error problem connecting to blogger api? I try post manually from blogspot editor and i don't get error? What browser are you using? firefox 3.6 What version of ScribeFire are you running? scribefire next 1.9 ``` ----- Original issue reported on code.google.com by `tonit...@gmail.com` on 20 Nov 2011 at 1:44
1.0
Blogspot post, Title and URL problem - ``` What's the problem? When we posted to blogspot, url blogspot and title almost similar but today, when i post an article to blogspot, the url post using first paragraph from article? Are scribefire error problem connecting to blogger api? I try post manually from blogspot editor and i don't get error? What browser are you using? firefox 3.6 What version of ScribeFire are you running? scribefire next 1.9 ``` ----- Original issue reported on code.google.com by `tonit...@gmail.com` on 20 Nov 2011 at 1:44
defect
blogspot post title and url problem what s the problem when we posted to blogspot url blogspot and title almost similar but today when i post an article to blogspot the url post using first paragraph from article are scribefire error problem connecting to blogger api i try post manually from blogspot editor and i don t get error what browser are you using firefox what version of scribefire are you running scribefire next original issue reported on code google com by tonit gmail com on nov at
1
637,740
20,676,479,361
IssuesEvent
2022-03-10 09:46:23
AY2122S2-TIC4002-F18-5/tp2
https://api.github.com/repos/AY2122S2-TIC4002-F18-5/tp2
closed
Add a log command to register a brief log update on a Person
priority.High
Add a log command to register a brief text log message giving a recent or important update relevant to a team member Person.
1.0
Add a log command to register a brief log update on a Person - Add a log command to register a brief text log message giving a recent or important update relevant to a team member Person.
non_defect
add a log command to register a brief log update on a person add a log command to register a brief text log message giving a recent or important update relevant to a team member person
0
713,853
24,541,622,515
IssuesEvent
2022-10-12 04:38:55
AY2223S1-CS2113-F11-1/tp
https://api.github.com/repos/AY2223S1-CS2113-F11-1/tp
closed
Parse Command
type.Chore priority.High
When the user initiates a command in String, parse and execute the relevant command.
1.0
Parse Command - When the user initiates a command in String, parse and execute the relevant command.
non_defect
parse command when the user initiates a command in string parse and execute the relevant command
0
72,472
24,132,662,091
IssuesEvent
2022-09-21 08:44:49
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
SEVERE Exception for SQL limit clause [HZ-1436]
Type: Defect Source: Internal to-jira Team: Platform
**Describe the bug** An exception appears in the logs for an SQL clause that stops because of LIMIT **Expected behavior** No exception in logs, noise **To Reproduce** ``` public static void main(String[] args) throws Exception { Config config = new Config(); config.getJetConfig().setEnabled(true); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); hazelcastInstance.getMap("neil").put("Hello", "World"); hazelcastInstance.getMap("neil").put("Goodbye", "World"); String mapping = "CREATE MAPPING neil TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + String.class.getName() + "'," + " 'valueFormat' = 'java'," + " 'valueJavaClass' = '" + String.class.getName() + "'" + " )"; String query = "SELECT * FROM neil LIMIT 1"; List.of(mapping, query).forEach(sql -> hazelcastInstance.getSql().execute(sql)); } ``` gives ``` com.hazelcast.jet.JetException: Execution on a member failed: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause at com.hazelcast.jet.impl.LightMasterContext.findError(LightMasterContext.java:310) at com.hazelcast.jet.impl.LightMasterContext.lambda$createContext$2(LightMasterContext.java:164) at com.hazelcast.jet.impl.LightMasterContext.lambda$invokeOnParticipant$8(LightMasterContext.java:281) at com.hazelcast.spi.impl.AbstractInvocationFuture$WhenCompleteNode.lambda$execute$0(AbstractInvocationFuture.java:1564) at java.base/java.util.concurrent.ForkJoinTask$RunnableExecuteAction.exec(ForkJoinTask.java:1395) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:373) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1182) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1655) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1622) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Caused by: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause ``` **Additional context** Hazelcast Platform 5.2-SNAPSHOT (20220819 - 111323f)
1.0
SEVERE Exception for SQL limit clause [HZ-1436] - **Describe the bug** An exception appears in the logs for an SQL clause that stops because of LIMIT **Expected behavior** No exception in logs, noise **To Reproduce** ``` public static void main(String[] args) throws Exception { Config config = new Config(); config.getJetConfig().setEnabled(true); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); hazelcastInstance.getMap("neil").put("Hello", "World"); hazelcastInstance.getMap("neil").put("Goodbye", "World"); String mapping = "CREATE MAPPING neil TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + String.class.getName() + "'," + " 'valueFormat' = 'java'," + " 'valueJavaClass' = '" + String.class.getName() + "'" + " )"; String query = "SELECT * FROM neil LIMIT 1"; List.of(mapping, query).forEach(sql -> hazelcastInstance.getSql().execute(sql)); } ``` gives ``` com.hazelcast.jet.JetException: Execution on a member failed: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause at com.hazelcast.jet.impl.LightMasterContext.findError(LightMasterContext.java:310) at com.hazelcast.jet.impl.LightMasterContext.lambda$createContext$2(LightMasterContext.java:164) at com.hazelcast.jet.impl.LightMasterContext.lambda$invokeOnParticipant$8(LightMasterContext.java:281) at com.hazelcast.spi.impl.AbstractInvocationFuture$WhenCompleteNode.lambda$execute$0(AbstractInvocationFuture.java:1564) at java.base/java.util.concurrent.ForkJoinTask$RunnableExecuteAction.exec(ForkJoinTask.java:1395) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:373) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1182) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1655) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1622) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Caused by: com.hazelcast.sql.impl.ResultLimitReachedException: Done by reaching the item number in SQL LIMIT clause ``` **Additional context** Hazelcast Platform 5.2-SNAPSHOT (20220819 - 111323f)
defect
severe exception for sql limit clause describe the bug an exception appears in the logs for an sql clause that stops because of limit expected behavior no exception in logs noise to reproduce public static void main string args throws exception config config new config config getjetconfig setenabled true hazelcastinstance hazelcastinstance hazelcast newhazelcastinstance config hazelcastinstance getmap neil put hello world hazelcastinstance getmap neil put goodbye world string mapping create mapping neil type imap options keyformat java keyjavaclass string class getname valueformat java valuejavaclass string class getname string query select from neil limit list of mapping query foreach sql hazelcastinstance getsql execute sql gives com hazelcast jet jetexception execution on a member failed com hazelcast sql impl resultlimitreachedexception done by reaching the item number in sql limit clause at com hazelcast jet impl lightmastercontext finderror lightmastercontext java at com hazelcast jet impl lightmastercontext lambda createcontext lightmastercontext java at com hazelcast jet impl lightmastercontext lambda invokeonparticipant lightmastercontext java at com hazelcast spi impl abstractinvocationfuture whencompletenode lambda execute abstractinvocationfuture java at java base java util concurrent forkjointask runnableexecuteaction exec forkjointask java at java base java util concurrent forkjointask doexec forkjointask java at java base java util concurrent forkjoinpool workqueue toplevelexec forkjoinpool java at java base java util concurrent forkjoinpool scan forkjoinpool java at java base java util concurrent forkjoinpool runworker forkjoinpool java at java base java util concurrent forkjoinworkerthread run forkjoinworkerthread java caused by com hazelcast sql impl resultlimitreachedexception done by reaching the item number in sql limit clause additional context hazelcast platform snapshot
1
24,880
4,117,831,080
IssuesEvent
2016-06-08 09:05:50
abc306396358/pytesser
https://api.github.com/repos/abc306396358/pytesser
closed
How to get whitelist to work with pytesseract
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Trying to use the code that makes a whitelist for Tesseract like follows ocr = tesseract.TessBaseAPI() ocr.SetVariable("tessedit_char_whitelist", "0123456789;") ocr.SetPageSegMode(tesseract.PSM_AUTO) ocr.Init("C:\\Program Files (x86)\\Tesseract-OCR\\","eng",tesseract.OEM_DEFAULT) What is the expected output? What do you see instead? Intended output is to have only "0123456789;" characters be recognized when using the image_to_string() function. Using code like what is above, image_to_string() just ignores it and grabs whatever characters it finds. What version of the product are you using? On what operating system? pytesseract-0.1, Python 2.7, Windows 8.1 Please provide any additional information below. I've been trying everything people use for Tesseract-OCR, but that doesn't work with pytesseract. I haven't been able to find any solution or method to whitelisting with the image_to_string() function anywhere, which would be immensely helpful in improving the accuracy of the function. Thanks in advance for any help on the matter. ``` Original issue reported on code.google.com by `darke...@yahoo.com` on 9 Jun 2015 at 6:58
1.0
How to get whitelist to work with pytesseract - ``` What steps will reproduce the problem? Trying to use the code that makes a whitelist for Tesseract like follows ocr = tesseract.TessBaseAPI() ocr.SetVariable("tessedit_char_whitelist", "0123456789;") ocr.SetPageSegMode(tesseract.PSM_AUTO) ocr.Init("C:\\Program Files (x86)\\Tesseract-OCR\\","eng",tesseract.OEM_DEFAULT) What is the expected output? What do you see instead? Intended output is to have only "0123456789;" characters be recognized when using the image_to_string() function. Using code like what is above, image_to_string() just ignores it and grabs whatever characters it finds. What version of the product are you using? On what operating system? pytesseract-0.1, Python 2.7, Windows 8.1 Please provide any additional information below. I've been trying everything people use for Tesseract-OCR, but that doesn't work with pytesseract. I haven't been able to find any solution or method to whitelisting with the image_to_string() function anywhere, which would be immensely helpful in improving the accuracy of the function. Thanks in advance for any help on the matter. ``` Original issue reported on code.google.com by `darke...@yahoo.com` on 9 Jun 2015 at 6:58
defect
how to get whitelist to work with pytesseract what steps will reproduce the problem trying to use the code that makes a whitelist for tesseract like follows ocr tesseract tessbaseapi ocr setvariable tessedit char whitelist ocr setpagesegmode tesseract psm auto ocr init c program files tesseract ocr eng tesseract oem default what is the expected output what do you see instead intended output is to have only characters be recognized when using the image to string function using code like what is above image to string just ignores it and grabs whatever characters it finds what version of the product are you using on what operating system pytesseract python windows please provide any additional information below i ve been trying everything people use for tesseract ocr but that doesn t work with pytesseract i haven t been able to find any solution or method to whitelisting with the image to string function anywhere which would be immensely helpful in improving the accuracy of the function thanks in advance for any help on the matter original issue reported on code google com by darke yahoo com on jun at
1
66,149
6,989,730,772
IssuesEvent
2017-12-14 17:03:22
saltstack/salt
https://api.github.com/repos/saltstack/salt
opened
Add timezone windows Tests to Auto Test Suite
auto-test Windows
This task is to monitor the addition of four tests to the auto test suite: - [ ] timezone.get_zone - [ ] timezone.get_offset - [ ] timezone.set_zone - [ ] timezone.zone_compare There are unit tests but need to add integration tests
1.0
Add timezone windows Tests to Auto Test Suite - This task is to monitor the addition of four tests to the auto test suite: - [ ] timezone.get_zone - [ ] timezone.get_offset - [ ] timezone.set_zone - [ ] timezone.zone_compare There are unit tests but need to add integration tests
non_defect
add timezone windows tests to auto test suite this task is to monitor the addition of four tests to the auto test suite timezone get zone timezone get offset timezone set zone timezone zone compare there are unit tests but need to add integration tests
0
426,599
12,375,000,534
IssuesEvent
2020-05-19 03:18:26
mozilla/fx-private-relay
https://api.github.com/repos/mozilla/fx-private-relay
closed
Add the ability to see all their aliases from the addon
enhancement priority-P1 size:M 🧩 add-on
It's unclear which addresses have been created without going to the dashboard on the website. Should also consider a "manage addresses" link that directs someone to the website dashboard to view and manage their created records.
1.0
Add the ability to see all their aliases from the addon - It's unclear which addresses have been created without going to the dashboard on the website. Should also consider a "manage addresses" link that directs someone to the website dashboard to view and manage their created records.
non_defect
add the ability to see all their aliases from the addon it s unclear which addresses have been created without going to the dashboard on the website should also consider a manage addresses link that directs someone to the website dashboard to view and manage their created records
0
31,020
25,258,493,804
IssuesEvent
2022-11-15 20:21:26
pulibrary/dspace-development
https://api.github.com/repos/pulibrary/dspace-development
closed
DataSpace `dev` environment server is inaccessible
infrastructure dataspace
The following was reported: > 2:00 If someone could look at whatever's going on with https://dataspace-dev.princeton.edu/ that would be amazing because I anticipate Lynn will be asking about it at the meeting, but otherwise I can take a look when I'm back. I currently cannot access the application using https://dataspace-dev.princeton.edu/
1.0
DataSpace `dev` environment server is inaccessible - The following was reported: > 2:00 If someone could look at whatever's going on with https://dataspace-dev.princeton.edu/ that would be amazing because I anticipate Lynn will be asking about it at the meeting, but otherwise I can take a look when I'm back. I currently cannot access the application using https://dataspace-dev.princeton.edu/
non_defect
dataspace dev environment server is inaccessible the following was reported if someone could look at whatever s going on with that would be amazing because i anticipate lynn will be asking about it at the meeting but otherwise i can take a look when i m back i currently cannot access the application using
0
9,574
2,615,162,823
IssuesEvent
2015-03-01 06:42:15
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
WARNING: Failed to associate with XX:XX:XX:XX:XX:XX
auto-migrated Priority-Triage Type-Defect
``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) reaver 1.4 1. What operating system are you using (Linux is the only supported OS) vmware workstation linux 2.6 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -62 4. What is the manufacturer and model # of the device you are trying to crack? tp-link 321g v4 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b xx:xx:xx:xx:xx -vv 6. Please describe what you think the issue is. no idea 7. Paste the output from Reaver below. [+] Waiting for beacon from XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX ``` Original issue reported on code.google.com by `adsa...@gmail.com` on 6 Jun 2012 at 3:15
1.0
WARNING: Failed to associate with XX:XX:XX:XX:XX:XX - ``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) reaver 1.4 1. What operating system are you using (Linux is the only supported OS) vmware workstation linux 2.6 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -62 4. What is the manufacturer and model # of the device you are trying to crack? tp-link 321g v4 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b xx:xx:xx:xx:xx -vv 6. Please describe what you think the issue is. no idea 7. Paste the output from Reaver below. [+] Waiting for beacon from XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX [!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX ``` Original issue reported on code.google.com by `adsa...@gmail.com` on 6 Jun 2012 at 3:15
defect
warning failed to associate with xx xx xx xx xx xx a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os vmware workstation linux is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack tp link what is the entire command line string you are supplying to reaver reaver i b xx xx xx xx xx vv please describe what you think the issue is no idea paste the output from reaver below waiting for beacon from xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx warning failed to associate with xx xx xx xx xx xx original issue reported on code google com by adsa gmail com on jun at
1
48,559
13,135,248,517
IssuesEvent
2020-08-07 02:12:00
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
ExtraNodesetGenerator does not work with distributed mesh
T: defect
## Bug Description When specify IDs of nodes to be added in the node set, nodes with these IDs may only be on one processor, thus make call `BoundaryInfo::add_node` error out. ## Steps to Reproduce All current tests for this mesh generator are restricted with replicated mesh. NS module only use this to pin pressure at a node. If this cannot be fixed, those NS tests cannot run with distributed mesh. It might be an easy fix by protecting `add_node` call by querying if the node exist on the mesh. ## Impact Enable tests using this mesh generator to run in parallel with distributed mesh.
1.0
ExtraNodesetGenerator does not work with distributed mesh - ## Bug Description When specify IDs of nodes to be added in the node set, nodes with these IDs may only be on one processor, thus make call `BoundaryInfo::add_node` error out. ## Steps to Reproduce All current tests for this mesh generator are restricted with replicated mesh. NS module only use this to pin pressure at a node. If this cannot be fixed, those NS tests cannot run with distributed mesh. It might be an easy fix by protecting `add_node` call by querying if the node exist on the mesh. ## Impact Enable tests using this mesh generator to run in parallel with distributed mesh.
defect
extranodesetgenerator does not work with distributed mesh bug description when specify ids of nodes to be added in the node set nodes with these ids may only be on one processor thus make call boundaryinfo add node error out steps to reproduce all current tests for this mesh generator are restricted with replicated mesh ns module only use this to pin pressure at a node if this cannot be fixed those ns tests cannot run with distributed mesh it might be an easy fix by protecting add node call by querying if the node exist on the mesh impact enable tests using this mesh generator to run in parallel with distributed mesh
1
150,332
19,603,612,824
IssuesEvent
2022-01-06 06:07:21
snykiotcubedev/react-17.0.2
https://api.github.com/repos/snykiotcubedev/react-17.0.2
opened
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz
security vulnerability
## CVE-2020-7645 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary> <p>Launch latest Chrome with the Devtools Protocol port open</p> <p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p> <p> Dependency Hierarchy: - :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/react-17.0.2/commit/c546697344431dcd6c04b4fea877c488ef3a6ad5">c546697344431dcd6c04b4fea877c488ef3a6ad5</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems. <p>Publish Date: 2020-05-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-7645">https://nvd.nist.gov/vuln/detail/CVE-2020-7645</a></p> <p>Release Date: 2020-05-07</p> <p>Fix Resolution: chrome-launcher - 0.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz - ## CVE-2020-7645 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary> <p>Launch latest Chrome with the Devtools Protocol port open</p> <p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p> <p> Dependency Hierarchy: - :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/react-17.0.2/commit/c546697344431dcd6c04b4fea877c488ef3a6ad5">c546697344431dcd6c04b4fea877c488ef3a6ad5</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems. <p>Publish Date: 2020-05-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-7645">https://nvd.nist.gov/vuln/detail/CVE-2020-7645</a></p> <p>Release Date: 2020-05-07</p> <p>Fix Resolution: chrome-launcher - 0.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in chrome launcher tgz cve high severity vulnerability vulnerable library chrome launcher tgz launch latest chrome with the devtools protocol port open library home page a href dependency hierarchy x chrome launcher tgz vulnerable library found in head commit a href found in base branch main vulnerability details all versions of chrome launcher allow execution of arbitrary commands by controlling the home environment variable in linux operating systems publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution chrome launcher step up your open source security game with whitesource
0
164,512
20,364,519,507
IssuesEvent
2022-02-21 02:57:46
mxttwoods/blog
https://api.github.com/repos/mxttwoods/blog
closed
WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed
security vulnerability
## WS-2022-0008 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - gatsby-3.2.1.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - selfsigned-1.10.8.tgz - :x: **node-forge-0.10.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way. <p>Publish Date: 2022-01-08 <p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p> <p>Release Date: 2022-01-08</p> <p>Fix Resolution (node-forge): 1.2.1</p> <p>Direct dependency fix Resolution (gatsby): 3.11.0-alpha-qe-sm.92</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed - ## WS-2022-0008 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - gatsby-3.2.1.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - selfsigned-1.10.8.tgz - :x: **node-forge-0.10.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way. <p>Publish Date: 2022-01-08 <p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p> <p>Release Date: 2022-01-08</p> <p>Fix Resolution (node-forge): 1.2.1</p> <p>Direct dependency fix Resolution (gatsby): 3.11.0-alpha-qe-sm.92</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws medium detected in node forge tgz autoclosed ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy gatsby tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in base branch main vulnerability details the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution gatsby alpha qe sm step up your open source security game with whitesource
0
108,132
23,537,849,075
IssuesEvent
2022-08-20 00:24:04
happycube/ld-decode
https://api.github.com/repos/happycube/ld-decode
closed
ld-compress help points to old wiki location
ld-decode-tools deficiency
The wiki documentation for ld-compress has moved to it's own page now: https://github.com/happycube/ld-decode/wiki/ld-compress but the application is still pointing to the old location: Output: ``` ld-compress --help Usage: /usr/local/bin/ld-compress [-c] [-a] [-u] [-v] [-p] [-h] [-l <1-12>] [-g] file(s) Modes: -c Compress (default): Takes one or more .lds files and compresses them to .ldf files in the current directory. -u Uncompress: Takes one or more .ldf/.raw.oga files and uncompresses them to .lds files in the current directory. -a GPU Acceleration. Uses OpenCL or CUDA to accelerate encoding. See https://github.com/happycube/ld-decode/wiki/ld-decode-utilities -v Verify: Returns md5 checksums of the given .ldf/.raw.oga files and their contained .lds files for verification purposes. Options -p Progress: displays progress bars - requires pv to be installed. -h Help: This dialog. -l Compression level 1 - 12 (1 - 11 for GPU encoding). Default is 11 (10 for GPU). 6 is recommended for faster but fair compression. -g Use .raw.oga extension instead of .ldf when compressing. ```
1.0
ld-compress help points to old wiki location - The wiki documentation for ld-compress has moved to it's own page now: https://github.com/happycube/ld-decode/wiki/ld-compress but the application is still pointing to the old location: Output: ``` ld-compress --help Usage: /usr/local/bin/ld-compress [-c] [-a] [-u] [-v] [-p] [-h] [-l <1-12>] [-g] file(s) Modes: -c Compress (default): Takes one or more .lds files and compresses them to .ldf files in the current directory. -u Uncompress: Takes one or more .ldf/.raw.oga files and uncompresses them to .lds files in the current directory. -a GPU Acceleration. Uses OpenCL or CUDA to accelerate encoding. See https://github.com/happycube/ld-decode/wiki/ld-decode-utilities -v Verify: Returns md5 checksums of the given .ldf/.raw.oga files and their contained .lds files for verification purposes. Options -p Progress: displays progress bars - requires pv to be installed. -h Help: This dialog. -l Compression level 1 - 12 (1 - 11 for GPU encoding). Default is 11 (10 for GPU). 6 is recommended for faster but fair compression. -g Use .raw.oga extension instead of .ldf when compressing. ```
non_defect
ld compress help points to old wiki location the wiki documentation for ld compress has moved to it s own page now but the application is still pointing to the old location output ld compress help usage usr local bin ld compress file s modes c compress default takes one or more lds files and compresses them to ldf files in the current directory u uncompress takes one or more ldf raw oga files and uncompresses them to lds files in the current directory a gpu acceleration uses opencl or cuda to accelerate encoding see v verify returns checksums of the given ldf raw oga files and their contained lds files for verification purposes options p progress displays progress bars requires pv to be installed h help this dialog l compression level for gpu encoding default is for gpu is recommended for faster but fair compression g use raw oga extension instead of ldf when compressing
0
42,550
11,134,925,344
IssuesEvent
2019-12-20 13:09:00
nanopb/nanopb
https://api.github.com/repos/nanopb/nanopb
closed
Passing spaces and # in protoc arguments does not work
Component-Generator Priority-Low Type-Defect
When I use the `-L` or `--library-include-format=_some_folder/pb.h`, I expect in my `.pb.h` file to have `#include <_some_folder/pb.h`. However the script generated only `_some_folder/pb.h` for the include, not wrapped in `#include <>`. When I try to instead to `--library-include-format=\"#include <_some_folder/pb.h>\", I get various parsing errors with nanopb & bash. Is this the intended behaviour of this flag? Becuase that wouldn't make much sense. Let me know if theres any code/other info you all need.
1.0
Passing spaces and # in protoc arguments does not work - When I use the `-L` or `--library-include-format=_some_folder/pb.h`, I expect in my `.pb.h` file to have `#include <_some_folder/pb.h`. However the script generated only `_some_folder/pb.h` for the include, not wrapped in `#include <>`. When I try to instead to `--library-include-format=\"#include <_some_folder/pb.h>\", I get various parsing errors with nanopb & bash. Is this the intended behaviour of this flag? Becuase that wouldn't make much sense. Let me know if theres any code/other info you all need.
defect
passing spaces and in protoc arguments does not work when i use the l or library include format some folder pb h i expect in my pb h file to have include when i try to instead to library include format include i get various parsing errors with nanopb bash is this the intended behaviour of this flag becuase that wouldn t make much sense let me know if theres any code other info you all need
1
41,404
10,440,130,101
IssuesEvent
2019-09-18 08:03:15
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Nine parser bugs / missing features
C: Parser E: All Editions P: Medium T: Defect
First; congratulations on the release of 3.12! I've been looking forward to seeing it :) I've been using the `org.jooq.meta.extensions.ddl.DDLDatabase` generator to generate my code from SQL. So far I've had to manually keep two versions; one with the scripts that will be executed by Flyway on the mysql DB and another that is used by JOOQ to generate the code. This is an obvious potential source of bugs in addition to the extra work so I'd like to get rid of it. I tried running the new 3.12 on my SQL migration scripts and ran into nine errors: 1. After clause in alter statement ``` Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [10:60] ...LE `example` ADD COLUMN `field3` varchar(32) NULL [*]AFTER `field2`; ``` My understanding is that the order of the fields can have an impact on the performance of the queries. But I don't really care that much. Removing the AFTER clause will make the parsing work. EDIT: duplicate of #7241 2. Autoincrement in ALTER clause ``` Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:73] ...ADD COLUMN `example_id` int(10) unsigned NOT NULL [*]AUTO_INCREMENT ``` I don't see a workaround, and this feature is needed in my DB. EDIT: Reported as #9129. 3. Duplicate foreign key name in alter statement ``` Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Token '(' expected: [14:46] ...ink` [ERROR] ADD CONSTRAINT `fk_example_link` FOREIGN KEY [*]`fk_example_link` (`example_id`) REFERENCES `example` (`example_id`) ON DELETE C... ``` Workaround is to remove the second occurrence of the name. EDIT: Reported as #9130 4. Alter statement with first clause ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:69] ...le` ADD COLUMN `field1` int(10) unsigned NOT NULL [*]FIRST; ``` Same as with AFTER clause. EDIT: duplicate of #7241 5. Multiple statements in an ALTER statement ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [9:46] ...ple` [ERROR] ADD COLUMN `field2` int(10) unsigned NOT NULL[*], [ERROR] ADD COLUMN `field3` int(10) unsigned NOT NULL; ``` Workaround is to manually split things into individual statements. EDIT: Duplicate of #8846. 6. A BIT field with a default ``` org.jooq.impl.ParserException: Token ')' expected: [3:35] ...BLE `example` ( `field` bit(1) NOT NULL DEFAULT b[*]'0' ); ``` Workaround is rewriting the SQL without the b and single quotes. EDIT: Reported as #9131. 7. Creating an index ``` Error while exporting schema: SQL [create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index"))]; Unknown data type: "index_idx"; SQL statement: [ERROR] create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index")) [50004-199] ``` I got the normal way of creating indexes to work with JOOQ. However I did run into an additional bug when the table name isn't quoted: ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Error while exporting schema: SQL [create index "index_idx" on example("index")]; Table "EXAMPLE" not found; SQL statement: [ERROR] create index "index_idx" on example("index") ``` 8. Disabling foreign-key-checks + foreign key to table further into the script ``` SET FOREIGN_KEY_CHECKS = 0; ... ``` The obvious workaround is to re-arrange the statements so that the needed tables and fields are available in the right order. 9. Using btree statement ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [6:54] ...ER TABLE `example` ADD PRIMARY KEY (`example_id`) [*]USING BTREE; ``` The workaround is simple here; just remove the USING BTREE part. EDIT: Reported as #9132. There is a strange tenth issue that I ran into where I think MySQL actually erroneously succeeds when it should fail. It has to do with deleting a primary key when it no longer exists, but because the clause is contained in a single ALTER statement MySQL allows it. I created an example project here: https://github.com/codemonstur/jooq-generate-errors. Also looked through issue #8452 and couldn't find these. All SQL was generated using Schemasync.
1.0
Nine parser bugs / missing features - First; congratulations on the release of 3.12! I've been looking forward to seeing it :) I've been using the `org.jooq.meta.extensions.ddl.DDLDatabase` generator to generate my code from SQL. So far I've had to manually keep two versions; one with the scripts that will be executed by Flyway on the mysql DB and another that is used by JOOQ to generate the code. This is an obvious potential source of bugs in addition to the extra work so I'd like to get rid of it. I tried running the new 3.12 on my SQL migration scripts and ran into nine errors: 1. After clause in alter statement ``` Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [10:60] ...LE `example` ADD COLUMN `field3` varchar(32) NULL [*]AFTER `field2`; ``` My understanding is that the order of the fields can have an impact on the performance of the queries. But I don't really care that much. Removing the AFTER clause will make the parsing work. EDIT: duplicate of #7241 2. Autoincrement in ALTER clause ``` Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:73] ...ADD COLUMN `example_id` int(10) unsigned NOT NULL [*]AUTO_INCREMENT ``` I don't see a workaround, and this feature is needed in my DB. EDIT: Reported as #9129. 3. Duplicate foreign key name in alter statement ``` Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Token '(' expected: [14:46] ...ink` [ERROR] ADD CONSTRAINT `fk_example_link` FOREIGN KEY [*]`fk_example_link` (`example_id`) REFERENCES `example` (`example_id`) ON DELETE C... ``` Workaround is to remove the second occurrence of the name. EDIT: Reported as #9130 4. Alter statement with first clause ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [7:69] ...le` ADD COLUMN `field1` int(10) unsigned NOT NULL [*]FIRST; ``` Same as with AFTER clause. EDIT: duplicate of #7241 5. Multiple statements in an ALTER statement ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [9:46] ...ple` [ERROR] ADD COLUMN `field2` int(10) unsigned NOT NULL[*], [ERROR] ADD COLUMN `field3` int(10) unsigned NOT NULL; ``` Workaround is to manually split things into individual statements. EDIT: Duplicate of #8846. 6. A BIT field with a default ``` org.jooq.impl.ParserException: Token ')' expected: [3:35] ...BLE `example` ( `field` bit(1) NOT NULL DEFAULT b[*]'0' ); ``` Workaround is rewriting the SQL without the b and single quotes. EDIT: Reported as #9131. 7. Creating an index ``` Error while exporting schema: SQL [create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index"))]; Unknown data type: "index_idx"; SQL statement: [ERROR] create table "example"("example_id" bigint not null auto_increment, "index" bigint not null, primary key ("id"), index "index_idx" ("index")) [50004-199] ``` I got the normal way of creating indexes to work with JOOQ. However I did run into an additional bug when the table name isn't quoted: ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Error while exporting schema: SQL [create index "index_idx" on example("index")]; Table "EXAMPLE" not found; SQL statement: [ERROR] create index "index_idx" on example("index") ``` 8. Disabling foreign-key-checks + foreign key to table further into the script ``` SET FOREIGN_KEY_CHECKS = 0; ... ``` The obvious workaround is to re-arrange the statements so that the needed tables and fields are available in the right order. 9. Using btree statement ``` [ERROR] Failed to execute goal org.jooq:jooq-codegen-maven:3.12.0:generate (default) on project jooq-test: Error running jOOQ code generation tool: Unexpected token or missing query delimiter: [6:54] ...ER TABLE `example` ADD PRIMARY KEY (`example_id`) [*]USING BTREE; ``` The workaround is simple here; just remove the USING BTREE part. EDIT: Reported as #9132. There is a strange tenth issue that I ran into where I think MySQL actually erroneously succeeds when it should fail. It has to do with deleting a primary key when it no longer exists, but because the clause is contained in a single ALTER statement MySQL allows it. I created an example project here: https://github.com/codemonstur/jooq-generate-errors. Also looked through issue #8452 and couldn't find these. All SQL was generated using Schemasync.
defect
nine parser bugs missing features first congratulations on the release of i ve been looking forward to seeing it i ve been using the org jooq meta extensions ddl ddldatabase generator to generate my code from sql so far i ve had to manually keep two versions one with the scripts that will be executed by flyway on the mysql db and another that is used by jooq to generate the code this is an obvious potential source of bugs in addition to the extra work so i d like to get rid of it i tried running the new on my sql migration scripts and ran into nine errors after clause in alter statement error running jooq code generation tool unexpected token or missing query delimiter le example add column varchar null after my understanding is that the order of the fields can have an impact on the performance of the queries but i don t really care that much removing the after clause will make the parsing work edit duplicate of autoincrement in alter clause failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter add column example id int unsigned not null auto increment i don t see a workaround and this feature is needed in my db edit reported as duplicate foreign key name in alter statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool token expected ink add constraint fk example link foreign key fk example link example id references example example id on delete c workaround is to remove the second occurrence of the name edit reported as alter statement with first clause failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter le add column int unsigned not null first same as with after clause edit duplicate of multiple statements in an alter statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter ple add column int unsigned not null add column int unsigned not null workaround is to manually split things into individual statements edit duplicate of a bit field with a default org jooq impl parserexception token expected ble example field bit not null default b workaround is rewriting the sql without the b and single quotes edit reported as creating an index error while exporting schema sql unknown data type index idx sql statement create table example example id bigint not null auto increment index bigint not null primary key id index index idx index i got the normal way of creating indexes to work with jooq however i did run into an additional bug when the table name isn t quoted failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool error while exporting schema sql table example not found sql statement create index index idx on example index disabling foreign key checks foreign key to table further into the script set foreign key checks the obvious workaround is to re arrange the statements so that the needed tables and fields are available in the right order using btree statement failed to execute goal org jooq jooq codegen maven generate default on project jooq test error running jooq code generation tool unexpected token or missing query delimiter er table example add primary key example id using btree the workaround is simple here just remove the using btree part edit reported as there is a strange tenth issue that i ran into where i think mysql actually erroneously succeeds when it should fail it has to do with deleting a primary key when it no longer exists but because the clause is contained in a single alter statement mysql allows it i created an example project here also looked through issue and couldn t find these all sql was generated using schemasync
1
82,234
32,074,548,429
IssuesEvent
2023-09-25 10:07:49
vector-im/element-call
https://api.github.com/repos/vector-im/element-call
opened
Tiles near the bottom of the window in grid mode become un-clickable half way down.
T-Defect
### Steps to reproduce 1. Participate in a call in Grid mode 2. Resize the window if necessary to have tiles scrollable underneath the controls ### Outcome #### What did you expect? - Tiles not directly underneath the controls should be clickable. - Tiles directly underneath the controls should be clickable outside the bounds of the control buttons. #### What happened instead? There appears to be an invisible bar housing the controls that swallows up the mouse focus. https://github.com/vector-im/element-call/assets/6060466/ee25753b-9b65-469b-a1d4-08c43fa81689 ### Operating system macOS ### Browser information Firefox & Safari. ### URL for webapp call.element.io ### Will you send logs? No
1.0
Tiles near the bottom of the window in grid mode become un-clickable half way down. - ### Steps to reproduce 1. Participate in a call in Grid mode 2. Resize the window if necessary to have tiles scrollable underneath the controls ### Outcome #### What did you expect? - Tiles not directly underneath the controls should be clickable. - Tiles directly underneath the controls should be clickable outside the bounds of the control buttons. #### What happened instead? There appears to be an invisible bar housing the controls that swallows up the mouse focus. https://github.com/vector-im/element-call/assets/6060466/ee25753b-9b65-469b-a1d4-08c43fa81689 ### Operating system macOS ### Browser information Firefox & Safari. ### URL for webapp call.element.io ### Will you send logs? No
defect
tiles near the bottom of the window in grid mode become un clickable half way down steps to reproduce participate in a call in grid mode resize the window if necessary to have tiles scrollable underneath the controls outcome what did you expect tiles not directly underneath the controls should be clickable tiles directly underneath the controls should be clickable outside the bounds of the control buttons what happened instead there appears to be an invisible bar housing the controls that swallows up the mouse focus operating system macos browser information firefox safari url for webapp call element io will you send logs no
1
621,248
19,581,527,781
IssuesEvent
2022-01-04 22:05:08
paperclipui/paperclip
https://api.github.com/repos/paperclipui/paperclip
closed
[MTI] diffing sprint
priority: high effort: medium impact: high
TODOs: - [x] generate HTML report for diffed images - [ ] include coverage information for each screen? - should navigate to coverage report - [x] watch mode for visual diffing - [ ] **ability to lock changes to UI screens** - [ ] figure out how to store artifacts somewhere, and pull them down to be analyzed (s3?) - [ ] ability to set threshold limits
1.0
[MTI] diffing sprint - TODOs: - [x] generate HTML report for diffed images - [ ] include coverage information for each screen? - should navigate to coverage report - [x] watch mode for visual diffing - [ ] **ability to lock changes to UI screens** - [ ] figure out how to store artifacts somewhere, and pull them down to be analyzed (s3?) - [ ] ability to set threshold limits
non_defect
diffing sprint todos generate html report for diffed images include coverage information for each screen should navigate to coverage report watch mode for visual diffing ability to lock changes to ui screens figure out how to store artifacts somewhere and pull them down to be analyzed ability to set threshold limits
0
75,152
7,460,860,694
IssuesEvent
2018-03-30 21:45:29
Microsoft/PTVS
https://api.github.com/repos/Microsoft/PTVS
opened
PythonToolsMockTests fail with MEF composition issues
Tests
```cs System.ComponentModel.Composition.ImportCardinalityMismatchException: No exports were found that match the constraint: ContractName Microsoft.PythonTools.Editor.PythonEditorServices RequiredTypeIdentity Microsoft.PythonTools.Editor.PythonEditorServices ```
1.0
PythonToolsMockTests fail with MEF composition issues - ```cs System.ComponentModel.Composition.ImportCardinalityMismatchException: No exports were found that match the constraint: ContractName Microsoft.PythonTools.Editor.PythonEditorServices RequiredTypeIdentity Microsoft.PythonTools.Editor.PythonEditorServices ```
non_defect
pythontoolsmocktests fail with mef composition issues cs system componentmodel composition importcardinalitymismatchexception no exports were found that match the constraint contractname microsoft pythontools editor pythoneditorservices requiredtypeidentity microsoft pythontools editor pythoneditorservices
0
79,333
15,586,130,695
IssuesEvent
2021-03-18 01:14:30
revfactory/task-api
https://api.github.com/repos/revfactory/task-api
opened
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.16.jar
security vulnerability
## CVE-2020-11996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: task-api/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.3.RELEASE.jar - :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive. <p>Publish Date: 2020-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p> <p>Release Date: 2020-06-26</p> <p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.16.jar - ## CVE-2020-11996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: task-api/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.3.RELEASE.jar - :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive. <p>Publish Date: 2020-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p> <p>Release Date: 2020-06-26</p> <p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file task api build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details a specially crafted sequence of http requests sent to apache tomcat to to and to could trigger high cpu usage for several seconds if a sufficient number of such requests were made on concurrent http connections the server could become unresponsive publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat coyote org apache tomcat embed org apache tomcat embed step up your open source security game with whitesource
0
88,755
25,497,747,821
IssuesEvent
2022-11-27 21:55:00
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
reopened
@fluentui/e2e: Cypress command produces wrong paths for tests
Type: Bug :bug: Area: Build System Resolution: Soft Close
### Environment Information - **Package version(s)**: `master` - **Browser and OS versions**: Windows ### Please provide a reproduction of the bug in a codepen: This happens only on Windows as custom command handles only Unix-like paths. https://github.com/microsoft/fluentui/blob/dab45a2afe609709735b9b6e604a1af40d50e809/packages/fluentui/e2e/cypress/support/commands.js#L3-L9 ``` // Cypress passes to a custom command const testFilePath = tests\carouselClickableContent.spec.ts ``` Then it produces a wrong path that contains `tests-`, for example `tests-carousel-clickable-content` instead of `carousel-clickable-content`. As requested page does not exist, everything fails 💣 #### Actual behavior: ``` 1) Popup - on ESC key press "before each" hook for "is not closed if ESC is handled by children": Invariant Violation: The following error originated from your application code, not from Cypress. > Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports. Check the render method of `E2EExample`. When Cypress detects uncaught errors originating from your application it will automatically fail the current test. This behavior is configurable, and you can choose to turn this off by listening to the `uncaught:exception` event. https://on.cypress.io/uncaught-exception-from-application ``` #### Expected behavior: It works. ### Priorities and help requested: Are you willing to submit a PR to fix? No Requested priority: Normal
1.0
@fluentui/e2e: Cypress command produces wrong paths for tests - ### Environment Information - **Package version(s)**: `master` - **Browser and OS versions**: Windows ### Please provide a reproduction of the bug in a codepen: This happens only on Windows as custom command handles only Unix-like paths. https://github.com/microsoft/fluentui/blob/dab45a2afe609709735b9b6e604a1af40d50e809/packages/fluentui/e2e/cypress/support/commands.js#L3-L9 ``` // Cypress passes to a custom command const testFilePath = tests\carouselClickableContent.spec.ts ``` Then it produces a wrong path that contains `tests-`, for example `tests-carousel-clickable-content` instead of `carousel-clickable-content`. As requested page does not exist, everything fails 💣 #### Actual behavior: ``` 1) Popup - on ESC key press "before each" hook for "is not closed if ESC is handled by children": Invariant Violation: The following error originated from your application code, not from Cypress. > Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports. Check the render method of `E2EExample`. When Cypress detects uncaught errors originating from your application it will automatically fail the current test. This behavior is configurable, and you can choose to turn this off by listening to the `uncaught:exception` event. https://on.cypress.io/uncaught-exception-from-application ``` #### Expected behavior: It works. ### Priorities and help requested: Are you willing to submit a PR to fix? No Requested priority: Normal
non_defect
fluentui cypress command produces wrong paths for tests environment information package version s master browser and os versions windows please provide a reproduction of the bug in a codepen this happens only on windows as custom command handles only unix like paths cypress passes to a custom command const testfilepath tests carouselclickablecontent spec ts then it produces a wrong path that contains tests for example tests carousel clickable content instead of carousel clickable content as requested page does not exist everything fails 💣 actual behavior popup on esc key press before each hook for is not closed if esc is handled by children invariant violation the following error originated from your application code not from cypress element type is invalid expected a string for built in components or a class function for composite components but got undefined you likely forgot to export your component from the file it s defined in or you might have mixed up default and named imports check the render method of when cypress detects uncaught errors originating from your application it will automatically fail the current test this behavior is configurable and you can choose to turn this off by listening to the uncaught exception event expected behavior it works priorities and help requested are you willing to submit a pr to fix no requested priority normal
0
52,096
7,747,842,358
IssuesEvent
2018-05-30 05:57:41
benwasd/ssah
https://api.github.com/repos/benwasd/ssah
closed
Workshops mit Fachpersonen
documentation
Use Cases besprechen mit und Verbesserungen einpflegen. Feedback dokumentieren. - Administration der Skischule SnowValley Linden. - Martina, Administration Skischule Saas-Fee
1.0
Workshops mit Fachpersonen - Use Cases besprechen mit und Verbesserungen einpflegen. Feedback dokumentieren. - Administration der Skischule SnowValley Linden. - Martina, Administration Skischule Saas-Fee
non_defect
workshops mit fachpersonen use cases besprechen mit und verbesserungen einpflegen feedback dokumentieren administration der skischule snowvalley linden martina administration skischule saas fee
0
27,391
13,239,087,655
IssuesEvent
2020-08-19 02:23:01
radis/radis
https://api.github.com/repos/radis/radis
closed
Improve definition of width-grid resolution (res_L, res_G)
performance physics
With the manuscript draft being updated, we now have a much better handle on how to precisely define the errors in the width-grid for the "DLM"-approach. It would be good to propagate these updates to the RADIS code, but some executive decisions have to be made on how to implement it. There are basically three candidates of how the user could specify the desired error: - **`p`** -> _relative increase between two width-gridpoints_ - **`log_p`** = np.log(p) -> _actual step-size of the logarithmic width-grid_ - **`err_rms`** = RMS[Y_i]*np.log(p)/12 -> _contribution to RMS-error by a single line._ (The factor 1/12 = 1/2 x 1/6, where 1/6 comes from integrating the grid-alignment assuming uniform distribution between 0 and 1) Each has their own advantage, namely: - **`p`**: unambiguous what is meant, and makes sense from a human perspective (e.g. `p = 1.2` means 20% increase between gridpoints) - **`log_p`**: equally unambiguous, but perhaps makes less sense from a human perspective (e.g. what really does `log_p = 0.2` mean?). Advantage is that when `log_p` is closer to zero, it means higher resolution, which perhaps is more intuitive. - **`err_rms`**: Makes most sense from human point of view, and resolution also gets better as `err_rms` gets closer to zero. Problem is that choosing the RMS-error as the defining error measure is somewhat arbitrary, and depending on the kind of weight used (1,2, or 3), the same value of `err_rms` will lead to different `p` and `log_p` values, which the user may not be aware of (but perhaps equally possibly they do not care about) Anyway so that's a decision that has to be made. My personal preference is in the presented order (with **`p`** most favorable), but of course I'd like to hear everyone's thoughts.
True
Improve definition of width-grid resolution (res_L, res_G) - With the manuscript draft being updated, we now have a much better handle on how to precisely define the errors in the width-grid for the "DLM"-approach. It would be good to propagate these updates to the RADIS code, but some executive decisions have to be made on how to implement it. There are basically three candidates of how the user could specify the desired error: - **`p`** -> _relative increase between two width-gridpoints_ - **`log_p`** = np.log(p) -> _actual step-size of the logarithmic width-grid_ - **`err_rms`** = RMS[Y_i]*np.log(p)/12 -> _contribution to RMS-error by a single line._ (The factor 1/12 = 1/2 x 1/6, where 1/6 comes from integrating the grid-alignment assuming uniform distribution between 0 and 1) Each has their own advantage, namely: - **`p`**: unambiguous what is meant, and makes sense from a human perspective (e.g. `p = 1.2` means 20% increase between gridpoints) - **`log_p`**: equally unambiguous, but perhaps makes less sense from a human perspective (e.g. what really does `log_p = 0.2` mean?). Advantage is that when `log_p` is closer to zero, it means higher resolution, which perhaps is more intuitive. - **`err_rms`**: Makes most sense from human point of view, and resolution also gets better as `err_rms` gets closer to zero. Problem is that choosing the RMS-error as the defining error measure is somewhat arbitrary, and depending on the kind of weight used (1,2, or 3), the same value of `err_rms` will lead to different `p` and `log_p` values, which the user may not be aware of (but perhaps equally possibly they do not care about) Anyway so that's a decision that has to be made. My personal preference is in the presented order (with **`p`** most favorable), but of course I'd like to hear everyone's thoughts.
non_defect
improve definition of width grid resolution res l res g with the manuscript draft being updated we now have a much better handle on how to precisely define the errors in the width grid for the dlm approach it would be good to propagate these updates to the radis code but some executive decisions have to be made on how to implement it there are basically three candidates of how the user could specify the desired error p relative increase between two width gridpoints log p np log p actual step size of the logarithmic width grid err rms rms np log p contribution to rms error by a single line the factor x where comes from integrating the grid alignment assuming uniform distribution between and each has their own advantage namely p unambiguous what is meant and makes sense from a human perspective e g p means increase between gridpoints log p equally unambiguous but perhaps makes less sense from a human perspective e g what really does log p mean advantage is that when log p is closer to zero it means higher resolution which perhaps is more intuitive err rms makes most sense from human point of view and resolution also gets better as err rms gets closer to zero problem is that choosing the rms error as the defining error measure is somewhat arbitrary and depending on the kind of weight used or the same value of err rms will lead to different p and log p values which the user may not be aware of but perhaps equally possibly they do not care about anyway so that s a decision that has to be made my personal preference is in the presented order with p most favorable but of course i d like to hear everyone s thoughts
0
29,405
5,682,679,605
IssuesEvent
2017-04-13 10:22:42
bancika/diy-layout-creator
https://api.github.com/repos/bancika/diy-layout-creator
closed
Trace mask -> export to png doesn't add .png to filename
auto-migrated Priority-Medium Type-Defect
``` When you export (at least) trace mask to png and just type the filename part the program won't add the extension part (.png). This might be true for all exports but I've only tested traces to png. ``` Original issue reported on code.google.com by `per.tre...@gmail.com` on 25 Jan 2011 at 10:26
1.0
Trace mask -> export to png doesn't add .png to filename - ``` When you export (at least) trace mask to png and just type the filename part the program won't add the extension part (.png). This might be true for all exports but I've only tested traces to png. ``` Original issue reported on code.google.com by `per.tre...@gmail.com` on 25 Jan 2011 at 10:26
defect
trace mask export to png doesn t add png to filename when you export at least trace mask to png and just type the filename part the program won t add the extension part png this might be true for all exports but i ve only tested traces to png original issue reported on code google com by per tre gmail com on jan at
1
42,501
11,092,169,162
IssuesEvent
2019-12-15 17:10:21
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
sparse eigs (arpack) shift-invert drops the smallest eigenvalue for some k
defect scipy.sparse.linalg
scipy.sparse.linalg.eigs shift-invert mode for the SM, smallest-magnitude eigenvalue `( A=sparse.rand(), k=k, which='LM', sigma=0 )` drops the smallest one for some k: k 1 shift: [ 17 +0j ] k 2 shift: [ 17 +0j 52 +0j ] k 3 shift: [ 52 +0j -23 -47j -23 +47j ] k 4 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j ] k 5 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j ] k 6 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j ] k 7 shift: [ 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j 1e+02 -36j 1e+02 +36j ] It's hard to believe that this is an ARPACK bug; what else could be wrong ? Versions: numpy 1.17.4 scipy 1.3.3 python 3.7.3 macos 10.10.5 3581068 24 Nov 18:25 .../libopenblasp-r0.3.7.dev.dylib 669500 24 Nov 18:25 .../_arpack.cpython-37m-darwin.so Notes: The sometimes-missing smallest eigenvalue from `sparse.rand` with various n, density, seed is, so far, real. I take `v0=ones`, not random -- this should be the default. The nice scipy arpack tutorial says "which = 'SM' may lead to slow execution time and/or anomalous results". That's too weak --in fact SM often fails on many `sparse.rand` matrices. #### Reproducing code example: <!-- If you place your code between the triple backticks below, it will be rendered as a code block. --> #!/usr/bin/env python """ test scipy.sparse.linalg.eigs LM shift-invert: some k drop the smallest real eigenvalue ? """ # https://docs.scipy.org/doc/scipy/reference/tutorial/arpack.html X.XT symmetric, easy # which = 'SM' may lead to slow execution time and/or anomalous results # https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.eigs.html # denis 10 Dec 2019 from __future__ import division, print_function import sys import numpy as np import scipy from scipy import sparse from scipy.sparse.linalg import eigs, ArpackError, ArpackNoConvergence try: import scikits.umfpack # eigs -> splu spsolve ? print( "umfpack:", scikits.umfpack.__version__ ) except ImportError: pass def randomsparse( n, density=0.1, seed=0 ): """ -> n x n scipy.sparse.rand csc, uniform -1 .. 1 """ random = np.random.RandomState( seed=seed ) A = sparse.rand( n, n, density, format="csc", random_state=random ) # uniform 0 .. 1 # if distrib: A.data *= 2 A.data -= 1 print( "randomsparse: n %d density %.2g seed %d A %.2g .. %.2g nnz %d " % ( n, density, seed, A.data.min(), A.data.max(), A.nnz )) return A def zsort( z ): """ sort z abs increasing, then imag """ jsort = np.lexsort( [np.imag(z), np.abs(z)] ) return z[jsort] np.set_printoptions( threshold=20, edgeitems=10, linewidth=140, formatter = dict( float = lambda x: "%.2g" % x, # float arrays %.2g complexfloat = " {0.real:.2g} {0.imag:+.2g}j " .format )) print( "\n" + 80 * "=" ) print( "versions: numpy %s scipy %s python %s " % ( np.__version__, scipy.__version__ , sys.version.split()[0] )) #............................................................................... # gen -- gen = randomsparse n = 10 density = .5 seed = 0 # some -> LinAlgWarning singular, lu_factor dense = True # test arpack dense first # eigs -- kmax = 8 sigma = 0 tol = 0 # 0: machine precision, eps ? v0 = 1 # None: random, 1: ones SM = False # to change these params, run this.py a=1 b=None 'c = expr' ... in sh or ipython -- for arg in sys.argv[1:]: exec( arg ) genparams = "%s n %d density %.2g seed %d dense %s " % ( gen.__name__, n, density, seed, dense ) eigsparams = "sigma %s tol %g v0 %s SM %d " % ( sigma, tol, v0, SM ) print( "gen params:", genparams ) print( "eigs params:", eigsparams ) kmax = min( kmax, n - 2 ) if v0 is not None: v0 = np.ones( n ) #............................................................................... # generate a test matrix A -- # may be singular for some seeds: arpack.py sometimes calls lu_factor, SM doesn't A = randomsparse( n, density=density, seed=seed ) A *= 100 if dense: # test arpack dense first A = A.A #............................................................................... print( "\n# scipy.sparse.linalg.eigs (arpack) shift-invert vs. SM" ) for k in range( 1, kmax+1 ): evals = eigs( A, k=k, which='LM', sigma=0, v0=v0, maxiter=None, tol=tol, return_eigenvectors=False ) # eigs(A, k=6, M=None, sigma=None, which='LM', v0=None, maxiter=None, tol=0, # return_eigenvectors=True, Minv=None, OPinv=None, OPpart=None) print( "k %d shift: %s" % (k, zsort( evals ))) if SM: for k in range( 1, kmax+1 ): try: evals = eigs( A, k=k, which='SM', v0=v0, maxiter=None, tol=tol, return_eigenvectors=False ) print( "k %d SM: %s" % (k, zsort( evals )[0] )) except (ArpackError, ArpackNoConvergence) as msg: print( "k %d SM: %s" % (k, msg) )
1.0
sparse eigs (arpack) shift-invert drops the smallest eigenvalue for some k - scipy.sparse.linalg.eigs shift-invert mode for the SM, smallest-magnitude eigenvalue `( A=sparse.rand(), k=k, which='LM', sigma=0 )` drops the smallest one for some k: k 1 shift: [ 17 +0j ] k 2 shift: [ 17 +0j 52 +0j ] k 3 shift: [ 52 +0j -23 -47j -23 +47j ] k 4 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j ] k 5 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j ] k 6 shift: [ 17 +0j 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j ] k 7 shift: [ 52 +0j -23 -47j -23 +47j 71 +0j -1e+02 +0j 1e+02 -36j 1e+02 +36j ] It's hard to believe that this is an ARPACK bug; what else could be wrong ? Versions: numpy 1.17.4 scipy 1.3.3 python 3.7.3 macos 10.10.5 3581068 24 Nov 18:25 .../libopenblasp-r0.3.7.dev.dylib 669500 24 Nov 18:25 .../_arpack.cpython-37m-darwin.so Notes: The sometimes-missing smallest eigenvalue from `sparse.rand` with various n, density, seed is, so far, real. I take `v0=ones`, not random -- this should be the default. The nice scipy arpack tutorial says "which = 'SM' may lead to slow execution time and/or anomalous results". That's too weak --in fact SM often fails on many `sparse.rand` matrices. #### Reproducing code example: <!-- If you place your code between the triple backticks below, it will be rendered as a code block. --> #!/usr/bin/env python """ test scipy.sparse.linalg.eigs LM shift-invert: some k drop the smallest real eigenvalue ? """ # https://docs.scipy.org/doc/scipy/reference/tutorial/arpack.html X.XT symmetric, easy # which = 'SM' may lead to slow execution time and/or anomalous results # https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.eigs.html # denis 10 Dec 2019 from __future__ import division, print_function import sys import numpy as np import scipy from scipy import sparse from scipy.sparse.linalg import eigs, ArpackError, ArpackNoConvergence try: import scikits.umfpack # eigs -> splu spsolve ? print( "umfpack:", scikits.umfpack.__version__ ) except ImportError: pass def randomsparse( n, density=0.1, seed=0 ): """ -> n x n scipy.sparse.rand csc, uniform -1 .. 1 """ random = np.random.RandomState( seed=seed ) A = sparse.rand( n, n, density, format="csc", random_state=random ) # uniform 0 .. 1 # if distrib: A.data *= 2 A.data -= 1 print( "randomsparse: n %d density %.2g seed %d A %.2g .. %.2g nnz %d " % ( n, density, seed, A.data.min(), A.data.max(), A.nnz )) return A def zsort( z ): """ sort z abs increasing, then imag """ jsort = np.lexsort( [np.imag(z), np.abs(z)] ) return z[jsort] np.set_printoptions( threshold=20, edgeitems=10, linewidth=140, formatter = dict( float = lambda x: "%.2g" % x, # float arrays %.2g complexfloat = " {0.real:.2g} {0.imag:+.2g}j " .format )) print( "\n" + 80 * "=" ) print( "versions: numpy %s scipy %s python %s " % ( np.__version__, scipy.__version__ , sys.version.split()[0] )) #............................................................................... # gen -- gen = randomsparse n = 10 density = .5 seed = 0 # some -> LinAlgWarning singular, lu_factor dense = True # test arpack dense first # eigs -- kmax = 8 sigma = 0 tol = 0 # 0: machine precision, eps ? v0 = 1 # None: random, 1: ones SM = False # to change these params, run this.py a=1 b=None 'c = expr' ... in sh or ipython -- for arg in sys.argv[1:]: exec( arg ) genparams = "%s n %d density %.2g seed %d dense %s " % ( gen.__name__, n, density, seed, dense ) eigsparams = "sigma %s tol %g v0 %s SM %d " % ( sigma, tol, v0, SM ) print( "gen params:", genparams ) print( "eigs params:", eigsparams ) kmax = min( kmax, n - 2 ) if v0 is not None: v0 = np.ones( n ) #............................................................................... # generate a test matrix A -- # may be singular for some seeds: arpack.py sometimes calls lu_factor, SM doesn't A = randomsparse( n, density=density, seed=seed ) A *= 100 if dense: # test arpack dense first A = A.A #............................................................................... print( "\n# scipy.sparse.linalg.eigs (arpack) shift-invert vs. SM" ) for k in range( 1, kmax+1 ): evals = eigs( A, k=k, which='LM', sigma=0, v0=v0, maxiter=None, tol=tol, return_eigenvectors=False ) # eigs(A, k=6, M=None, sigma=None, which='LM', v0=None, maxiter=None, tol=0, # return_eigenvectors=True, Minv=None, OPinv=None, OPpart=None) print( "k %d shift: %s" % (k, zsort( evals ))) if SM: for k in range( 1, kmax+1 ): try: evals = eigs( A, k=k, which='SM', v0=v0, maxiter=None, tol=tol, return_eigenvectors=False ) print( "k %d SM: %s" % (k, zsort( evals )[0] )) except (ArpackError, ArpackNoConvergence) as msg: print( "k %d SM: %s" % (k, msg) )
defect
sparse eigs arpack shift invert drops the smallest eigenvalue for some k scipy sparse linalg eigs shift invert mode for the sm smallest magnitude eigenvalue a sparse rand k k which lm sigma drops the smallest one for some k k shift k shift k shift k shift k shift k shift k shift it s hard to believe that this is an arpack bug what else could be wrong versions numpy scipy python macos nov libopenblasp dev dylib nov arpack cpython darwin so notes the sometimes missing smallest eigenvalue from sparse rand with various n density seed is so far real i take ones not random this should be the default the nice scipy arpack tutorial says which sm may lead to slow execution time and or anomalous results that s too weak in fact sm often fails on many sparse rand matrices reproducing code example if you place your code between the triple backticks below it will be rendered as a code block usr bin env python test scipy sparse linalg eigs lm shift invert some k drop the smallest real eigenvalue x xt symmetric easy which sm may lead to slow execution time and or anomalous results denis dec from future import division print function import sys import numpy as np import scipy from scipy import sparse from scipy sparse linalg import eigs arpackerror arpacknoconvergence try import scikits umfpack eigs splu spsolve print umfpack scikits umfpack version except importerror pass def randomsparse n density seed n x n scipy sparse rand csc uniform random np random randomstate seed seed a sparse rand n n density format csc random state random uniform if distrib a data a data print randomsparse n d density seed d a nnz d n density seed a data min a data max a nnz return a def zsort z sort z abs increasing then imag jsort np lexsort return z np set printoptions threshold edgeitems linewidth formatter dict float lambda x x float arrays complexfloat real imag j format print n print versions numpy s scipy s python s np version scipy version sys version split gen gen randomsparse n density seed some linalgwarning singular lu factor dense true test arpack dense first eigs kmax sigma tol machine precision eps none random ones sm false to change these params run this py a b none c expr in sh or ipython for arg in sys argv exec arg genparams s n d density seed d dense s gen name n density seed dense eigsparams sigma s tol g s sm d sigma tol sm print gen params genparams print eigs params eigsparams kmax min kmax n if is not none np ones n generate a test matrix a may be singular for some seeds arpack py sometimes calls lu factor sm doesn t a randomsparse n density density seed seed a if dense test arpack dense first a a a print n scipy sparse linalg eigs arpack shift invert vs sm for k in range kmax evals eigs a k k which lm sigma maxiter none tol tol return eigenvectors false eigs a k m none sigma none which lm none maxiter none tol return eigenvectors true minv none opinv none oppart none print k d shift s k zsort evals if sm for k in range kmax try evals eigs a k k which sm maxiter none tol tol return eigenvectors false print k d sm s k zsort evals except arpackerror arpacknoconvergence as msg print k d sm s k msg
1
19,513
3,218,564,272
IssuesEvent
2015-10-08 02:32:13
darkfeline/python-abx
https://api.github.com/repos/darkfeline/python-abx
closed
ImportError: No module named gst
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Download source zip and extract 2. ./abx.py What is the expected output? An ABX comparator What do you see instead? Traceback (most recent call last): File "./abx.py", line 6, in <module> import gst ImportError: No module named gst What version of the product are you using? 2012-05-26 On what operating system? Mint 17 Cinnamon 64-bit live Please provide any additional information below. Python 2.7.6 ``` Original issue reported on code.google.com by `southi...@gmail.com` on 1 Mar 2015 at 4:53
1.0
ImportError: No module named gst - ``` What steps will reproduce the problem? 1. Download source zip and extract 2. ./abx.py What is the expected output? An ABX comparator What do you see instead? Traceback (most recent call last): File "./abx.py", line 6, in <module> import gst ImportError: No module named gst What version of the product are you using? 2012-05-26 On what operating system? Mint 17 Cinnamon 64-bit live Please provide any additional information below. Python 2.7.6 ``` Original issue reported on code.google.com by `southi...@gmail.com` on 1 Mar 2015 at 4:53
defect
importerror no module named gst what steps will reproduce the problem download source zip and extract abx py what is the expected output an abx comparator what do you see instead traceback most recent call last file abx py line in import gst importerror no module named gst what version of the product are you using on what operating system mint cinnamon bit live please provide any additional information below python original issue reported on code google com by southi gmail com on mar at
1
406,269
27,557,434,044
IssuesEvent
2023-03-07 19:04:25
pharmaverse/admiral
https://api.github.com/repos/pharmaverse/admiral
closed
Documentation: derive_vars_duration additional example
documentation good first issue release Q1-2023
derive_vars_duration start_date and end_date inputs can be counter-intuitive when start_date needs to be an end data variable and end_date needs to be a start date variable. "Last Dose Relative Time" LDRELTM requires start_date = LDOSEDTM, end_date = ASTDTM so that LDRELTM will be positive (unless there are data issues). Suggestion is to include an example for "Derive adverse event start time since last dose in hours" showing missing data examples. ``` data <- tribble( ~USUBJID, ~ASTDTM, ~LDOSEDTM, "P01", ymd_hms("2019-08-09T04:30:56"), ymd_hms("2019-08-08T10:05:00"), "P02", ymd_hms("2019-11-11T23:59:59"), ymd_hms("2019-10-11T11:37:00"), "P03", ymd_hms("2019-11-11T00:00:00"), ymd_hms("2019-11-10T23:59:59"), "P04", ymd_hms("2019-11-11T12:34:56"),NA, "P05", NA, ymd_hms("2019-09-28T12:34:56") ) derive_vars_duration( data, new_var = LDRELTM, new_var_unit = LDRELTMU, start_date = LDOSEDTM, end_date = ASTDTM, in_unit = "hours", out_unit = "hours", add_one = FALSE ) ```
1.0
Documentation: derive_vars_duration additional example - derive_vars_duration start_date and end_date inputs can be counter-intuitive when start_date needs to be an end data variable and end_date needs to be a start date variable. "Last Dose Relative Time" LDRELTM requires start_date = LDOSEDTM, end_date = ASTDTM so that LDRELTM will be positive (unless there are data issues). Suggestion is to include an example for "Derive adverse event start time since last dose in hours" showing missing data examples. ``` data <- tribble( ~USUBJID, ~ASTDTM, ~LDOSEDTM, "P01", ymd_hms("2019-08-09T04:30:56"), ymd_hms("2019-08-08T10:05:00"), "P02", ymd_hms("2019-11-11T23:59:59"), ymd_hms("2019-10-11T11:37:00"), "P03", ymd_hms("2019-11-11T00:00:00"), ymd_hms("2019-11-10T23:59:59"), "P04", ymd_hms("2019-11-11T12:34:56"),NA, "P05", NA, ymd_hms("2019-09-28T12:34:56") ) derive_vars_duration( data, new_var = LDRELTM, new_var_unit = LDRELTMU, start_date = LDOSEDTM, end_date = ASTDTM, in_unit = "hours", out_unit = "hours", add_one = FALSE ) ```
non_defect
documentation derive vars duration additional example derive vars duration start date and end date inputs can be counter intuitive when start date needs to be an end data variable and end date needs to be a start date variable last dose relative time ldreltm requires start date ldosedtm end date astdtm so that ldreltm will be positive unless there are data issues suggestion is to include an example for derive adverse event start time since last dose in hours showing missing data examples data tribble usubjid astdtm ldosedtm ymd hms ymd hms ymd hms ymd hms ymd hms ymd hms ymd hms na na ymd hms derive vars duration data new var ldreltm new var unit ldreltmu start date ldosedtm end date astdtm in unit hours out unit hours add one false
0
33,425
7,121,468,065
IssuesEvent
2018-01-19 07:54:08
MethodicalAcceleratorDesign/MAD-X
https://api.github.com/repos/MethodicalAcceleratorDesign/MAD-X
closed
AC-dipole in PTC but not in convertor
defect
*Issue migrated from trac ticket # 108* **component:** ptc_proper | **priority:** major ___ #### 2012-01-30 13:50:56: frank.schmidt commented date: 2011.11.19 The AC-dipole is in PTC and lacking in the convertor. ___ #### 2012-01-30 13:51:16: @ldeniau changed title from *AC-dipole is in PTC but not in convertor* to *AC-dipole in PTC but not in convertor* ___ #### 2012-12-07 13:02:38: @ldeniau changed owner from ** to *piotr.skowronski* ___ #### 2012-12-07 13:02:38: @ldeniau changed status from *new* to *assigned*
1.0
AC-dipole in PTC but not in convertor - *Issue migrated from trac ticket # 108* **component:** ptc_proper | **priority:** major ___ #### 2012-01-30 13:50:56: frank.schmidt commented date: 2011.11.19 The AC-dipole is in PTC and lacking in the convertor. ___ #### 2012-01-30 13:51:16: @ldeniau changed title from *AC-dipole is in PTC but not in convertor* to *AC-dipole in PTC but not in convertor* ___ #### 2012-12-07 13:02:38: @ldeniau changed owner from ** to *piotr.skowronski* ___ #### 2012-12-07 13:02:38: @ldeniau changed status from *new* to *assigned*
defect
ac dipole in ptc but not in convertor issue migrated from trac ticket component ptc proper priority major frank schmidt commented date the ac dipole is in ptc and lacking in the convertor ldeniau changed title from ac dipole is in ptc but not in convertor to ac dipole in ptc but not in convertor ldeniau changed owner from to piotr skowronski ldeniau changed status from new to assigned
1
347,388
10,429,048,757
IssuesEvent
2019-09-17 01:07:28
eveporcello/graphql-fun
https://api.github.com/repos/eveporcello/graphql-fun
opened
Review Notes target Game Prototypes
high-priority planning
Together we need to review the notes and see what game features we should prototype. - [ ] Identify Games (easiest - hardest) - [ ] Create a new task for each game - [ ] Draft Schemas for each game - [ ] Plan Steps/Issues for Each Game - [ ] Each prototype gets a milestone
1.0
Review Notes target Game Prototypes - Together we need to review the notes and see what game features we should prototype. - [ ] Identify Games (easiest - hardest) - [ ] Create a new task for each game - [ ] Draft Schemas for each game - [ ] Plan Steps/Issues for Each Game - [ ] Each prototype gets a milestone
non_defect
review notes target game prototypes together we need to review the notes and see what game features we should prototype identify games easiest hardest create a new task for each game draft schemas for each game plan steps issues for each game each prototype gets a milestone
0
379,308
11,219,741,738
IssuesEvent
2020-01-07 14:30:28
aiidateam/plumpy
https://api.github.com/repos/aiidateam/plumpy
closed
Make `breadcrumbs` integrated attribute of `PortNamespace`
priority/nice to have topic/ports type/requested feature
The original design is that a `PortNamespace` should have no knowledge whether it is part of another namespace. This led to problems when trying to raise exceptions with useful messages during validation. If a nested port did not know exactly where it is nested, it could only provide its own name in the error message. For heavily nested namespaces, however, there can be multiple ports with the same name, leaving the user guessing which one threw the error. This was solved ad-hoc by passing an `breadcrumbs` argument along during recursive port validation. However, this concept is starting to become more important in other use cases as well. The original design of not having a bidirectional dependency/knowledge within the nesting should be kept, but we can make the breadcrumbs an attribute of a port when it is created. If it just has the namespace path but no actual object references this should be fine.
1.0
Make `breadcrumbs` integrated attribute of `PortNamespace` - The original design is that a `PortNamespace` should have no knowledge whether it is part of another namespace. This led to problems when trying to raise exceptions with useful messages during validation. If a nested port did not know exactly where it is nested, it could only provide its own name in the error message. For heavily nested namespaces, however, there can be multiple ports with the same name, leaving the user guessing which one threw the error. This was solved ad-hoc by passing an `breadcrumbs` argument along during recursive port validation. However, this concept is starting to become more important in other use cases as well. The original design of not having a bidirectional dependency/knowledge within the nesting should be kept, but we can make the breadcrumbs an attribute of a port when it is created. If it just has the namespace path but no actual object references this should be fine.
non_defect
make breadcrumbs integrated attribute of portnamespace the original design is that a portnamespace should have no knowledge whether it is part of another namespace this led to problems when trying to raise exceptions with useful messages during validation if a nested port did not know exactly where it is nested it could only provide its own name in the error message for heavily nested namespaces however there can be multiple ports with the same name leaving the user guessing which one threw the error this was solved ad hoc by passing an breadcrumbs argument along during recursive port validation however this concept is starting to become more important in other use cases as well the original design of not having a bidirectional dependency knowledge within the nesting should be kept but we can make the breadcrumbs an attribute of a port when it is created if it just has the namespace path but no actual object references this should be fine
0
67,591
21,009,060,049
IssuesEvent
2022-03-30 03:26:29
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: scipy.stats.multiscale_graphcorr p-values are computed differently from literature and other packages
defect scipy.stats
### Describe your issue. This bug is as described as here: https://github.com/neurodata/hyppo/issues/124. The issue is reproduced here: p-values do not appear to be computed correctly. Literature review/review of similar independence testing code seems to suggest most people use the approach suggested in Phipson et al., 2011 described in 6.2 to always include the given ordering as a permutation. Should be updated for all permutation-based approaches to my knowledge. ### Reproducing Code Example ```python from scipy.stats import multiscale_graphcorr import numpy as np X = np.arange(0, 25) Y = np.arange(0, 25) stat, pval, _ = multiscale_graphcorr(X, Y, reps=100) print(pval) print(stat) >> .01 >> 1 ``` p-value obtained is 1/100. R example using energy package (contains other nonparametric multivariate independence tests using permutation tests), with energy version 1.7-7 and R version 4.0.2: ```R require(energy) X = 0:25; Y = 0:25; result = dcor.test(X, Y, R=100) print(result$p.value) >> .0099... print(result$stat) >> 1 ``` p-value obtained is 1/(100 + 1). We should always use the result from Phipson et. al 2011, rather than only in the case where the p-value would otherwise be 0. ### Error message ```shell There is no error message for this issue. The proposed change is very small. It would require changing this line (https://github.com/neurodata/hyppo/pull/223): # calculate p-value and significant permutation map through list pvalue = (null_dist >= stat).sum() / reps # correct for a p-value of 0. This is because, with bootstrapping # permutations, a p-value of 0 is incorrect if pvalue == 0: pvalue = 1 / reps ``` to: ``` pvalue = ((null_dist >= stat).sum() + 1) / (1 + reps) ``` ### SciPy/NumPy/Python version information ``` 1.7.0 1.21.1 sys.version_info(major=3, minor=8, micro=5, releaselevel='final', serial=0)d ```
1.0
BUG: scipy.stats.multiscale_graphcorr p-values are computed differently from literature and other packages - ### Describe your issue. This bug is as described as here: https://github.com/neurodata/hyppo/issues/124. The issue is reproduced here: p-values do not appear to be computed correctly. Literature review/review of similar independence testing code seems to suggest most people use the approach suggested in Phipson et al., 2011 described in 6.2 to always include the given ordering as a permutation. Should be updated for all permutation-based approaches to my knowledge. ### Reproducing Code Example ```python from scipy.stats import multiscale_graphcorr import numpy as np X = np.arange(0, 25) Y = np.arange(0, 25) stat, pval, _ = multiscale_graphcorr(X, Y, reps=100) print(pval) print(stat) >> .01 >> 1 ``` p-value obtained is 1/100. R example using energy package (contains other nonparametric multivariate independence tests using permutation tests), with energy version 1.7-7 and R version 4.0.2: ```R require(energy) X = 0:25; Y = 0:25; result = dcor.test(X, Y, R=100) print(result$p.value) >> .0099... print(result$stat) >> 1 ``` p-value obtained is 1/(100 + 1). We should always use the result from Phipson et. al 2011, rather than only in the case where the p-value would otherwise be 0. ### Error message ```shell There is no error message for this issue. The proposed change is very small. It would require changing this line (https://github.com/neurodata/hyppo/pull/223): # calculate p-value and significant permutation map through list pvalue = (null_dist >= stat).sum() / reps # correct for a p-value of 0. This is because, with bootstrapping # permutations, a p-value of 0 is incorrect if pvalue == 0: pvalue = 1 / reps ``` to: ``` pvalue = ((null_dist >= stat).sum() + 1) / (1 + reps) ``` ### SciPy/NumPy/Python version information ``` 1.7.0 1.21.1 sys.version_info(major=3, minor=8, micro=5, releaselevel='final', serial=0)d ```
defect
bug scipy stats multiscale graphcorr p values are computed differently from literature and other packages describe your issue this bug is as described as here the issue is reproduced here p values do not appear to be computed correctly literature review review of similar independence testing code seems to suggest most people use the approach suggested in phipson et al described in to always include the given ordering as a permutation should be updated for all permutation based approaches to my knowledge reproducing code example python from scipy stats import multiscale graphcorr import numpy as np x np arange y np arange stat pval multiscale graphcorr x y reps print pval print stat p value obtained is r example using energy package contains other nonparametric multivariate independence tests using permutation tests with energy version and r version r require energy x y result dcor test x y r print result p value print result stat p value obtained is we should always use the result from phipson et al rather than only in the case where the p value would otherwise be error message shell there is no error message for this issue the proposed change is very small it would require changing this line calculate p value and significant permutation map through list pvalue null dist stat sum reps correct for a p value of this is because with bootstrapping permutations a p value of is incorrect if pvalue pvalue reps to pvalue null dist stat sum reps scipy numpy python version information sys version info major minor micro releaselevel final serial d
1
60,023
17,023,314,365
IssuesEvent
2021-07-03 01:23:04
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
inkscape version 0.44.1 can export large png
Component: tilesathome Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 11.05am, Thursday, 23rd October 2008]** The current check in TahConf.pm just checks for versions of 0.46 vs earlier, while there are other earlier versions that can export the needed 8192x8192px png files.
1.0
inkscape version 0.44.1 can export large png - **[Submitted to the original trac issue database at 11.05am, Thursday, 23rd October 2008]** The current check in TahConf.pm just checks for versions of 0.46 vs earlier, while there are other earlier versions that can export the needed 8192x8192px png files.
defect
inkscape version can export large png the current check in tahconf pm just checks for versions of vs earlier while there are other earlier versions that can export the needed png files
1
45,694
13,023,360,559
IssuesEvent
2020-07-27 09:52:10
pymc-devs/pymc3
https://api.github.com/repos/pymc-devs/pymc3
closed
Shape argument passed to an observed variable is ignored
defects shape problem
If you specify a `shape` argument for an observed variable, this parameter is silently ignored in favor of the shape of the `observed` value. Specifying `shape` should return an error if specified and does not correspond to the passed data.
1.0
Shape argument passed to an observed variable is ignored - If you specify a `shape` argument for an observed variable, this parameter is silently ignored in favor of the shape of the `observed` value. Specifying `shape` should return an error if specified and does not correspond to the passed data.
defect
shape argument passed to an observed variable is ignored if you specify a shape argument for an observed variable this parameter is silently ignored in favor of the shape of the observed value specifying shape should return an error if specified and does not correspond to the passed data
1
627,143
19,896,937,380
IssuesEvent
2022-01-25 00:52:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.att.com - Unable to access the site
priority-important priority-normal severity-critical action-needssitepatch browser-fenix engine-gecko QA_triaged
<!-- @browser: Firefox Mobile 98.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/98232 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.att.com/ **Browser / Version**: Firefox Mobile 98.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: The page says access denied. This may require users to be in the US. Tried reference browser and release Firefox without success. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/09759a9d-26c0-4df0-aaf7-feeab98eade0.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220111093827</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/1/2e1854c7-7ff0-4ffc-b447-a2860c2fbb0c) _From [webcompat.com](https://webcompat.com/) with ❤️_
2.0
www.att.com - Unable to access the site - <!-- @browser: Firefox Mobile 98.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/98232 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.att.com/ **Browser / Version**: Firefox Mobile 98.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: The page says access denied. This may require users to be in the US. Tried reference browser and release Firefox without success. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/09759a9d-26c0-4df0-aaf7-feeab98eade0.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220111093827</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/1/2e1854c7-7ff0-4ffc-b447-a2860c2fbb0c) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
unable to access the site url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce the page says access denied this may require users to be in the us tried reference browser and release firefox without success view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
241,892
7,835,397,009
IssuesEvent
2018-06-17 05:00:29
rathena/rathena
https://api.github.com/repos/rathena/rathena
closed
Rand attack in nonfunctional refinement
component:core mode:renewal priority:low type:bug
Currently I checked the equipment refining system with random bonus ATK is not working. A weapons LV4 for example, after the fifth refining should add in a weapons attack rand (0 ~ 14), which now appears to be constant. In refine_db.txt have the correct information, but does not appear to work. In the right situation an armament for example, +12 refining could have difference in the damage other equal also +12. Link: http://irowiki.org/wiki/Refinement_System **(Effects Gained)** NOTE: The link follows a model similar to the refine_db the rAthena, but the end result is strange. There is no difference between weapons refinements.
1.0
Rand attack in nonfunctional refinement - Currently I checked the equipment refining system with random bonus ATK is not working. A weapons LV4 for example, after the fifth refining should add in a weapons attack rand (0 ~ 14), which now appears to be constant. In refine_db.txt have the correct information, but does not appear to work. In the right situation an armament for example, +12 refining could have difference in the damage other equal also +12. Link: http://irowiki.org/wiki/Refinement_System **(Effects Gained)** NOTE: The link follows a model similar to the refine_db the rAthena, but the end result is strange. There is no difference between weapons refinements.
non_defect
rand attack in nonfunctional refinement currently i checked the equipment refining system with random bonus atk is not working a weapons for example after the fifth refining should add in a weapons attack rand which now appears to be constant in refine db txt have the correct information but does not appear to work in the right situation an armament for example refining could have difference in the damage other equal also link effects gained note the link follows a model similar to the refine db the rathena but the end result is strange there is no difference between weapons refinements
0
284,112
8,735,807,612
IssuesEvent
2018-12-11 17:43:59
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
Cracks Clipper is broken in 2.x
bug crash likelihood medium priority reviewed severity high wrong results
The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 402 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Cracks Clipper is broken in 2.x Assigned to: Kathleen Biagas Category: Target version: 2.1.1 Author: Kathleen Biagas Start: 09/22/2010 Due date: % Done: 0 Estimated time: Created: 09/22/2010 12:54 pm Updated: 09/29/2010 02:53 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.0.0 Impact: Expected Use: OS: All Support Group: Any Description: The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap. Comments: Restored functionality of CracksClipper operator.SVN revisions 12594 (2.1 RC) 12596 (trunk).
1.0
Cracks Clipper is broken in 2.x - The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 402 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Cracks Clipper is broken in 2.x Assigned to: Kathleen Biagas Category: Target version: 2.1.1 Author: Kathleen Biagas Start: 09/22/2010 Due date: % Done: 0 Estimated time: Created: 09/22/2010 12:54 pm Updated: 09/29/2010 02:53 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.0.0 Impact: Expected Use: OS: All Support Group: Any Description: The CracksClipper operator no longer works, as of 2.0. Greg Burton has need of this functionality, and would like it fixed asap. Comments: Restored functionality of CracksClipper operator.SVN revisions 12594 (2.1 RC) 12596 (trunk).
non_defect
cracks clipper is broken in x the cracksclipper operator no longer works as of greg burton has need of this functionality and would like it fixed asap redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject cracks clipper is broken in x assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description the cracksclipper operator no longer works as of greg burton has need of this functionality and would like it fixed asap comments restored functionality of cracksclipper operator svn revisions rc trunk
0
17,577
3,012,747,511
IssuesEvent
2015-07-29 02:09:30
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
[CLOSED] Unchain/unpile should ge greyed out in certain circumstances
auto-migrated Priority-Low Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:20 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/25_ ---- ``` In the Edit Profile screen the unchain/unpile should be greyed out if the user does not have those priviliges and/or no tasks are selected. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 21 Jul 2008 at 8:05
1.0
[CLOSED] Unchain/unpile should ge greyed out in certain circumstances - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:20 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/25_ ---- ``` In the Edit Profile screen the unchain/unpile should be greyed out if the user does not have those priviliges and/or no tasks are selected. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 21 Jul 2008 at 8:05
defect
unchain unpile should ge greyed out in certain circumstances issue by monday jul at gmt originally opened as in the edit profile screen the unchain unpile should be greyed out if the user does not have those priviliges and or no tasks are selected original issue reported on code google com by arthurte gmail com on jul at
1
58,448
14,398,713,499
IssuesEvent
2020-12-03 09:57:29
joncampbell123/dosbox-x
https://api.github.com/repos/joncampbell123/dosbox-x
closed
flatpak build problem
build issues platform: Linux
**Describe the bug** Ok, that only took one release :-) For some reason, that I'm not sure about the flatpak build process no longer works. The build seems to get stuck in a continuous loop. It seems it successfully does the autogen.sh and configure steps, then starts the make step and shows the following: ``` ... config.status: executing depfiles commands Running: make make: Warning: File 'Makefile.am' has modification time 1275 s in the future CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing aclocal-1.16 cd . && /bin/sh /run/build/dosbox-x/missing automake-1.16 --foreign CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing autoconf /bin/sh ./config.status --recheck running CONFIG_SHELL=/bin/sh /bin/sh ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 CFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection LDFLAGS=-L/app/lib -Wl,-z,relro,-z,now -Wl,--as-needed CXXFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection --no-create --no-recursion checking build system type... x86_64-pc-linux-gnu ... ``` And the configure and make just keeps looping. **To Reproduce** Steps to reproduce the behavior: ``` git clone --recursive https://github.com/flathub/com.dosbox_x.DOSBox-X.git cd com.dosbox_x.DOSBox-X flatpak install flathub org.freedesktop.Sdk//20.08 -y flatpak-builder --force-clean --install --user -y builddir com.dosbox_x.DOSBox-X.yaml ``` **Additional context** By default the yaml file will use buildsystem=autotool and that worked with the last version. I tried switching to buildsystem=simple and specifying the build-commands, but that also gets into the same circular configure issue. e.g. ``` - name: dosbox-x buildsystem: simple build-commands: - ./autogen.sh - ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 - make - make install sources: - type: archive url: https://github.com/joncampbell123/dosbox-x/archive/dosbox-x-v0.83.7.tar.gz sha256: 9cdfa3267c340a869255d8eb1c4ebf4adde47c22854e1d013da22190350bfbb3 post-install: - install -Dm644 /app/share/icons/hicolor/scalable/apps/dosbox-x.svg /app/share/icons/hicolor/scalable/apps/${FLATPAK_ID}.svg - desktop-file-edit --set-key=Icon --set-value=${FLATPAK_ID} /app/share/applications/${FLATPAK_ID}.desktop ``` I also tried replacing the build-commands with just a simple ``./build-debug-sdl2`` and then it crashes out with a: ``` checking for a BSD-compatible install... /usr/bin/install -c checking whether build environment is sane... configure: error: newly created file is older than distributed files! Check your system clock Error: module dosbox-x: Child process exited with code 1 ``` To clarify the flatpak build runs in an environment with EPOCH set to zero. This is intentional, and I don't think you can disable it. Loop during regular build: ``` ```
1.0
flatpak build problem - **Describe the bug** Ok, that only took one release :-) For some reason, that I'm not sure about the flatpak build process no longer works. The build seems to get stuck in a continuous loop. It seems it successfully does the autogen.sh and configure steps, then starts the make step and shows the following: ``` ... config.status: executing depfiles commands Running: make make: Warning: File 'Makefile.am' has modification time 1275 s in the future CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing aclocal-1.16 cd . && /bin/sh /run/build/dosbox-x/missing automake-1.16 --foreign CDPATH="${ZSH_VERSION+.}:" && cd . && /bin/sh /run/build/dosbox-x/missing autoconf /bin/sh ./config.status --recheck running CONFIG_SHELL=/bin/sh /bin/sh ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 CFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection LDFLAGS=-L/app/lib -Wl,-z,relro,-z,now -Wl,--as-needed CXXFLAGS=-O2 -g -pipe -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fexceptions -fstack-protector-strong -grecord-gcc-switches -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection --no-create --no-recursion checking build system type... x86_64-pc-linux-gnu ... ``` And the configure and make just keeps looping. **To Reproduce** Steps to reproduce the behavior: ``` git clone --recursive https://github.com/flathub/com.dosbox_x.DOSBox-X.git cd com.dosbox_x.DOSBox-X flatpak install flathub org.freedesktop.Sdk//20.08 -y flatpak-builder --force-clean --install --user -y builddir com.dosbox_x.DOSBox-X.yaml ``` **Additional context** By default the yaml file will use buildsystem=autotool and that worked with the last version. I tried switching to buildsystem=simple and specifying the build-commands, but that also gets into the same circular configure issue. e.g. ``` - name: dosbox-x buildsystem: simple build-commands: - ./autogen.sh - ./configure --enable-core-inline --enable-debug=heavy --enable-sdl2 - make - make install sources: - type: archive url: https://github.com/joncampbell123/dosbox-x/archive/dosbox-x-v0.83.7.tar.gz sha256: 9cdfa3267c340a869255d8eb1c4ebf4adde47c22854e1d013da22190350bfbb3 post-install: - install -Dm644 /app/share/icons/hicolor/scalable/apps/dosbox-x.svg /app/share/icons/hicolor/scalable/apps/${FLATPAK_ID}.svg - desktop-file-edit --set-key=Icon --set-value=${FLATPAK_ID} /app/share/applications/${FLATPAK_ID}.desktop ``` I also tried replacing the build-commands with just a simple ``./build-debug-sdl2`` and then it crashes out with a: ``` checking for a BSD-compatible install... /usr/bin/install -c checking whether build environment is sane... configure: error: newly created file is older than distributed files! Check your system clock Error: module dosbox-x: Child process exited with code 1 ``` To clarify the flatpak build runs in an environment with EPOCH set to zero. This is intentional, and I don't think you can disable it. Loop during regular build: ``` ```
non_defect
flatpak build problem describe the bug ok that only took one release for some reason that i m not sure about the flatpak build process no longer works the build seems to get stuck in a continuous loop it seems it successfully does the autogen sh and configure steps then starts the make step and shows the following config status executing depfiles commands running make make warning file makefile am has modification time s in the future cdpath zsh version cd bin sh run build dosbox x missing aclocal cd bin sh run build dosbox x missing automake foreign cdpath zsh version cd bin sh run build dosbox x missing autoconf bin sh config status recheck running config shell bin sh bin sh configure enable core inline enable debug heavy enable cflags g pipe wp d fortify source wp d glibcxx assertions fexceptions fstack protector strong grecord gcc switches fasynchronous unwind tables fstack clash protection fcf protection ldflags l app lib wl z relro z now wl as needed cxxflags g pipe wp d fortify source wp d glibcxx assertions fexceptions fstack protector strong grecord gcc switches fasynchronous unwind tables fstack clash protection fcf protection no create no recursion checking build system type pc linux gnu and the configure and make just keeps looping to reproduce steps to reproduce the behavior git clone recursive cd com dosbox x dosbox x flatpak install flathub org freedesktop sdk y flatpak builder force clean install user y builddir com dosbox x dosbox x yaml additional context by default the yaml file will use buildsystem autotool and that worked with the last version i tried switching to buildsystem simple and specifying the build commands but that also gets into the same circular configure issue e g name dosbox x buildsystem simple build commands autogen sh configure enable core inline enable debug heavy enable make make install sources type archive url post install install app share icons hicolor scalable apps dosbox x svg app share icons hicolor scalable apps flatpak id svg desktop file edit set key icon set value flatpak id app share applications flatpak id desktop i also tried replacing the build commands with just a simple build debug and then it crashes out with a checking for a bsd compatible install usr bin install c checking whether build environment is sane configure error newly created file is older than distributed files check your system clock error module dosbox x child process exited with code to clarify the flatpak build runs in an environment with epoch set to zero this is intentional and i don t think you can disable it loop during regular build
0
54,826
23,341,619,081
IssuesEvent
2022-08-09 14:27:04
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Ability to create and use alias for index pattern
enhancement loe:hours Feature:Data Views Team:AppServicesSv impact:low
**Describe the feature:** Would like to have the ability to add a meaningful displayname for an indexpattern. This name should be shown in the locations where a user can choose an indexpattern for his/her search. Not all users are tech savvy and indexnames often are quite technical. For those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use. **Describe a specific use case for the feature:** On the discover tab there is this dropdown to choose an index pattern, right below the 'add filter' link. I would like to see different names here, so instead of lnd-clust:idx-ms3-sw-ds,lp-clust:idx-ms3-sw-ds and lnd-clust::idx-ms3-sw-ds , Users could see the names London-and-liverpool-subway-stats and London-subway-stats Above is just an example, but it would be helpful, especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences. I can image this is just an alias one could set when creating the index pattern and if it is not set, the indexpattern itselves is shown like it is now.
1.0
Ability to create and use alias for index pattern - **Describe the feature:** Would like to have the ability to add a meaningful displayname for an indexpattern. This name should be shown in the locations where a user can choose an indexpattern for his/her search. Not all users are tech savvy and indexnames often are quite technical. For those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use. **Describe a specific use case for the feature:** On the discover tab there is this dropdown to choose an index pattern, right below the 'add filter' link. I would like to see different names here, so instead of lnd-clust:idx-ms3-sw-ds,lp-clust:idx-ms3-sw-ds and lnd-clust::idx-ms3-sw-ds , Users could see the names London-and-liverpool-subway-stats and London-subway-stats Above is just an example, but it would be helpful, especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences. I can image this is just an alias one could set when creating the index pattern and if it is not set, the indexpattern itselves is shown like it is now.
non_defect
ability to create and use alias for index pattern describe the feature would like to have the ability to add a meaningful displayname for an indexpattern this name should be shown in the locations where a user can choose an indexpattern for his her search not all users are tech savvy and indexnames often are quite technical for those whore are not a more human readable name without wildcards and abbreviations can be quite helpful in quickly deciding which index to use describe a specific use case for the feature on the discover tab there is this dropdown to choose an index pattern right below the add filter link i would like to see different names here so instead of lnd clust idx sw ds lp clust idx sw ds and lnd clust idx sw ds users could see the names london and liverpool subway stats and london subway stats above is just an example but it would be helpful especially when using remoteclusters together with localclusters and the dropdown is filled with indices with the same names with slight differences i can image this is just an alias one could set when creating the index pattern and if it is not set the indexpattern itselves is shown like it is now
0
71,708
18,848,499,757
IssuesEvent
2021-11-11 17:36:07
spack/spack
https://api.github.com/repos/spack/spack
closed
Fenics fails to run after installation with spack
build-error
Fenics fails to run after installation with spack ### Expected Result To be able to execute simple fenics scripts using the spack-built fenics. ### Actual Result Simply importing the fenics module into the spack-provided python fails with: ```console ImportError: No module named _common ``` ### Steps to reproduce the issue ``` bridges@wheeler-sn[500]> module load fenics-2017.1.0-intel-17.0.4-impi-mkl-python2-zre3f5b Autoloading eigen-3.3.3-intel-17.0.4-impi-mkl-lnbosvu Autoloading metis-5.1.0-intel-17.0.4-xyonw4o Autoloading fftw-3.3.6-pl2-intel-17.0.4-impi-bwlmvhs Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc Autoloading intel-mkl-17.0.4-intel-17.0.4-z2nms2r Autoloading gmp-6.1.2-intel-17.0.4-asgihpm Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk Autoloading mpfr-3.1.5-intel-17.0.4-e2wn26r Autoloading py-mpmath-0.19-intel-17.0.4-python2-bub4nhr Autoloading python-2.7.13-intel-17.0.4-ufhcwdr Autoloading bzip2-1.0.6-intel-17.0.4-zjtu25w Autoloading ncurses-6.0-intel-17.0.4-j3ftwhg Autoloading openssl-1.0.2k-intel-17.0.4-6n2jfwi Autoloading sqlite-3.18.0-intel-17.0.4-cyhl22m Autoloading readline-7.0-intel-17.0.4-psasazv Autoloading readline-7.0-intel-17.0.4-psasazv Autoloading py-ply-3.8-intel-17.0.4-python2-ntmnh7h Autoloading py-instant-2017.1.0-intel-17.0.4-mkl-python2-ebd5pyo Autoloading swig-3.0.12-intel-17.0.4-bbmqam2 Autoloading pcre-8.40-intel-17.0.4-72ed5lu Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx Autoloading py-appdirs-1.4.3-intel-17.0.4-python2-mupwe3g Autoloading py-packaging-16.8-intel-17.0.4-python2-js2s5wi Autoloading py-pyparsing-2.2.0-intel-17.0.4-python2-vh7hq5e Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl Autoloading py-slepc4py-3.7.0-intel-17.0.4-impi-mkl-python2-iahwthv Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6 Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-3x6zx5l Autoloading hypre-2.11.2-intel-17.0.4-impi-mkl-g5srp2y Autoloading superlu-dist-5.1.3-intel-17.0.4-impi-mkl-ymsttjs Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2 Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading py-mpi4py-2.0.0-intel-17.0.4-impi-python2-ozw6g4g Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-fy5wf4h Autoloading arpack-ng-3.5.0-intel-17.0.4-impi-mkl-l7ewls2 Autoloading py-ufl-2017.1.0-intel-17.0.4-mkl-python2-2xalq65 Autoloading cmake-3.8.1-intel-17.0.4-m63t5kd Autoloading py-sympy-1.0-intel-17.0.4-python2-tap7mf2 Autoloading py-dijitso-2017.1.0-intel-17.0.4-mkl-python2-ljegoot Autoloading boost-1.63.0-intel-17.0.4-bjbzbtc Autoloading trilinos-12.10.1-intel-17.0.4-impi-mkl-l67gizf Autoloading glm-0.9.7.1-intel-17.0.4-wmre6yo Autoloading netcdf-4.4.1.1-intel-17.0.4-impi-bdktit6 Autoloading mumps-5.1.1-intel-17.0.4-impi-mkl-6enszkq Autoloading matio-1.5.9-intel-17.0.4-impi-sjmlxnf Autoloading python-2.7.13-intel-17.0.4-ufhcwdr Autoloading swig-3.0.12-intel-17.0.4-bbmqam2 Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2 Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-python2-fy5wf4h Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-python2-3x6zx5l Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6 Autoloading py-fiat-2017.1.0-intel-17.0.4-mkl-python2-6krzj5q Autoloading py-ffc-2017.1.0-intel-17.0.4-mkl-python2-mleunix Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp bridges@wheeler-sn[501]> which python python is /opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/bin/python bridges@wheeler-sn[502]> python -c 'import fenics' Traceback (most recent call last): File "<string>", line 1, in <module> File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/fenics/__init__.py", line 7, in <module> from dolfin import * File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/__init__.py", line 17, in <module> from . import cpp File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/__init__.py", line 43, in <module> exec("from . import %s" % module_name) File "<string>", line 1, in <module> File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 21, in <module> _common = swig_import_helper() File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 20, in swig_import_helper return importlib.import_module('_common') File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) ImportError: No module named _common bridges@wheeler-sn[503]> ``` ### Information on your system This includes: 1. OpenHPC Running on CentOS 7.3
1.0
Fenics fails to run after installation with spack - Fenics fails to run after installation with spack ### Expected Result To be able to execute simple fenics scripts using the spack-built fenics. ### Actual Result Simply importing the fenics module into the spack-provided python fails with: ```console ImportError: No module named _common ``` ### Steps to reproduce the issue ``` bridges@wheeler-sn[500]> module load fenics-2017.1.0-intel-17.0.4-impi-mkl-python2-zre3f5b Autoloading eigen-3.3.3-intel-17.0.4-impi-mkl-lnbosvu Autoloading metis-5.1.0-intel-17.0.4-xyonw4o Autoloading fftw-3.3.6-pl2-intel-17.0.4-impi-bwlmvhs Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc Autoloading intel-mkl-17.0.4-intel-17.0.4-z2nms2r Autoloading gmp-6.1.2-intel-17.0.4-asgihpm Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk Autoloading mpfr-3.1.5-intel-17.0.4-e2wn26r Autoloading py-mpmath-0.19-intel-17.0.4-python2-bub4nhr Autoloading python-2.7.13-intel-17.0.4-ufhcwdr Autoloading bzip2-1.0.6-intel-17.0.4-zjtu25w Autoloading ncurses-6.0-intel-17.0.4-j3ftwhg Autoloading openssl-1.0.2k-intel-17.0.4-6n2jfwi Autoloading sqlite-3.18.0-intel-17.0.4-cyhl22m Autoloading readline-7.0-intel-17.0.4-psasazv Autoloading readline-7.0-intel-17.0.4-psasazv Autoloading py-ply-3.8-intel-17.0.4-python2-ntmnh7h Autoloading py-instant-2017.1.0-intel-17.0.4-mkl-python2-ebd5pyo Autoloading swig-3.0.12-intel-17.0.4-bbmqam2 Autoloading pcre-8.40-intel-17.0.4-72ed5lu Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx Autoloading py-appdirs-1.4.3-intel-17.0.4-python2-mupwe3g Autoloading py-packaging-16.8-intel-17.0.4-python2-js2s5wi Autoloading py-pyparsing-2.2.0-intel-17.0.4-python2-vh7hq5e Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl Autoloading py-subprocess32-3.2.7-intel-17.0.4-python2-5aviwgl Autoloading py-slepc4py-3.7.0-intel-17.0.4-impi-mkl-python2-iahwthv Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6 Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-3x6zx5l Autoloading hypre-2.11.2-intel-17.0.4-impi-mkl-g5srp2y Autoloading superlu-dist-5.1.3-intel-17.0.4-impi-mkl-ymsttjs Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2 Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading py-mpi4py-2.0.0-intel-17.0.4-impi-python2-ozw6g4g Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-fy5wf4h Autoloading arpack-ng-3.5.0-intel-17.0.4-impi-mkl-l7ewls2 Autoloading py-ufl-2017.1.0-intel-17.0.4-mkl-python2-2xalq65 Autoloading cmake-3.8.1-intel-17.0.4-m63t5kd Autoloading py-sympy-1.0-intel-17.0.4-python2-tap7mf2 Autoloading py-dijitso-2017.1.0-intel-17.0.4-mkl-python2-ljegoot Autoloading boost-1.63.0-intel-17.0.4-bjbzbtc Autoloading trilinos-12.10.1-intel-17.0.4-impi-mkl-l67gizf Autoloading glm-0.9.7.1-intel-17.0.4-wmre6yo Autoloading netcdf-4.4.1.1-intel-17.0.4-impi-bdktit6 Autoloading mumps-5.1.1-intel-17.0.4-impi-mkl-6enszkq Autoloading matio-1.5.9-intel-17.0.4-impi-sjmlxnf Autoloading python-2.7.13-intel-17.0.4-ufhcwdr Autoloading swig-3.0.12-intel-17.0.4-bbmqam2 Autoloading py-six-1.10.0-intel-17.0.4-python2-mj2hiwx Autoloading suite-sparse-4.5.5-intel-17.0.4-mkl-v6mhhmc Autoloading py-setuptools-35.0.2-intel-17.0.4-python2-jg3axmx Autoloading parmetis-4.0.3-intel-17.0.4-impi-3bbdt5v Autoloading hdf5-1.10.1-intel-17.0.4-impi-63jbam2 Autoloading slepc-3.7.4-intel-17.0.4-impi-mkl-python2-fy5wf4h Autoloading py-numpy-1.13.0-intel-17.0.4-mkl-python2-bby27xc Autoloading petsc-3.7.6-intel-17.0.4-impi-mkl-python2-3x6zx5l Autoloading py-petsc4py-3.7.0-intel-17.0.4-impi-mkl-python2-y6jgyt6 Autoloading py-fiat-2017.1.0-intel-17.0.4-mkl-python2-6krzj5q Autoloading py-ffc-2017.1.0-intel-17.0.4-mkl-python2-mleunix Autoloading scotch-6.0.4-intel-17.0.4-impi-dxrzrfk Autoloading intel-mpi-2017.4-intel-17.0.4-ataurnp bridges@wheeler-sn[501]> which python python is /opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/bin/python bridges@wheeler-sn[502]> python -c 'import fenics' Traceback (most recent call last): File "<string>", line 1, in <module> File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/fenics/__init__.py", line 7, in <module> from dolfin import * File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/__init__.py", line 17, in <module> from . import cpp File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/__init__.py", line 43, in <module> exec("from . import %s" % module_name) File "<string>", line 1, in <module> File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 21, in <module> _common = swig_import_helper() File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/fenics-2017.1.0-zre3f5b4xo4u2z45jjnoeqvzzfv4l35m/lib/python2.7/site-packages/dolfin/cpp/common.py", line 20, in swig_import_helper return importlib.import_module('_common') File "/opt/spack/opt/spack/linux-centos7-x86_64/intel-17.0.4/python-2.7.13-ufhcwdrlqo7lt2i2otrgvmgh5f4jrzyt/lib/python2.7/importlib/__init__.py", line 37, in import_module __import__(name) ImportError: No module named _common bridges@wheeler-sn[503]> ``` ### Information on your system This includes: 1. OpenHPC Running on CentOS 7.3
non_defect
fenics fails to run after installation with spack fenics fails to run after installation with spack expected result to be able to execute simple fenics scripts using the spack built fenics actual result simply importing the fenics module into the spack provided python fails with console importerror no module named common steps to reproduce the issue bridges wheeler sn module load fenics intel impi mkl autoloading eigen intel impi mkl lnbosvu autoloading metis intel autoloading fftw intel impi bwlmvhs autoloading intel mpi intel ataurnp autoloading suite sparse intel mkl autoloading intel mkl intel autoloading gmp intel asgihpm autoloading scotch intel impi dxrzrfk autoloading mpfr intel autoloading py mpmath intel autoloading python intel ufhcwdr autoloading intel autoloading ncurses intel autoloading openssl intel autoloading sqlite intel autoloading readline intel psasazv autoloading readline intel psasazv autoloading py ply intel autoloading py instant intel mkl autoloading swig intel autoloading pcre intel autoloading py setuptools intel autoloading py six intel autoloading py appdirs intel autoloading py packaging intel autoloading py pyparsing intel autoloading py numpy intel mkl autoloading py intel autoloading py intel autoloading py intel impi mkl iahwthv autoloading py intel impi mkl autoloading petsc intel impi mkl autoloading hypre intel impi mkl autoloading superlu dist intel impi mkl ymsttjs autoloading parmetis intel impi autoloading intel impi autoloading parmetis intel impi autoloading py intel impi autoloading slepc intel impi mkl autoloading arpack ng intel impi mkl autoloading py ufl intel mkl autoloading cmake intel autoloading py sympy intel autoloading py dijitso intel mkl ljegoot autoloading boost intel bjbzbtc autoloading trilinos intel impi mkl autoloading glm intel autoloading netcdf intel impi autoloading mumps intel impi mkl autoloading matio intel impi sjmlxnf autoloading python intel ufhcwdr autoloading swig intel autoloading py six intel autoloading suite sparse intel mkl autoloading py setuptools intel autoloading parmetis intel impi autoloading intel impi autoloading slepc intel impi mkl autoloading py numpy intel mkl autoloading petsc intel impi mkl autoloading py intel impi mkl autoloading py fiat intel mkl autoloading py ffc intel mkl mleunix autoloading scotch intel impi dxrzrfk autoloading intel mpi intel ataurnp bridges wheeler sn which python python is opt spack opt spack linux intel python bin python bridges wheeler sn python c import fenics traceback most recent call last file line in file opt spack opt spack linux intel fenics lib site packages fenics init py line in from dolfin import file opt spack opt spack linux intel fenics lib site packages dolfin init py line in from import cpp file opt spack opt spack linux intel fenics lib site packages dolfin cpp init py line in exec from import s module name file line in file opt spack opt spack linux intel fenics lib site packages dolfin cpp common py line in common swig import helper file opt spack opt spack linux intel fenics lib site packages dolfin cpp common py line in swig import helper return importlib import module common file opt spack opt spack linux intel python lib importlib init py line in import module import name importerror no module named common bridges wheeler sn information on your system this includes openhpc running on centos
0
72,769
24,280,758,850
IssuesEvent
2022-09-28 17:10:51
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
PARA LO DE PEPSI, NO ESTÁ RESPETANDO EL USO DE CFDI ELEGIDO
bug resolved defect
Te mandé un correo donde el uso de CFDI no está coincidiendo el uso de CFDI en el PDF con el XML. Ellos eligen I08 y lo incluye en el PDF, pero en el XML lo pone distinto.
1.0
PARA LO DE PEPSI, NO ESTÁ RESPETANDO EL USO DE CFDI ELEGIDO - Te mandé un correo donde el uso de CFDI no está coincidiendo el uso de CFDI en el PDF con el XML. Ellos eligen I08 y lo incluye en el PDF, pero en el XML lo pone distinto.
defect
para lo de pepsi no está respetando el uso de cfdi elegido te mandé un correo donde el uso de cfdi no está coincidiendo el uso de cfdi en el pdf con el xml ellos eligen y lo incluye en el pdf pero en el xml lo pone distinto
1
399,064
11,742,661,888
IssuesEvent
2020-03-12 01:37:31
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
Revisit urls schemes
priority: low refactor technical change
In GitLab by @se-bastiaan on Apr 28, 2019, 21:30 ### One-sentence description Revisit urls schemes ### Why? It's currently a bit messy and not consistent. ### Current implementation We have one central urls.py for our application in thaliawebsite which decides the namespacing of all the modules. ### Suggested implementation The central urls.py only includes the urls.py files of all the packages. These packages are then allowed to fill in their url schemes how they see fit. This allows us to re-use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page. All urls used in the django admin and implemented via the central urls.py (like push notifications and newsletters) should be implemented using the ModelAdmin's `get_urls` method like it was done in registrations and payments.
1.0
Revisit urls schemes - In GitLab by @se-bastiaan on Apr 28, 2019, 21:30 ### One-sentence description Revisit urls schemes ### Why? It's currently a bit messy and not consistent. ### Current implementation We have one central urls.py for our application in thaliawebsite which decides the namespacing of all the modules. ### Suggested implementation The central urls.py only includes the urls.py files of all the packages. These packages are then allowed to fill in their url schemes how they see fit. This allows us to re-use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page. All urls used in the django admin and implemented via the central urls.py (like push notifications and newsletters) should be implemented using the ModelAdmin's `get_urls` method like it was done in registrations and payments.
non_defect
revisit urls schemes in gitlab by se bastiaan on apr one sentence description revisit urls schemes why it s currently a bit messy and not consistent current implementation we have one central urls py for our application in thaliawebsite which decides the namespacing of all the modules suggested implementation the central urls py only includes the urls py files of all the packages these packages are then allowed to fill in their url schemes how they see fit this allows us to re use the same namespace in multiple modules and follow the way our menu is structured to decide the full url of a page all urls used in the django admin and implemented via the central urls py like push notifications and newsletters should be implemented using the modeladmin s get urls method like it was done in registrations and payments
0
61,952
17,023,817,712
IssuesEvent
2021-07-03 04:00:52
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Query like "75 Gegužių g., Šiauliai" does not find address
Component: nominatim Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 8.55am, Thursday, 23rd August 2012]** This building can be found at http://www.openstreetmap.org/edit?lat=55.91334&lon=23.25451&zoom=16 , but nominatim query results show only streets, most of which are located outside the city.
1.0
Query like "75 Gegužių g., Šiauliai" does not find address - **[Submitted to the original trac issue database at 8.55am, Thursday, 23rd August 2012]** This building can be found at http://www.openstreetmap.org/edit?lat=55.91334&lon=23.25451&zoom=16 , but nominatim query results show only streets, most of which are located outside the city.
defect
query like geguå¾iå³ g å iauliai does not find address this building can be found at but nominatim query results show only streets most of which are located outside the city
1
333,089
29,508,088,184
IssuesEvent
2023-06-03 15:13:09
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix decompositions.test_numpy_qr
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix decompositions.test_numpy_qr - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5101840802/jobs/9171003812" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
non_defect
fix decompositions test numpy qr tensorflow img src torch img src numpy img src jax img src paddle img src
0
24,270
3,947,090,343
IssuesEvent
2016-04-28 08:28:03
buildo/github-workflow-pal
https://api.github.com/repos/buildo/github-workflow-pal
opened
[subissue] should filter out "customers" label
defect
## description "customers" label should be filtered out like "macro" to avoid adding cards to prisma by mistake ![image](https://cloud.githubusercontent.com/assets/4029499/14879843/e1b5dd56-0d2b-11e6-9c47-1a1a7d4d97e5.png) ## how to reproduce - add subissue to customer issue ## specs filter out "customers" too here: https://github.com/buildo/github-workflow-pal/blob/master/src/plugin/addSubIssueButton.js#L32
1.0
[subissue] should filter out "customers" label - ## description "customers" label should be filtered out like "macro" to avoid adding cards to prisma by mistake ![image](https://cloud.githubusercontent.com/assets/4029499/14879843/e1b5dd56-0d2b-11e6-9c47-1a1a7d4d97e5.png) ## how to reproduce - add subissue to customer issue ## specs filter out "customers" too here: https://github.com/buildo/github-workflow-pal/blob/master/src/plugin/addSubIssueButton.js#L32
defect
should filter out customers label description customers label should be filtered out like macro to avoid adding cards to prisma by mistake how to reproduce add subissue to customer issue specs filter out customers too here
1
140,173
11,303,578,150
IssuesEvent
2020-01-17 20:28:50
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898
CWE-134 bug flawfinder flawfinder_level_4 static analysis testing triage
```yaml { "severity": { "severity-description": "", "cvss-score": 0, "rvss-vector": "", "cvss-vector": "", "rvss-score": 0 }, "id": 1, "mitigation": { "description": "Use a constant for the format specification", "date-mitigation": "", "pull-request": "" }, "flaw": { "date-detected": "2020-01-17 (20:28)", "package": "N/A", "issue": "", "detected-by": "Alias Robotics", "reported-by-relationship": "automatic", "detected-by-method": "testing static", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/407292189/artifacts/download", "date-reported": "2020-01-17 (20:28)", "reproduction": "See artifacts below (if available)", "specificity": "subject-specific", "subsystem": "N/A", "languages": "None", "trace": "(context) int vsnprintf(char *s, size_t sz, const char *fmt, va_list args)", "reproducibility": "always", "application": "N/A", "reported-by": "Alias Robotics", "phase": "testing", "architectural-location": "application-specific" }, "vendor": null, "cve": "None", "system": "./install/motoman_driver/share/motoman_driver/MotoPlus/Controller.c:898:5", "links": "", "keywords": [ "flawfinder", "flawfinder_level_4", "static analysis", "testing", "triage", "CWE-134", "bug" ], "description": "If format strings can be influenced by an attacker, they can be exploited, and note that sprintf variations do not always \\\\0-terminate (CWE-134). Use a constant for the format specification. . Happening @ ...ver/MotoPlus/Controller.c:898", "title": "CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898", "exploitation": { "description": "", "exploitation-vector": "", "exploitation-image": "" }, "cwe": [ "CWE-134" ], "type": "bug" } ```
1.0
CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898 - ```yaml { "severity": { "severity-description": "", "cvss-score": 0, "rvss-vector": "", "cvss-vector": "", "rvss-score": 0 }, "id": 1, "mitigation": { "description": "Use a constant for the format specification", "date-mitigation": "", "pull-request": "" }, "flaw": { "date-detected": "2020-01-17 (20:28)", "package": "N/A", "issue": "", "detected-by": "Alias Robotics", "reported-by-relationship": "automatic", "detected-by-method": "testing static", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/407292189/artifacts/download", "date-reported": "2020-01-17 (20:28)", "reproduction": "See artifacts below (if available)", "specificity": "subject-specific", "subsystem": "N/A", "languages": "None", "trace": "(context) int vsnprintf(char *s, size_t sz, const char *fmt, va_list args)", "reproducibility": "always", "application": "N/A", "reported-by": "Alias Robotics", "phase": "testing", "architectural-location": "application-specific" }, "vendor": null, "cve": "None", "system": "./install/motoman_driver/share/motoman_driver/MotoPlus/Controller.c:898:5", "links": "", "keywords": [ "flawfinder", "flawfinder_level_4", "static analysis", "testing", "triage", "CWE-134", "bug" ], "description": "If format strings can be influenced by an attacker, they can be exploited, and note that sprintf variations do not always \\\\0-terminate (CWE-134). Use a constant for the format specification. . Happening @ ...ver/MotoPlus/Controller.c:898", "title": "CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ ver/MotoPlus/Controller.c:898", "exploitation": { "description": "", "exploitation-vector": "", "exploitation-image": "" }, "cwe": [ "CWE-134" ], "type": "bug" } ```
non_defect
cwe format if format strings can be influenced by an attacker they can be exploi ver motoplus controller c yaml severity severity description cvss score rvss vector cvss vector rvss score id mitigation description use a constant for the format specification date mitigation pull request flaw date detected package n a issue detected by alias robotics reported by relationship automatic detected by method testing static reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros industrial jobs artifacts download date reported reproduction see artifacts below if available specificity subject specific subsystem n a languages none trace context int vsnprintf char s size t sz const char fmt va list args reproducibility always application n a reported by alias robotics phase testing architectural location application specific vendor null cve none system install motoman driver share motoman driver motoplus controller c links keywords flawfinder flawfinder level static analysis testing triage cwe bug description if format strings can be influenced by an attacker they can be exploited and note that sprintf variations do not always terminate cwe use a constant for the format specification happening ver motoplus controller c title cwe format if format strings can be influenced by an attacker they can be exploi ver motoplus controller c exploitation description exploitation vector exploitation image cwe cwe type bug
0
52,696
13,224,940,356
IssuesEvent
2020-08-17 20:09:41
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
ACER Failed To Converge for all Events (Trac #170)
Migrated from Trac combo reconstruction defect
AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/170">https://code.icecube.wisc.edu/projects/icecube/ticket/170</a>, reported by pankninand owned by panknin</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:08:15", "_ts": "1547240895681654", "description": "AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.", "reporter": "panknin", "cc": "", "resolution": "fixed", "time": "2009-10-27T10:10:23", "component": "combo reconstruction", "summary": "ACER Failed To Converge for all Events", "priority": "critical", "keywords": "", "milestone": "", "owner": "panknin", "type": "defect" } ``` </p> </details>
1.0
ACER Failed To Converge for all Events (Trac #170) - AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/170">https://code.icecube.wisc.edu/projects/icecube/ticket/170</a>, reported by pankninand owned by panknin</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:08:15", "_ts": "1547240895681654", "description": "AtmCscdEneryReco fails on all event, even though they passed in earlier versions. I likey induced some bug, if not it has to be documented.", "reporter": "panknin", "cc": "", "resolution": "fixed", "time": "2009-10-27T10:10:23", "component": "combo reconstruction", "summary": "ACER Failed To Converge for all Events", "priority": "critical", "keywords": "", "milestone": "", "owner": "panknin", "type": "defect" } ``` </p> </details>
defect
acer failed to converge for all events trac atmcscdeneryreco fails on all event even though they passed in earlier versions i likey induced some bug if not it has to be documented migrated from json status closed changetime ts description atmcscdeneryreco fails on all event even though they passed in earlier versions i likey induced some bug if not it has to be documented reporter panknin cc resolution fixed time component combo reconstruction summary acer failed to converge for all events priority critical keywords milestone owner panknin type defect
1
52,379
13,224,707,379
IssuesEvent
2020-08-17 19:40:59
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
Inconsistent segment parameter requirements (Trac #2135)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2135">https://code.icecube.wisc.edu/projects/icecube/ticket/2135</a>, reported by cweaverand owned by cweaver</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:03:41", "_ts": "1547240621058514", "description": "`I3Tray.AddSegment` assumes that segments take two parameters (the tray and a name), but `icetray.traysegment` enforces only that they must have one parameter (the tray):\n\n{{{\n@icetray.traysegment\ndef mySegment(tray):\n\tpass\ntray=I3Tray()\ntray.AddSegment(mySegment)\n}}}\n\nresults in:\n\n{{{\n ...\n File \"/Users/cweaver/Work/IceCube/combo/build/lib/I3Tray.py\", line 224, in AddSegment\n return _segment(self, _name, **kwargs)\nTypeError: mySegment() takes exactly 1 argument (2 given)\n}}}\n\nThis is not caught by [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/python/traysegment.py#L23 the relevant check in icetray.traysegment]. ", "reporter": "cweaver", "cc": "", "resolution": "fixed", "time": "2018-02-12T23:26:32", "component": "combo core", "summary": "Inconsistent segment parameter requirements", "priority": "normal", "keywords": "", "milestone": "", "owner": "cweaver", "type": "defect" } ``` </p> </details>
1.0
Inconsistent segment parameter requirements (Trac #2135) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2135">https://code.icecube.wisc.edu/projects/icecube/ticket/2135</a>, reported by cweaverand owned by cweaver</em></summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:03:41", "_ts": "1547240621058514", "description": "`I3Tray.AddSegment` assumes that segments take two parameters (the tray and a name), but `icetray.traysegment` enforces only that they must have one parameter (the tray):\n\n{{{\n@icetray.traysegment\ndef mySegment(tray):\n\tpass\ntray=I3Tray()\ntray.AddSegment(mySegment)\n}}}\n\nresults in:\n\n{{{\n ...\n File \"/Users/cweaver/Work/IceCube/combo/build/lib/I3Tray.py\", line 224, in AddSegment\n return _segment(self, _name, **kwargs)\nTypeError: mySegment() takes exactly 1 argument (2 given)\n}}}\n\nThis is not caught by [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/python/traysegment.py#L23 the relevant check in icetray.traysegment]. ", "reporter": "cweaver", "cc": "", "resolution": "fixed", "time": "2018-02-12T23:26:32", "component": "combo core", "summary": "Inconsistent segment parameter requirements", "priority": "normal", "keywords": "", "milestone": "", "owner": "cweaver", "type": "defect" } ``` </p> </details>
defect
inconsistent segment parameter requirements trac migrated from json status closed changetime ts description addsegment assumes that segments take two parameters the tray and a name but icetray traysegment enforces only that they must have one parameter the tray n n n icetray traysegment ndef mysegment tray n tpass ntray ntray addsegment mysegment n n nresults in n n n n file users cweaver work icecube combo build lib py line in addsegment n return segment self name kwargs ntypeerror mysegment takes exactly argument given n n nthis is not caught by reporter cweaver cc resolution fixed time component combo core summary inconsistent segment parameter requirements priority normal keywords milestone owner cweaver type defect
1
199,297
6,987,939,885
IssuesEvent
2017-12-14 10:59:15
canonical-websites/www.ubuntu.com
https://api.github.com/repos/canonical-websites/www.ubuntu.com
opened
All vertical pages should have a condensed version of their url
Priority: Medium
All vertical pages should have a condensed version of their url. For example, https://www.ubuntu.com/digital-signage should reach https://www.ubuntu.com/internet-of-things/digital-signage. Same for all the other vertical pages - eg; /financial-services, /telecommunications, /security, /digital-signage, /robotics, /gateways, /digital-signage, /education, /government, /enterprise --- *Reported from: https://www.ubuntu.com/digital-signage*
1.0
All vertical pages should have a condensed version of their url - All vertical pages should have a condensed version of their url. For example, https://www.ubuntu.com/digital-signage should reach https://www.ubuntu.com/internet-of-things/digital-signage. Same for all the other vertical pages - eg; /financial-services, /telecommunications, /security, /digital-signage, /robotics, /gateways, /digital-signage, /education, /government, /enterprise --- *Reported from: https://www.ubuntu.com/digital-signage*
non_defect
all vertical pages should have a condensed version of their url all vertical pages should have a condensed version of their url for example should reach same for all the other vertical pages eg financial services telecommunications security digital signage robotics gateways digital signage education government enterprise reported from
0
377,100
26,232,288,761
IssuesEvent
2023-01-05 02:04:09
firebase/flutterfire
https://api.github.com/repos/firebase/flutterfire
closed
[📚] Cloud Messaging - Notifications: More Android code sample broken
type: documentation good first issue blocked: customer-response Stale in triage
Please see #10128 for a related issue. The ["Foreground Notifications" section](https://firebase.flutter.dev/docs/messaging/notifications#foreground-notifications) shows another Android code sample: ```dart FirebaseMessaging.onMessage.listen((RemoteMessage message) { RemoteNotification notification = message.notification; AndroidNotification android = message.notification?.android; // If `onMessage` is triggered with a notification, construct our own // local notification to show to users using the created channel. if (notification != null && android != null) { flutterLocalNotificationsPlugin.show( notification.hashCode, notification.title, notification.body, NotificationDetails( android: AndroidNotificationDetails( channel.id, channel.name, channel.description, icon: android?.smallIcon, // other properties... ), )); } }); ``` With `flutter_local_notifications: 12.0.4`, this doesn't compile anymore: - `notification` and `android` are nullable. - The channel description is now a named parameter. ![image](https://user-images.githubusercontent.com/1061209/208081728-ad88bbf5-b84f-4ceb-b71d-751423acc1a5.png) Here's a version that does compile: ```dart FirebaseMessaging.onMessage.listen((RemoteMessage message) { RemoteNotification? notification = message.notification; AndroidNotification? android = message.notification?.android; // If `onMessage` is triggered with a notification, construct our own // local notification to show to users using the created channel. if (notification != null && android != null) { flutterLocalNotificationsPlugin.show( notification.hashCode, notification.title, notification.body, NotificationDetails( android: AndroidNotificationDetails( channel.id, channel.name, channelDescription: channel.description, icon: android.smallIcon, // other properties... ), )); } }); ```
1.0
[📚] Cloud Messaging - Notifications: More Android code sample broken - Please see #10128 for a related issue. The ["Foreground Notifications" section](https://firebase.flutter.dev/docs/messaging/notifications#foreground-notifications) shows another Android code sample: ```dart FirebaseMessaging.onMessage.listen((RemoteMessage message) { RemoteNotification notification = message.notification; AndroidNotification android = message.notification?.android; // If `onMessage` is triggered with a notification, construct our own // local notification to show to users using the created channel. if (notification != null && android != null) { flutterLocalNotificationsPlugin.show( notification.hashCode, notification.title, notification.body, NotificationDetails( android: AndroidNotificationDetails( channel.id, channel.name, channel.description, icon: android?.smallIcon, // other properties... ), )); } }); ``` With `flutter_local_notifications: 12.0.4`, this doesn't compile anymore: - `notification` and `android` are nullable. - The channel description is now a named parameter. ![image](https://user-images.githubusercontent.com/1061209/208081728-ad88bbf5-b84f-4ceb-b71d-751423acc1a5.png) Here's a version that does compile: ```dart FirebaseMessaging.onMessage.listen((RemoteMessage message) { RemoteNotification? notification = message.notification; AndroidNotification? android = message.notification?.android; // If `onMessage` is triggered with a notification, construct our own // local notification to show to users using the created channel. if (notification != null && android != null) { flutterLocalNotificationsPlugin.show( notification.hashCode, notification.title, notification.body, NotificationDetails( android: AndroidNotificationDetails( channel.id, channel.name, channelDescription: channel.description, icon: android.smallIcon, // other properties... ), )); } }); ```
non_defect
cloud messaging notifications more android code sample broken please see for a related issue the shows another android code sample dart firebasemessaging onmessage listen remotemessage message remotenotification notification message notification androidnotification android message notification android if onmessage is triggered with a notification construct our own local notification to show to users using the created channel if notification null android null flutterlocalnotificationsplugin show notification hashcode notification title notification body notificationdetails android androidnotificationdetails channel id channel name channel description icon android smallicon other properties with flutter local notifications this doesn t compile anymore notification and android are nullable the channel description is now a named parameter here s a version that does compile dart firebasemessaging onmessage listen remotemessage message remotenotification notification message notification androidnotification android message notification android if onmessage is triggered with a notification construct our own local notification to show to users using the created channel if notification null android null flutterlocalnotificationsplugin show notification hashcode notification title notification body notificationdetails android androidnotificationdetails channel id channel name channeldescription channel description icon android smallicon other properties
0
1,505
16,620,343,934
IssuesEvent
2021-06-02 23:21:17
timberio/vector
https://api.github.com/repos/timberio/vector
opened
BatchNotifier should support partial failures.
domain: performance domain: reliability type: enhancement
Currently, `BatchNotifier` can be attached to multiple events, but it does not report partial failures. This means that even if only one event in a batch fails, the entire batch could be marked as a failure. In order to not only unlock higher-performance acknowledgement via batching, but to provide maximally correct processing, `BatchNotifier` should be modified/extended to support tracking the status of individual events that have been attached to it.
True
BatchNotifier should support partial failures. - Currently, `BatchNotifier` can be attached to multiple events, but it does not report partial failures. This means that even if only one event in a batch fails, the entire batch could be marked as a failure. In order to not only unlock higher-performance acknowledgement via batching, but to provide maximally correct processing, `BatchNotifier` should be modified/extended to support tracking the status of individual events that have been attached to it.
non_defect
batchnotifier should support partial failures currently batchnotifier can be attached to multiple events but it does not report partial failures this means that even if only one event in a batch fails the entire batch could be marked as a failure in order to not only unlock higher performance acknowledgement via batching but to provide maximally correct processing batchnotifier should be modified extended to support tracking the status of individual events that have been attached to it
0
19,910
3,274,680,101
IssuesEvent
2015-10-26 12:18:22
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Query inconsistency
Team: Core Team: QuSP Type: Critical Type: Defect
The behavior of retrieving data, deserializing it and setting it on the record is racy since an update could have happened in the same time. This can lead to permanently inconsistent query results.
1.0
Query inconsistency - The behavior of retrieving data, deserializing it and setting it on the record is racy since an update could have happened in the same time. This can lead to permanently inconsistent query results.
defect
query inconsistency the behavior of retrieving data deserializing it and setting it on the record is racy since an update could have happened in the same time this can lead to permanently inconsistent query results
1
143,777
22,154,480,401
IssuesEvent
2022-06-03 20:42:30
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
opened
Use bool.TrueString and bool.FalseString instead of normal strings in HostingAbstractionsWebHostBuilderExtensions
design-proposal
## Summary Hello there folks. I think we use bool.TrueString and bool.FalseString instead of actual strings of boolean in this class https://github.com/dotnet/aspnetcore/blob/main/src/Hosting/Abstractions/src/HostingAbstractionsWebHostBuilderExtensions.cs ## Motivation and goals We can reduce magic strings. We can Find references to those static fields easily. We can reuse those fields. ## In scope - ## Out of scope - ## Risks / unknowns I think there isn't any risks about this change. ## Examples - Please let me know and I will create a PR. Thanks
1.0
Use bool.TrueString and bool.FalseString instead of normal strings in HostingAbstractionsWebHostBuilderExtensions - ## Summary Hello there folks. I think we use bool.TrueString and bool.FalseString instead of actual strings of boolean in this class https://github.com/dotnet/aspnetcore/blob/main/src/Hosting/Abstractions/src/HostingAbstractionsWebHostBuilderExtensions.cs ## Motivation and goals We can reduce magic strings. We can Find references to those static fields easily. We can reuse those fields. ## In scope - ## Out of scope - ## Risks / unknowns I think there isn't any risks about this change. ## Examples - Please let me know and I will create a PR. Thanks
non_defect
use bool truestring and bool falsestring instead of normal strings in hostingabstractionswebhostbuilderextensions summary hello there folks i think we use bool truestring and bool falsestring instead of actual strings of boolean in this class motivation and goals we can reduce magic strings we can find references to those static fields easily we can reuse those fields in scope out of scope risks unknowns i think there isn t any risks about this change examples please let me know and i will create a pr thanks
0