Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
222,549
17,081,725,929
IssuesEvent
2021-07-08 06:33:31
technology-outreach-hawaii/technology-outreach-hawaii.github.io
https://api.github.com/repos/technology-outreach-hawaii/technology-outreach-hawaii.github.io
closed
Rename Files
documentation
Please do work for this task in a branch called issue-29. Rename and delete files.
1.0
Rename Files - Please do work for this task in a branch called issue-29. Rename and delete files.
non_process
rename files please do work for this task in a branch called issue rename and delete files
0
10,825
8,745,089,431
IssuesEvent
2018-12-13 00:59:49
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Ingest pipeline for filebeat postgresql module fails on multi-core thread_id
:infrastructure Filebeat enhancement module
The ingest pipeline for the filebeat postgesql module defined here: beats/filebeat/module/postgresql/log/ingest/pipeline.json fails to parse postgresql log messages from multi-core machines because of the grok pattern for THREAD_ID. On multi-core postgresql, the thread_id looks like this: [832-1] Which will cause the whole grok pattern to fail because it expects to see a thread_id that looks like this: [832]
1.0
Ingest pipeline for filebeat postgresql module fails on multi-core thread_id - The ingest pipeline for the filebeat postgesql module defined here: beats/filebeat/module/postgresql/log/ingest/pipeline.json fails to parse postgresql log messages from multi-core machines because of the grok pattern for THREAD_ID. On multi-core postgresql, the thread_id looks like this: [832-1] Which will cause the whole grok pattern to fail because it expects to see a thread_id that looks like this: [832]
non_process
ingest pipeline for filebeat postgresql module fails on multi core thread id the ingest pipeline for the filebeat postgesql module defined here beats filebeat module postgresql log ingest pipeline json fails to parse postgresql log messages from multi core machines because of the grok pattern for thread id on multi core postgresql the thread id looks like this which will cause the whole grok pattern to fail because it expects to see a thread id that looks like this
0
6,130
8,997,406,426
IssuesEvent
2019-02-02 11:57:54
lxqt/lxqt
https://api.github.com/repos/lxqt/lxqt
closed
Compiling lxqt from source - problem with display managers
wont-process-this wontfix
Hello. I compile most software from source and I tried my luck with lxqt. I actually could manage to compile it just fine - but starting it does not work that well; or perhaps, it is a bit confusing to me ... The URLs I was using were: https://downloads.lxqt.org/downloads/libfm-qt/0.14.0/libfm-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/liblxqt/0.14.0/liblxqt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/libqtxdg/3.3.0/libqtxdg-3.3.0.tar.xz https://downloads.lxqt.org/downloads/libqtxdg/3.2.0/?C=M;O=D https://downloads.lxqt.org/downloads/libsysstat/0.4.2/libsysstat-0.4.2.tar.xz https://downloads.lxqt.org/downloads/libsysstat/0.4.1/?C=M;O=D https://downloads.lxqt.org/downloads/lxqt-about/0.14.0/lxqt-about-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-admin/0.14.0/lxqt-admin-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-build-tools/0.6.0/lxqt-build-tools-0.6.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-build-tools/ https://downloads.lxqt.org/downloads/lxqt-config/0.14.0/lxqt-config-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-config/ https://downloads.lxqt.org/downloads/lxqt-globalkeys/0.14.0/lxqt-globalkeys-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-globalkeys/ https://downloads.lxqt.org/downloads/lxqt-l10n/0.13.0/lxqt-l10n-0.13.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-globalkeys/ https://downloads.lxqt.org/downloads/lxqt-notificationd/0.14.0/lxqt-notificationd-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-openssh-askpass/0.14.0/lxqt-openssh-askpass-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-openssh-askpass/ https://downloads.lxqt.org/downloads/lxqt-panel/0.14.0/lxqt-panel-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-policykit/0.14.0/lxqt-policykit-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-policykit/ https://downloads.lxqt.org/downloads/lxqt-powermanagement/0.14.0/lxqt-powermanagement-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-qtplugin/0.14.0/lxqt-qtplugin-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-runner/0.14.0/lxqt-runner-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/lxqt-session/0.14.0/lxqt-session-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-session/ https://downloads.lxqt.org/downloads/lxqt-sudo/0.14.0/lxqt-sudo-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-sudo/ https://downloads.lxqt.org/downloads/lxqt-themes/0.14.0/lxqt-themes-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/obconf-qt/0.14.0/obconf-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/obconf-qt/ https://downloads.lxqt.org/downloads/pavucontrol-qt/0.14.0/pavucontrol-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/pcmanfm-qt/0.14.0/pcmanfm-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/pcmanfm-qt/ https://downloads.lxqt.org/downloads/qterminal/0.14.0/qterminal-0.14.0.tar.xz https://downloads.lxqt.org/downloads/qterminal/ https://downloads.lxqt.org/downloads/qtermwidget/0.14.0/qtermwidget-0.14.0.tar.xz https://downloads.lxqt.org/downloads/qterminal/ (don't mind the second entries there, they are a bit wrong; but the URLs to the .tar.xz should be correct. The above output was generated from a script which I just lazily copy/pasted into this textfield widget) I start **lxqt** via: exec ck-launch-session dbus-launch --exit-with-session startlxqt in .xinitrc I also manage to start ... something. lxqt comes up with a "choose a display manager". This is the first thing that confuses me - do I need a display manager? I login through .xinitrc for xfce, mate-desktop, fluxbox etc... I don't even HAVE a display manager. This is actually what stops me altogether - that is a blockage to me. Since I have no display manager, the popup widget stops me from continuing to go. I tried to use xdm as display manager but that widget does not even allow me to pick anything ... which is quite frustrating (I can navigate through the listing of files but not pick anything). It may be that I am missing something in my setup, but even then I think it would be useful if lxqt could be changed a little bit to allow for more flexibility. I will next give some possible entry points which could be improved here - please pick whatever you deem useful and just ignore the rest; I have no real insight into the lxqt project so I am quite clueless, but I am coming here from the point of view of a medium-advanced user, give or take. (1) Consider making that strange widget that popups optional. I am not sure if this is possible but do we need any external display manager? Perhaps a new button could be added "e. g. skip" or something like that. (2) if lxqt requires a window manager, as a mandatory part, perhaps a minimal one could be added to the lxqt codebase? (3) It would be helpful if somewhere in the widget, or anywhere else, some info box or so, a listing of supported display managers could be shown. For example, would "kwin" work? Is that even a display manager? I have no idea; I also can't tell from that widget either. (4) IF a display manager is necessary, could some compilation support be added to whatever needs it? That way I could use a configure option to cmake and then bypass that check altogether; additionally OR optionally, perhaps an additional ENVIRONMENT variable could be used to specify which display manager to use then. (5) I would perhaps postpone this whole check; perhaps first start lxqt anyway, and then use some notification widget where users can click on to set a DM. That way I could click away that widget and use lxqt. Right now I do not even get past that point at all (it may be that something is wrong on my system in general, but even then I think it would be helpful if more information is provided to the user in this event. Perhaps also have a toggle check box to disallow starting this notification widget at a later time, in the event that the user does not need any display manager. I understand that many people need this and use this but I don't need it. Hmm. I think I'll stop here - the most important point is not the 5 suggestions but more the general wish to see that I can successfully use lxqt. I have compiled all these components but can not use lxqt right now due to the widget popping up (I am actually quite sure that I also did something wrong because no other panels appear, but even in this case I think it may be helpful to show more context and more information, so that I can find out what has gone wrong altogether). As for my system, it's a linux system with gcc 8.2.0, latest binutils git cloned source, all of xorg compiled from latest stable sources, glibc quite old though at 2.23 and linux kernel also quite old at 4.4.14 - but despite being old, the latter two work quite ok for my use cases so far. Anyway thanks for reading this! I hope the next lxqt release can be a bit improved - I can compile mate-desktop and xfce just fine and start them (I am writing this from a self-compiled xfce right now for example). PS: Sorry if I mixed up the terms window manager with display manager; I know these are not the same but I am often confused so please read the above accordingly.
1.0
Compiling lxqt from source - problem with display managers - Hello. I compile most software from source and I tried my luck with lxqt. I actually could manage to compile it just fine - but starting it does not work that well; or perhaps, it is a bit confusing to me ... The URLs I was using were: https://downloads.lxqt.org/downloads/libfm-qt/0.14.0/libfm-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/liblxqt/0.14.0/liblxqt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/libqtxdg/3.3.0/libqtxdg-3.3.0.tar.xz https://downloads.lxqt.org/downloads/libqtxdg/3.2.0/?C=M;O=D https://downloads.lxqt.org/downloads/libsysstat/0.4.2/libsysstat-0.4.2.tar.xz https://downloads.lxqt.org/downloads/libsysstat/0.4.1/?C=M;O=D https://downloads.lxqt.org/downloads/lxqt-about/0.14.0/lxqt-about-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-admin/0.14.0/lxqt-admin-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-build-tools/0.6.0/lxqt-build-tools-0.6.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-build-tools/ https://downloads.lxqt.org/downloads/lxqt-config/0.14.0/lxqt-config-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-config/ https://downloads.lxqt.org/downloads/lxqt-globalkeys/0.14.0/lxqt-globalkeys-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-globalkeys/ https://downloads.lxqt.org/downloads/lxqt-l10n/0.13.0/lxqt-l10n-0.13.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-globalkeys/ https://downloads.lxqt.org/downloads/lxqt-notificationd/0.14.0/lxqt-notificationd-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-openssh-askpass/0.14.0/lxqt-openssh-askpass-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-openssh-askpass/ https://downloads.lxqt.org/downloads/lxqt-panel/0.14.0/lxqt-panel-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-policykit/0.14.0/lxqt-policykit-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-policykit/ https://downloads.lxqt.org/downloads/lxqt-powermanagement/0.14.0/lxqt-powermanagement-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-qtplugin/0.14.0/lxqt-qtplugin-0.14.0.tar.xz https://downloads.lxqt.org/downloads/liblxqt/ https://downloads.lxqt.org/downloads/lxqt-runner/0.14.0/lxqt-runner-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/lxqt-session/0.14.0/lxqt-session-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-session/ https://downloads.lxqt.org/downloads/lxqt-sudo/0.14.0/lxqt-sudo-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-sudo/ https://downloads.lxqt.org/downloads/lxqt-themes/0.14.0/lxqt-themes-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/obconf-qt/0.14.0/obconf-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/obconf-qt/ https://downloads.lxqt.org/downloads/pavucontrol-qt/0.14.0/pavucontrol-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/lxqt-runner/ https://downloads.lxqt.org/downloads/pcmanfm-qt/0.14.0/pcmanfm-qt-0.14.0.tar.xz https://downloads.lxqt.org/downloads/pcmanfm-qt/ https://downloads.lxqt.org/downloads/qterminal/0.14.0/qterminal-0.14.0.tar.xz https://downloads.lxqt.org/downloads/qterminal/ https://downloads.lxqt.org/downloads/qtermwidget/0.14.0/qtermwidget-0.14.0.tar.xz https://downloads.lxqt.org/downloads/qterminal/ (don't mind the second entries there, they are a bit wrong; but the URLs to the .tar.xz should be correct. The above output was generated from a script which I just lazily copy/pasted into this textfield widget) I start **lxqt** via: exec ck-launch-session dbus-launch --exit-with-session startlxqt in .xinitrc I also manage to start ... something. lxqt comes up with a "choose a display manager". This is the first thing that confuses me - do I need a display manager? I login through .xinitrc for xfce, mate-desktop, fluxbox etc... I don't even HAVE a display manager. This is actually what stops me altogether - that is a blockage to me. Since I have no display manager, the popup widget stops me from continuing to go. I tried to use xdm as display manager but that widget does not even allow me to pick anything ... which is quite frustrating (I can navigate through the listing of files but not pick anything). It may be that I am missing something in my setup, but even then I think it would be useful if lxqt could be changed a little bit to allow for more flexibility. I will next give some possible entry points which could be improved here - please pick whatever you deem useful and just ignore the rest; I have no real insight into the lxqt project so I am quite clueless, but I am coming here from the point of view of a medium-advanced user, give or take. (1) Consider making that strange widget that popups optional. I am not sure if this is possible but do we need any external display manager? Perhaps a new button could be added "e. g. skip" or something like that. (2) if lxqt requires a window manager, as a mandatory part, perhaps a minimal one could be added to the lxqt codebase? (3) It would be helpful if somewhere in the widget, or anywhere else, some info box or so, a listing of supported display managers could be shown. For example, would "kwin" work? Is that even a display manager? I have no idea; I also can't tell from that widget either. (4) IF a display manager is necessary, could some compilation support be added to whatever needs it? That way I could use a configure option to cmake and then bypass that check altogether; additionally OR optionally, perhaps an additional ENVIRONMENT variable could be used to specify which display manager to use then. (5) I would perhaps postpone this whole check; perhaps first start lxqt anyway, and then use some notification widget where users can click on to set a DM. That way I could click away that widget and use lxqt. Right now I do not even get past that point at all (it may be that something is wrong on my system in general, but even then I think it would be helpful if more information is provided to the user in this event. Perhaps also have a toggle check box to disallow starting this notification widget at a later time, in the event that the user does not need any display manager. I understand that many people need this and use this but I don't need it. Hmm. I think I'll stop here - the most important point is not the 5 suggestions but more the general wish to see that I can successfully use lxqt. I have compiled all these components but can not use lxqt right now due to the widget popping up (I am actually quite sure that I also did something wrong because no other panels appear, but even in this case I think it may be helpful to show more context and more information, so that I can find out what has gone wrong altogether). As for my system, it's a linux system with gcc 8.2.0, latest binutils git cloned source, all of xorg compiled from latest stable sources, glibc quite old though at 2.23 and linux kernel also quite old at 4.4.14 - but despite being old, the latter two work quite ok for my use cases so far. Anyway thanks for reading this! I hope the next lxqt release can be a bit improved - I can compile mate-desktop and xfce just fine and start them (I am writing this from a self-compiled xfce right now for example). PS: Sorry if I mixed up the terms window manager with display manager; I know these are not the same but I am often confused so please read the above accordingly.
process
compiling lxqt from source problem with display managers hello i compile most software from source and i tried my luck with lxqt i actually could manage to compile it just fine but starting it does not work that well or perhaps it is a bit confusing to me the urls i was using were don t mind the second entries there they are a bit wrong but the urls to the tar xz should be correct the above output was generated from a script which i just lazily copy pasted into this textfield widget i start lxqt via exec ck launch session dbus launch exit with session startlxqt in xinitrc i also manage to start something lxqt comes up with a choose a display manager this is the first thing that confuses me do i need a display manager i login through xinitrc for xfce mate desktop fluxbox etc i don t even have a display manager this is actually what stops me altogether that is a blockage to me since i have no display manager the popup widget stops me from continuing to go i tried to use xdm as display manager but that widget does not even allow me to pick anything which is quite frustrating i can navigate through the listing of files but not pick anything it may be that i am missing something in my setup but even then i think it would be useful if lxqt could be changed a little bit to allow for more flexibility i will next give some possible entry points which could be improved here please pick whatever you deem useful and just ignore the rest i have no real insight into the lxqt project so i am quite clueless but i am coming here from the point of view of a medium advanced user give or take consider making that strange widget that popups optional i am not sure if this is possible but do we need any external display manager perhaps a new button could be added e g skip or something like that if lxqt requires a window manager as a mandatory part perhaps a minimal one could be added to the lxqt codebase it would be helpful if somewhere in the widget or anywhere else some info box or so a listing of supported display managers could be shown for example would kwin work is that even a display manager i have no idea i also can t tell from that widget either if a display manager is necessary could some compilation support be added to whatever needs it that way i could use a configure option to cmake and then bypass that check altogether additionally or optionally perhaps an additional environment variable could be used to specify which display manager to use then i would perhaps postpone this whole check perhaps first start lxqt anyway and then use some notification widget where users can click on to set a dm that way i could click away that widget and use lxqt right now i do not even get past that point at all it may be that something is wrong on my system in general but even then i think it would be helpful if more information is provided to the user in this event perhaps also have a toggle check box to disallow starting this notification widget at a later time in the event that the user does not need any display manager i understand that many people need this and use this but i don t need it hmm i think i ll stop here the most important point is not the suggestions but more the general wish to see that i can successfully use lxqt i have compiled all these components but can not use lxqt right now due to the widget popping up i am actually quite sure that i also did something wrong because no other panels appear but even in this case i think it may be helpful to show more context and more information so that i can find out what has gone wrong altogether as for my system it s a linux system with gcc latest binutils git cloned source all of xorg compiled from latest stable sources glibc quite old though at and linux kernel also quite old at but despite being old the latter two work quite ok for my use cases so far anyway thanks for reading this i hope the next lxqt release can be a bit improved i can compile mate desktop and xfce just fine and start them i am writing this from a self compiled xfce right now for example ps sorry if i mixed up the terms window manager with display manager i know these are not the same but i am often confused so please read the above accordingly
1
3,338
6,473,182,212
IssuesEvent
2017-08-17 15:23:50
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
reopened
Error while removing asd from already physically removed disk
process_duplicate type_bug
Got the following stacktrace after removing the asd. The disk got replaced before it was removed from the GUI. ``` 2017-06-30 08:11:35 41400 -0400 - NY1SRV0019 - 24198/140118751631104 - celery/celery.worker.job - 570617 - ERROR - Task albanode.reset_asd[5daf5393-8b32-4b29-88a2-aeaed37c1e97] raised unexpected: RuntimeError(u'Disk with alias scsi-35000cca25d1a29d4 not available',) Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task R = retval = fun(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__ return self.run(*args, **kwargs) File "/opt/OpenvStorage/ovs/lib/albanode.py", line 374, in reset_asd result = node.client.add_asd(disk_id=disk_aliases[0].split('/')[-1]) File "/opt/OpenvStorage/ovs/extensions/plugins/asdmanager.py", line 206, in add_asd return self._call(requests.post, 'disks/{0}/asds'.format(disk_id), timeout=30) File "/opt/OpenvStorage/ovs/extensions/plugins/asdmanager.py", line 107, in _call raise RuntimeError(error_message) RuntimeError: Disk with alias scsi-35000cca25d1a29d4 not available ```
1.0
Error while removing asd from already physically removed disk - Got the following stacktrace after removing the asd. The disk got replaced before it was removed from the GUI. ``` 2017-06-30 08:11:35 41400 -0400 - NY1SRV0019 - 24198/140118751631104 - celery/celery.worker.job - 570617 - ERROR - Task albanode.reset_asd[5daf5393-8b32-4b29-88a2-aeaed37c1e97] raised unexpected: RuntimeError(u'Disk with alias scsi-35000cca25d1a29d4 not available',) Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task R = retval = fun(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__ return self.run(*args, **kwargs) File "/opt/OpenvStorage/ovs/lib/albanode.py", line 374, in reset_asd result = node.client.add_asd(disk_id=disk_aliases[0].split('/')[-1]) File "/opt/OpenvStorage/ovs/extensions/plugins/asdmanager.py", line 206, in add_asd return self._call(requests.post, 'disks/{0}/asds'.format(disk_id), timeout=30) File "/opt/OpenvStorage/ovs/extensions/plugins/asdmanager.py", line 107, in _call raise RuntimeError(error_message) RuntimeError: Disk with alias scsi-35000cca25d1a29d4 not available ```
process
error while removing asd from already physically removed disk got the following stacktrace after removing the asd the disk got replaced before it was removed from the gui celery celery worker job error task albanode reset asd raised unexpected runtimeerror u disk with alias scsi not available traceback most recent call last file usr lib dist packages celery app trace py line in trace task r retval fun args kwargs file usr lib dist packages celery app trace py line in protected call return self run args kwargs file opt openvstorage ovs lib albanode py line in reset asd result node client add asd disk id disk aliases split file opt openvstorage ovs extensions plugins asdmanager py line in add asd return self call requests post disks asds format disk id timeout file opt openvstorage ovs extensions plugins asdmanager py line in call raise runtimeerror error message runtimeerror disk with alias scsi not available
1
10,470
13,245,982,581
IssuesEvent
2020-08-19 15:05:24
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
internal error with self relation query
process/candidate
<!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description self relation query doesn't work ## How to reproduce ```graphql query { findManyPost(where: { morePosts: { some: { title: { equals: "my" } } } }) { id title userId postId } } ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Prisma information ```prisma generator photon { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = env("DB") } model User { id String @default(cuid()) @id posts Post[] } model Post { id String @default(cuid()) @id title String morePosts Post[] @relation("PostToPost") User User? @relation(fields: [userId], references: [id]) userId String? Post Post? @relation("PostToPost", fields: [postId], references: [id]) postId String? } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> postgres - Node.js version: <!--[Run `node -v` to see your Node.js version]--> - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` @prisma/cli : 2.5.0 Current platform : darwin Query Engine : query-engine 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/query-engine-darwin) Migration Engine : migration-engine-cli 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/migration-engine-darwin) Introspection Engine : introspection-core 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/introspection-engine-darwin) Format Binary : prisma-fmt 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/prisma-fmt-darwin) Studio : 0.259.0 ```
1.0
internal error with self relation query - <!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description self relation query doesn't work ## How to reproduce ```graphql query { findManyPost(where: { morePosts: { some: { title: { equals: "my" } } } }) { id title userId postId } } ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Prisma information ```prisma generator photon { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = env("DB") } model User { id String @default(cuid()) @id posts Post[] } model Post { id String @default(cuid()) @id title String morePosts Post[] @relation("PostToPost") User User? @relation(fields: [userId], references: [id]) userId String? Post Post? @relation("PostToPost", fields: [postId], references: [id]) postId String? } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> postgres - Node.js version: <!--[Run `node -v` to see your Node.js version]--> - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` @prisma/cli : 2.5.0 Current platform : darwin Query Engine : query-engine 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/query-engine-darwin) Migration Engine : migration-engine-cli 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/migration-engine-darwin) Introspection Engine : introspection-core 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/introspection-engine-darwin) Format Binary : prisma-fmt 9a670138b1db276001d785a2adcba1584c869d24 (at node_modules/@prisma/cli/prisma-fmt-darwin) Studio : 0.259.0 ```
process
internal error with self relation query thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description self relation query doesn t work how to reproduce graphql query findmanypost where moreposts some title equals my id title userid postid expected behavior prisma information prisma generator photon provider prisma client js datasource db provider postgresql url env db model user id string default cuid id posts post model post id string default cuid id title string moreposts post relation posttopost user user relation fields references userid string post post relation posttopost fields references postid string environment setup os database postgres node js version prisma version prisma cli current platform darwin query engine query engine at node modules prisma cli query engine darwin migration engine migration engine cli at node modules prisma cli migration engine darwin introspection engine introspection core at node modules prisma cli introspection engine darwin format binary prisma fmt at node modules prisma cli prisma fmt darwin studio
1
9,284
12,304,290,172
IssuesEvent
2020-05-11 20:14:01
nextgenhealthcare/connect
https://api.github.com/repos/nextgenhealthcare/connect
closed
Queued count does not always reset
count queued reprocess
The scenario is this. I have a database writer connector with queue always on. There is a message that errors, so the queue builds up. I stop the channel, change the connector properties to never queue, and start the channel again as paused. I browse the queued messages, and select reprocess all. The messages reprocess (or error), but the queued count stays the same. Imported Issue. Original Details: Jira Issue Key: MIRTH-3021 Reporter: seaston Created: 2013-10-21T01:27:35.000-0700
1.0
Queued count does not always reset - The scenario is this. I have a database writer connector with queue always on. There is a message that errors, so the queue builds up. I stop the channel, change the connector properties to never queue, and start the channel again as paused. I browse the queued messages, and select reprocess all. The messages reprocess (or error), but the queued count stays the same. Imported Issue. Original Details: Jira Issue Key: MIRTH-3021 Reporter: seaston Created: 2013-10-21T01:27:35.000-0700
process
queued count does not always reset the scenario is this i have a database writer connector with queue always on there is a message that errors so the queue builds up i stop the channel change the connector properties to never queue and start the channel again as paused i browse the queued messages and select reprocess all the messages reprocess or error but the queued count stays the same imported issue original details jira issue key mirth reporter seaston created
1
2,277
5,105,182,201
IssuesEvent
2017-01-05 05:52:09
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
introduce posterior_nograph option
bug preprocessor
I think it would be useful to have an option like: `options_.posterior_nograph` with default value at 0 like standard nograph. This option would allow performing all possible posterior subdraws for irfs, smoothed variables etc, but avoiding to produce thousands of useless graphs that are currently done [for large models at least]. I think it would also be useful to make a distinction w.r.t. `options_.nograph` since the latter would not trigger prior/posterior plots of parameter estimates which would be required in any case. And by the way, `pm3` and `posterior_irf` do not seem to honor the nograph option: plots are made in any case. So at least we should make sure `nograph` would work properly? If you agree, I would make the appropriate changes to the matlab `pm3` and `posterior_irf` routines with `posterior_nograph` [the preprocessor could be updated with this extra option, then].
1.0
introduce posterior_nograph option - I think it would be useful to have an option like: `options_.posterior_nograph` with default value at 0 like standard nograph. This option would allow performing all possible posterior subdraws for irfs, smoothed variables etc, but avoiding to produce thousands of useless graphs that are currently done [for large models at least]. I think it would also be useful to make a distinction w.r.t. `options_.nograph` since the latter would not trigger prior/posterior plots of parameter estimates which would be required in any case. And by the way, `pm3` and `posterior_irf` do not seem to honor the nograph option: plots are made in any case. So at least we should make sure `nograph` would work properly? If you agree, I would make the appropriate changes to the matlab `pm3` and `posterior_irf` routines with `posterior_nograph` [the preprocessor could be updated with this extra option, then].
process
introduce posterior nograph option i think it would be useful to have an option like options posterior nograph with default value at like standard nograph this option would allow performing all possible posterior subdraws for irfs smoothed variables etc but avoiding to produce thousands of useless graphs that are currently done i think it would also be useful to make a distinction w r t options nograph since the latter would not trigger prior posterior plots of parameter estimates which would be required in any case and by the way and posterior irf do not seem to honor the nograph option plots are made in any case so at least we should make sure nograph would work properly if you agree i would make the appropriate changes to the matlab and posterior irf routines with posterior nograph
1
320,198
9,777,758,705
IssuesEvent
2019-06-07 09:59:14
thonny/thonny
https://api.github.com/repos/thonny/thonny
closed
Win Xp, can't install....
bug low priority windows specifics wontfix
**[Original report](https://bitbucket.org/bitbucket-issue-migration\thonny-issues.zip/issue/642) by Anonymous.** ---------------------------------------- Hello, I'm trying to install thonny on xp machine with python 3.4.4. This is the error i get when i follow the instructions: [code]https://pastebin.com/0LUq46me[/code] Upgrading pip does not work also... Thanks in advance.
1.0
Win Xp, can't install.... - **[Original report](https://bitbucket.org/bitbucket-issue-migration\thonny-issues.zip/issue/642) by Anonymous.** ---------------------------------------- Hello, I'm trying to install thonny on xp machine with python 3.4.4. This is the error i get when i follow the instructions: [code]https://pastebin.com/0LUq46me[/code] Upgrading pip does not work also... Thanks in advance.
non_process
win xp can t install by anonymous hello i m trying to install thonny on xp machine with python this is the error i get when i follow the instructions upgrading pip does not work also thanks in advance
0
69,809
22,680,630,796
IssuesEvent
2022-07-04 09:33:27
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
[Poll] - Wrong votes in closed poll after removing 2 previous polls
T-Defect A-Polls Z-Labs
### Steps to reproduce - create a poll 1 (closed or opened) - create a poll 2 (closed or opened) - remove poll 1 - remove poll 2 - create a closed poll ### Outcome #### What did you expect? Votes for new closed poll should be not visible and empty. #### What happened instead? Votes of poll 1 are visible for the new closed poll. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? Yes
1.0
[Poll] - Wrong votes in closed poll after removing 2 previous polls - ### Steps to reproduce - create a poll 1 (closed or opened) - create a poll 2 (closed or opened) - remove poll 1 - remove poll 2 - create a closed poll ### Outcome #### What did you expect? Votes for new closed poll should be not visible and empty. #### What happened instead? Votes of poll 1 are visible for the new closed poll. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? Yes
non_process
wrong votes in closed poll after removing previous polls steps to reproduce create a poll closed or opened create a poll closed or opened remove poll remove poll create a closed poll outcome what did you expect votes for new closed poll should be not visible and empty what happened instead votes of poll are visible for the new closed poll your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no are you willing to provide a pr yes
0
129,430
5,096,928,631
IssuesEvent
2017-01-03 19:47:59
autofac/Autofac
https://api.github.com/repos/autofac/Autofac
closed
Autofac 4.0 and ASP.NET Core / .NET Core Tasks
priority-high task
This issue contains a list of (potential) tasks that need to be performed for Autofac 4.0 and adding support ASP.NET vNext. - [x] Create a develop branch following Gitflow. - [x] Make sure MyGet can build from the develop branch. - [x] Update the Autofac project to Profile259 to make the PCL compatible with aspnetcore50 framework target. - [x] Add .kproj files for Autofac and Autofac.AspNet. - [x] Add sample application for ASP.NET vNext. - [x] Remove NuGet dependencies on Autofac package while changes being made. - [x] Split out non-core projects to separate repositories. - [x] ~~Use MSBuild target with .nuspec in each project to create NuGet packages.~~ Package build is now done as part of AppVeyor build. - [x] ~~Fix the assembly, file and package version numbers.~~ Versioning is now controlled through appveyor.yml file. - [x] Tag the master branch with 3.5 prior to merging 4.0 into master (part of Gitflow anyway, but tags are required for doc site versions). - [x] Tag the master branch with 4.0 after bringing the develop code in (also part of Gitflow, also required for doc site). - [x] Update Autofac.Extensions.DependencyInjection to be strong named when [the strong named version of Microsoft.Extensions.DependencyInjection is released](https://github.com/aspnet/DependencyInjection/blob/1e21385ab12efacf1c8456809fbae87f85fa98eb/src/Microsoft.Extensions.DependencyInjection/project.json#L10). - [x] Update to netstandard from dotnet TFM when released. - [x] [Add support for `IGetRequiredServiceAware`](https://github.com/aspnet/DependencyInjection/issues/378) in Autofac.Extensions.DepedencyInjection - [x] Add release notes to project.json/nuspec files to point to the wiki release note pages for each package. (Moved to #590) - [x] Update the package icon URL in project.json/nuspec to point to the GitHub URL for the icon rather than Google Code. - [x] Create a new version of the documentation site for 4.0. - [x] Finalize the 4.0 behavior based on [changes that are being looked at for ASP.NET Core integration](https://github.com/aspnet/DependencyInjection/issues/433). - [x] Generate an updated version of the API documentation for 4.0. - [x] Update the autofac.org web site for 4.0. (Moved to #823) - [x] Finish porting [the list of integration libraries](https://github.com/autofac/Autofac/issues/594#issuecomment-139563831) - [x] Update the Examples repository to the latest integrations. - [x] Create an upgrade guide for 3.x to 4.0 for core Autofac and integration libraries. (Moved to autofac/Documentation#29)
1.0
Autofac 4.0 and ASP.NET Core / .NET Core Tasks - This issue contains a list of (potential) tasks that need to be performed for Autofac 4.0 and adding support ASP.NET vNext. - [x] Create a develop branch following Gitflow. - [x] Make sure MyGet can build from the develop branch. - [x] Update the Autofac project to Profile259 to make the PCL compatible with aspnetcore50 framework target. - [x] Add .kproj files for Autofac and Autofac.AspNet. - [x] Add sample application for ASP.NET vNext. - [x] Remove NuGet dependencies on Autofac package while changes being made. - [x] Split out non-core projects to separate repositories. - [x] ~~Use MSBuild target with .nuspec in each project to create NuGet packages.~~ Package build is now done as part of AppVeyor build. - [x] ~~Fix the assembly, file and package version numbers.~~ Versioning is now controlled through appveyor.yml file. - [x] Tag the master branch with 3.5 prior to merging 4.0 into master (part of Gitflow anyway, but tags are required for doc site versions). - [x] Tag the master branch with 4.0 after bringing the develop code in (also part of Gitflow, also required for doc site). - [x] Update Autofac.Extensions.DependencyInjection to be strong named when [the strong named version of Microsoft.Extensions.DependencyInjection is released](https://github.com/aspnet/DependencyInjection/blob/1e21385ab12efacf1c8456809fbae87f85fa98eb/src/Microsoft.Extensions.DependencyInjection/project.json#L10). - [x] Update to netstandard from dotnet TFM when released. - [x] [Add support for `IGetRequiredServiceAware`](https://github.com/aspnet/DependencyInjection/issues/378) in Autofac.Extensions.DepedencyInjection - [x] Add release notes to project.json/nuspec files to point to the wiki release note pages for each package. (Moved to #590) - [x] Update the package icon URL in project.json/nuspec to point to the GitHub URL for the icon rather than Google Code. - [x] Create a new version of the documentation site for 4.0. - [x] Finalize the 4.0 behavior based on [changes that are being looked at for ASP.NET Core integration](https://github.com/aspnet/DependencyInjection/issues/433). - [x] Generate an updated version of the API documentation for 4.0. - [x] Update the autofac.org web site for 4.0. (Moved to #823) - [x] Finish porting [the list of integration libraries](https://github.com/autofac/Autofac/issues/594#issuecomment-139563831) - [x] Update the Examples repository to the latest integrations. - [x] Create an upgrade guide for 3.x to 4.0 for core Autofac and integration libraries. (Moved to autofac/Documentation#29)
non_process
autofac and asp net core net core tasks this issue contains a list of potential tasks that need to be performed for autofac and adding support asp net vnext create a develop branch following gitflow make sure myget can build from the develop branch update the autofac project to to make the pcl compatible with framework target add kproj files for autofac and autofac aspnet add sample application for asp net vnext remove nuget dependencies on autofac package while changes being made split out non core projects to separate repositories use msbuild target with nuspec in each project to create nuget packages package build is now done as part of appveyor build fix the assembly file and package version numbers versioning is now controlled through appveyor yml file tag the master branch with prior to merging into master part of gitflow anyway but tags are required for doc site versions tag the master branch with after bringing the develop code in also part of gitflow also required for doc site update autofac extensions dependencyinjection to be strong named when update to netstandard from dotnet tfm when released in autofac extensions depedencyinjection add release notes to project json nuspec files to point to the wiki release note pages for each package moved to update the package icon url in project json nuspec to point to the github url for the icon rather than google code create a new version of the documentation site for finalize the behavior based on generate an updated version of the api documentation for update the autofac org web site for moved to finish porting update the examples repository to the latest integrations create an upgrade guide for x to for core autofac and integration libraries moved to autofac documentation
0
10,411
13,205,578,299
IssuesEvent
2020-08-14 18:15:34
timberio/vector
https://api.github.com/repos/timberio/vector
opened
Allow for inline `remap` and `filter` options
domain: config domain: processing domain: setup meta: idea needs: approval type: enhancement
I'm submitting this as a mini RFC since I think it helps to express the problem best. An idea that has some overlap with #257 and #406 is the ability to express inline filtering and remapping. ## Motivation Building pipelines in Vector can be very verbose, which is largely caused by the need to perform basic operations like filtering and remapping. For example, the following pipeline attempts to add a `type` field depending on the field the event originated from: ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] [transforms.split-by-file] type = “swimlanes” inputs = [“log-files”] [transforms.split-by-file.lanes.nginx-logs] “file.eq” = “/var/log/nginx.log” [transforms.split-by-file.lanes.postgres-logs] “file.eq” = “/var/log/postgres.log” [transforms.split-by-file.lanes.app-logs] “file.eq” = “/var/log/app.log” [transforms.add-nginx-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "nginx" [transforms.add-postgres-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "postgres" [transforms.add-app-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "app" # ... ``` This config is very verbose. ## Proposal I propose that we allow for filtering and remapping directly within transforms. We've set precedence for this with the `coerce` transform being embedded as a [`types option](https://vector.dev/docs/reference/transforms/regex_parser/#types) within various parsing transforms. We should do something similar with `filter` and `remap`: ### Inlining the `filter` option ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] [transforms.add-nginx-type] type = "add_fields" inputs = ["log-files"] fields.type = "nginx" filter."file.eq" = “/var/log/nginx.log” [transforms.add-postgres-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "postgres" filter."file.eq" = “/var/log/postgres.log” [transforms.add-app-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "app" filter."file.eq" = “/var/log/app.log” # ... ``` ### Inlining the `remap` option And inlining a `remap` option reduces it even further: ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] remap = """ if .file = “/var/log/nginx.log” { .type = "nginx" } else if .file = “/var/log/postgres.log” { .type = "postgres" } else if .file = “/var/log/app.log” { .type = "app" } """ # ... ``` ## Rationale 1. It reduces the boilerplate required to build Vector pipelines making them easier to read and understand. 2. It'll facilitate the transition from Logstash pipelines since they have a similar concept with their [conditionals](https://www.elastic.co/guide/en/logstash/current/config-examples.html#using-conditionals). ## Drawbacks 1. There are now 2 ways to do something, which could make Vector's config files harder to understand and follow. ## Open Questions 1. Do we want to generalize this solution for all transforms or continue to cherry-pick specific transforms that we know have value? It's interesting to think about this as a succinct snytax for describing pipelines instead of options. 2. Are we fighting with TOML? Would we be better off adopting another config syntax that allows for succinct, easier to read pipelines.
1.0
Allow for inline `remap` and `filter` options - I'm submitting this as a mini RFC since I think it helps to express the problem best. An idea that has some overlap with #257 and #406 is the ability to express inline filtering and remapping. ## Motivation Building pipelines in Vector can be very verbose, which is largely caused by the need to perform basic operations like filtering and remapping. For example, the following pipeline attempts to add a `type` field depending on the field the event originated from: ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] [transforms.split-by-file] type = “swimlanes” inputs = [“log-files”] [transforms.split-by-file.lanes.nginx-logs] “file.eq” = “/var/log/nginx.log” [transforms.split-by-file.lanes.postgres-logs] “file.eq” = “/var/log/postgres.log” [transforms.split-by-file.lanes.app-logs] “file.eq” = “/var/log/app.log” [transforms.add-nginx-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "nginx" [transforms.add-postgres-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "postgres" [transforms.add-app-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "app" # ... ``` This config is very verbose. ## Proposal I propose that we allow for filtering and remapping directly within transforms. We've set precedence for this with the `coerce` transform being embedded as a [`types option](https://vector.dev/docs/reference/transforms/regex_parser/#types) within various parsing transforms. We should do something similar with `filter` and `remap`: ### Inlining the `filter` option ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] [transforms.add-nginx-type] type = "add_fields" inputs = ["log-files"] fields.type = "nginx" filter."file.eq" = “/var/log/nginx.log” [transforms.add-postgres-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "postgres" filter."file.eq" = “/var/log/postgres.log” [transforms.add-app-type] type = "add_fields" inputs = ["nginx-logs"] fields.type = "app" filter."file.eq" = “/var/log/app.log” # ... ``` ### Inlining the `remap` option And inlining a `remap` option reduces it even further: ```toml [sources.log-files] type = "file" include = ["/var/log/*.log"] remap = """ if .file = “/var/log/nginx.log” { .type = "nginx" } else if .file = “/var/log/postgres.log” { .type = "postgres" } else if .file = “/var/log/app.log” { .type = "app" } """ # ... ``` ## Rationale 1. It reduces the boilerplate required to build Vector pipelines making them easier to read and understand. 2. It'll facilitate the transition from Logstash pipelines since they have a similar concept with their [conditionals](https://www.elastic.co/guide/en/logstash/current/config-examples.html#using-conditionals). ## Drawbacks 1. There are now 2 ways to do something, which could make Vector's config files harder to understand and follow. ## Open Questions 1. Do we want to generalize this solution for all transforms or continue to cherry-pick specific transforms that we know have value? It's interesting to think about this as a succinct snytax for describing pipelines instead of options. 2. Are we fighting with TOML? Would we be better off adopting another config syntax that allows for succinct, easier to read pipelines.
process
allow for inline remap and filter options i m submitting this as a mini rfc since i think it helps to express the problem best an idea that has some overlap with and is the ability to express inline filtering and remapping motivation building pipelines in vector can be very verbose which is largely caused by the need to perform basic operations like filtering and remapping for example the following pipeline attempts to add a type field depending on the field the event originated from toml type file include type “swimlanes” inputs “file eq” “ var log nginx log” “file eq” “ var log postgres log” “file eq” “ var log app log” type add fields inputs fields type nginx type add fields inputs fields type postgres type add fields inputs fields type app this config is very verbose proposal i propose that we allow for filtering and remapping directly within transforms we ve set precedence for this with the coerce transform being embedded as a within various parsing transforms we should do something similar with filter and remap inlining the filter option toml type file include type add fields inputs fields type nginx filter file eq “ var log nginx log” type add fields inputs fields type postgres filter file eq “ var log postgres log” type add fields inputs fields type app filter file eq “ var log app log” inlining the remap option and inlining a remap option reduces it even further toml type file include remap if file “ var log nginx log” type nginx else if file “ var log postgres log” type postgres else if file “ var log app log” type app rationale it reduces the boilerplate required to build vector pipelines making them easier to read and understand it ll facilitate the transition from logstash pipelines since they have a similar concept with their drawbacks there are now ways to do something which could make vector s config files harder to understand and follow open questions do we want to generalize this solution for all transforms or continue to cherry pick specific transforms that we know have value it s interesting to think about this as a succinct snytax for describing pipelines instead of options are we fighting with toml would we be better off adopting another config syntax that allows for succinct easier to read pipelines
1
8,404
11,569,535,545
IssuesEvent
2020-02-20 17:44:16
18F/tts-tech-portfolio
https://api.github.com/repos/18F/tts-tech-portfolio
opened
New Relic - Setup SIMPLE synthetics monitor/dashboard/alert method for any TTS Domains
DE.AE - Anomalies and Events - DE.CM - Security Continuous Monitoring DE.DP- Detection Processes epic: alerting
## User story As the TTS tech portfolio, we want to set up basic monitoring for all of our public urls. As a TTS System owner I want to have (atleast) basic insight into my cumulative uptime/downtime monitoring (dashboard) and be alerted for outages. ## Background New Relic Synthetics provides a no-cost tier for a basic ping service to test if a TTS domain endpoint is reachable and returns 200. Uptrends may not be available, there are currently no approved alternatives for ping/uptime metrics monitoring. ## Acceptance Criteria - [ ] Test a sample SIMPLE synthetics check - [ ] Test a sample synthetics dashboard - [ ] Test a sample synthetics alert - [ ] email - [ ] slack Related to https://github.com/18F/tts-tech-portfolio-private/issues/927
1.0
New Relic - Setup SIMPLE synthetics monitor/dashboard/alert method for any TTS Domains - ## User story As the TTS tech portfolio, we want to set up basic monitoring for all of our public urls. As a TTS System owner I want to have (atleast) basic insight into my cumulative uptime/downtime monitoring (dashboard) and be alerted for outages. ## Background New Relic Synthetics provides a no-cost tier for a basic ping service to test if a TTS domain endpoint is reachable and returns 200. Uptrends may not be available, there are currently no approved alternatives for ping/uptime metrics monitoring. ## Acceptance Criteria - [ ] Test a sample SIMPLE synthetics check - [ ] Test a sample synthetics dashboard - [ ] Test a sample synthetics alert - [ ] email - [ ] slack Related to https://github.com/18F/tts-tech-portfolio-private/issues/927
process
new relic setup simple synthetics monitor dashboard alert method for any tts domains user story as the tts tech portfolio we want to set up basic monitoring for all of our public urls as a tts system owner i want to have atleast basic insight into my cumulative uptime downtime monitoring dashboard and be alerted for outages background new relic synthetics provides a no cost tier for a basic ping service to test if a tts domain endpoint is reachable and returns uptrends may not be available there are currently no approved alternatives for ping uptime metrics monitoring acceptance criteria test a sample simple synthetics check test a sample synthetics dashboard test a sample synthetics alert email slack related to
1
198,283
14,971,223,957
IssuesEvent
2021-01-27 20:48:51
rstudio/rstudio
https://api.github.com/repos/rstudio/rstudio
closed
Can't run rsession unit tests on macOS
automation developer test
### System details RStudio Edition : dev environment RStudio Version : 1.4 master or juliet-rose branch OS Version : macOS 11.1 R Version : 3.6.1 ### Steps to reproduce the problem - build RStudio locally - from build directory, run `rstudio-tests` script ### Describe the problem in detail - the core tests run, but then rsession fails as follows: ``` Running rsession tests... dyld: Symbol not found: _GEkillDevice Referenced from: /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession Expected in: flat namespace in /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession /bin/bash: line 1: 83111 Abort trap: 6 env SEGFAULT_SIGNALS='abrt segv' LD_PRELOAD= /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession --run-tests --config-file=/Users/gary/rstudio/src/cpp/cmake-build-debug/conf/rsession-dev.conf Running r tests... dyld: Symbol not found: _GEkillDevice Referenced from: /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession Expected in: flat namespace in /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession /bin/bash: line 1: 83114 Abort trap: 6 env SEGFAULT_SIGNALS='abrt segv' LD_PRELOAD= /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession --run-script "source('/Users/gary/rstudio/src/cpp/tests/testthat/run-tests.R'); runAllTests('/Users/gary/rstudio/src/cpp', '/Users/gary/rstudio/src/cpp/cmake-build-debug')" --config-file=/Users/gary/rstudio/src/cpp/cmake-build-debug/conf/rsession-dev.conf ``` ### Describe the behavior you expected Able to run unit tests on macOS. Note that due to #6890 we haven't hit this in the official builds. Likely due to #8619. - [x] I have read the guide for [submitting good bug reports](https://github.com/rstudio/rstudio/wiki/Writing-Good-Bug-Reports). - [x] I have installed the latest version of RStudio, and confirmed that the issue still persists. - [x] If I am reporting a RStudio crash, I have included a [diagnostics report](https://support.rstudio.com/hc/en-us/articles/200321257-Running-a-Diagnostics-Report). - [x] I have done my best to include a minimal, self-contained set of instructions for consistently reproducing the issue.
1.0
Can't run rsession unit tests on macOS - ### System details RStudio Edition : dev environment RStudio Version : 1.4 master or juliet-rose branch OS Version : macOS 11.1 R Version : 3.6.1 ### Steps to reproduce the problem - build RStudio locally - from build directory, run `rstudio-tests` script ### Describe the problem in detail - the core tests run, but then rsession fails as follows: ``` Running rsession tests... dyld: Symbol not found: _GEkillDevice Referenced from: /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession Expected in: flat namespace in /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession /bin/bash: line 1: 83111 Abort trap: 6 env SEGFAULT_SIGNALS='abrt segv' LD_PRELOAD= /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession --run-tests --config-file=/Users/gary/rstudio/src/cpp/cmake-build-debug/conf/rsession-dev.conf Running r tests... dyld: Symbol not found: _GEkillDevice Referenced from: /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession Expected in: flat namespace in /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession /bin/bash: line 1: 83114 Abort trap: 6 env SEGFAULT_SIGNALS='abrt segv' LD_PRELOAD= /Users/gary/rstudio/src/cpp/cmake-build-debug/session/rsession --run-script "source('/Users/gary/rstudio/src/cpp/tests/testthat/run-tests.R'); runAllTests('/Users/gary/rstudio/src/cpp', '/Users/gary/rstudio/src/cpp/cmake-build-debug')" --config-file=/Users/gary/rstudio/src/cpp/cmake-build-debug/conf/rsession-dev.conf ``` ### Describe the behavior you expected Able to run unit tests on macOS. Note that due to #6890 we haven't hit this in the official builds. Likely due to #8619. - [x] I have read the guide for [submitting good bug reports](https://github.com/rstudio/rstudio/wiki/Writing-Good-Bug-Reports). - [x] I have installed the latest version of RStudio, and confirmed that the issue still persists. - [x] If I am reporting a RStudio crash, I have included a [diagnostics report](https://support.rstudio.com/hc/en-us/articles/200321257-Running-a-Diagnostics-Report). - [x] I have done my best to include a minimal, self-contained set of instructions for consistently reproducing the issue.
non_process
can t run rsession unit tests on macos system details rstudio edition dev environment rstudio version master or juliet rose branch os version macos r version steps to reproduce the problem build rstudio locally from build directory run rstudio tests script describe the problem in detail the core tests run but then rsession fails as follows running rsession tests dyld symbol not found gekilldevice referenced from users gary rstudio src cpp cmake build debug session rsession expected in flat namespace in users gary rstudio src cpp cmake build debug session rsession bin bash line abort trap env segfault signals abrt segv ld preload users gary rstudio src cpp cmake build debug session rsession run tests config file users gary rstudio src cpp cmake build debug conf rsession dev conf running r tests dyld symbol not found gekilldevice referenced from users gary rstudio src cpp cmake build debug session rsession expected in flat namespace in users gary rstudio src cpp cmake build debug session rsession bin bash line abort trap env segfault signals abrt segv ld preload users gary rstudio src cpp cmake build debug session rsession run script source users gary rstudio src cpp tests testthat run tests r runalltests users gary rstudio src cpp users gary rstudio src cpp cmake build debug config file users gary rstudio src cpp cmake build debug conf rsession dev conf describe the behavior you expected able to run unit tests on macos note that due to we haven t hit this in the official builds likely due to i have read the guide for i have installed the latest version of rstudio and confirmed that the issue still persists if i am reporting a rstudio crash i have included a i have done my best to include a minimal self contained set of instructions for consistently reproducing the issue
0
7,850
11,026,928,327
IssuesEvent
2019-12-06 08:12:29
wirecard/shop-systems-coding-guidelines
https://api.github.com/repos/wirecard/shop-systems-coding-guidelines
opened
Handling TODOs
processes
* TODO must always be accompanied by ticket number; e.g. you find a bug/issue/improvement, you take a shortcut in the code, but you know how a cleaner approach would look like; naturally, you write a TODO in the code; as part of writing a TODO, you have to create a separate ticket and use that ticket number in your newly written TODO, as a reference to keep things in check * related to previous, boyscout rule: when you see a TODO with a ticket number in it, check if the referenced ticket has already been tackled, and implement it if so * related to previous, boyscout rule: when you see a TODO without a ticket number in it, create a ticket for it and update the TODO * TODOs without ticket numbers are fine during development, as long as they don't reach the code review; useful when you want to get a prototype working without bothering with details, but you need a reminder to address those details before starting a code review
1.0
Handling TODOs - * TODO must always be accompanied by ticket number; e.g. you find a bug/issue/improvement, you take a shortcut in the code, but you know how a cleaner approach would look like; naturally, you write a TODO in the code; as part of writing a TODO, you have to create a separate ticket and use that ticket number in your newly written TODO, as a reference to keep things in check * related to previous, boyscout rule: when you see a TODO with a ticket number in it, check if the referenced ticket has already been tackled, and implement it if so * related to previous, boyscout rule: when you see a TODO without a ticket number in it, create a ticket for it and update the TODO * TODOs without ticket numbers are fine during development, as long as they don't reach the code review; useful when you want to get a prototype working without bothering with details, but you need a reminder to address those details before starting a code review
process
handling todos todo must always be accompanied by ticket number e g you find a bug issue improvement you take a shortcut in the code but you know how a cleaner approach would look like naturally you write a todo in the code as part of writing a todo you have to create a separate ticket and use that ticket number in your newly written todo as a reference to keep things in check related to previous boyscout rule when you see a todo with a ticket number in it check if the referenced ticket has already been tackled and implement it if so related to previous boyscout rule when you see a todo without a ticket number in it create a ticket for it and update the todo todos without ticket numbers are fine during development as long as they don t reach the code review useful when you want to get a prototype working without bothering with details but you need a reminder to address those details before starting a code review
1
21,275
28,442,352,419
IssuesEvent
2023-04-16 03:18:18
cse442-at-ub/project_s23-one-belt-one-road
https://api.github.com/repos/cse442-at-ub/project_s23-one-belt-one-road
closed
Convert login/register to an api with documentation
Processing Task Sprint 3
Task Tests: Test 1: 1) Go to our project GitHub at https://github.com/cse442-at-ub/project_s23-one-belt-one-road and verify there is a file called apiFunctions 2) Verify there is a function named 'login', which takes parameters 'username' and 'password' 3) Verify there is a function named 'register', which takes parameters 'username', 'password', 'password2', and 'email' Test 2: 1) Go to our project GitHub at https://github.com/cse442-at-ub/project_s23-one-belt-one-road and verify there is a file called apiFunctions 2) Verify there are instructions on how to use the login api and registration api 3) Verify there is documentation indicate the succusses of an API call to the developer
1.0
Convert login/register to an api with documentation - Task Tests: Test 1: 1) Go to our project GitHub at https://github.com/cse442-at-ub/project_s23-one-belt-one-road and verify there is a file called apiFunctions 2) Verify there is a function named 'login', which takes parameters 'username' and 'password' 3) Verify there is a function named 'register', which takes parameters 'username', 'password', 'password2', and 'email' Test 2: 1) Go to our project GitHub at https://github.com/cse442-at-ub/project_s23-one-belt-one-road and verify there is a file called apiFunctions 2) Verify there are instructions on how to use the login api and registration api 3) Verify there is documentation indicate the succusses of an API call to the developer
process
convert login register to an api with documentation task tests test go to our project github at and verify there is a file called apifunctions verify there is a function named login which takes parameters username and password verify there is a function named register which takes parameters username password and email test go to our project github at and verify there is a file called apifunctions verify there are instructions on how to use the login api and registration api verify there is documentation indicate the succusses of an api call to the developer
1
6,684
9,805,500,972
IssuesEvent
2019-06-12 09:08:01
EthVM/EthVM
https://api.github.com/repos/EthVM/EthVM
closed
Only map objects are supported when there is no schema
bug project:processing
* **I'm submitting a ...** - [ ] feature request - [x] bug report * **Bug Report** While deploying processing has thrown this exception: ``` [2019-06-11 22:30:20,448] DEBUG Finished fetching range: 178279..178534 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,669] DEBUG Latest block number = 178774, nextRange last = 178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,669] DEBUG Next = 178775..178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,811] DEBUG Finished fetching range: 178535..178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,026] DEBUG Total blocks fetched = 179303, fetch count = 1024. Elapsed = 4582 ms, target fetch = 3000 ms, % of target fetch = 1.5273334, trace count = 11730, avg trace count = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,026] DEBUG Reorg size = 5, records size = 5121 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,026] DEBUG Polled 5126 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask) [2019-06-11 22:30:21,039] ERROR WorkerSinkTask{id=postgres-block-sink-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask) org.apache.kafka.connect.errors.ConnectException: Tolerance exceeded in error handler at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:178) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execute(RetryWithToleranceOperator.java:104) at org.apache.kafka.connect.runtime.TransformationChain.apply(TransformationChain.java:50) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertAndTransformRecord(WorkerSinkTask.java:506) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertMessages(WorkerSinkTask.java:464) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:320) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:224) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:192) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:175) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:219) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.IllegalArgumentException: Only map objects are supported when there is no schema at com.ethvm.kafka.connect.transforms.SnakeCase.applySchemaless(SnakeCase.kt:62) at com.ethvm.kafka.connect.transforms.SnakeCase.apply(SnakeCase.kt:35) at org.apache.kafka.connect.runtime.TransformationChain.lambda$apply$0(TransformationChain.java:50) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndRetry(RetryWithToleranceOperator.java:128) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:162) ... 14 more [2019-06-11 22:30:21,044] ERROR WorkerSinkTask{id=postgres-block-sink-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask) [2019-06-11 22:30:21,970] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask) [2019-06-11 22:30:21,973] DEBUG Next range. Current tail = 179303, current head = 5775865, maxSize = 512 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker) [2019-06-11 22:30:21,973] DEBUG Next range from chain tracker: 179303..179814 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker) [2019-06-11 22:30:21,973] DEBUG Range = 179303..179814, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,973] DEBUG Fetching range: 179303..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,973] DEBUG Next = 179303..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179431..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179431..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179687..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179687..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179559..179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179559..179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,788] DEBUG Latest block number = 179765, nextRange last = 179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,788] DEBUG Next = 179766..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,903] DEBUG Latest block number = 179398, nextRange last = 179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,905] DEBUG Next = 179399..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,435] DEBUG Latest block number = 179516, nextRange last = 179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,435] DEBUG Next = 179517..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,455] ERROR WorkerSinkTask{id=postgres-block-sink-1} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask) org.apache.kafka.connect.errors.ConnectException: Tolerance exceeded in error handler at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:178) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execute(RetryWithToleranceOperator.java:104) at org.apache.kafka.connect.runtime.TransformationChain.apply(TransformationChain.java:50) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertAndTransformRecord(WorkerSinkTask.java:506) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertMessages(WorkerSinkTask.java:464) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:320) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:224) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:192) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:175) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:219) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.IllegalArgumentException: Only map objects are supported when there is no schema at com.ethvm.kafka.connect.transforms.SnakeCase.applySchemaless(SnakeCase.kt:62) at com.ethvm.kafka.connect.transforms.SnakeCase.apply(SnakeCase.kt:35) at org.apache.kafka.connect.runtime.TransformationChain.lambda$apply$0(TransformationChain.java:50) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndRetry(RetryWithToleranceOperator.java:128) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:162) ... 14 more [2019-06-11 22:30:23,466] ERROR WorkerSinkTask{id=postgres-block-sink-1} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask) [2019-06-11 22:30:23,501] DEBUG Latest block number = 179635, nextRange last = 179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) ````
1.0
Only map objects are supported when there is no schema - * **I'm submitting a ...** - [ ] feature request - [x] bug report * **Bug Report** While deploying processing has thrown this exception: ``` [2019-06-11 22:30:20,448] DEBUG Finished fetching range: 178279..178534 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,669] DEBUG Latest block number = 178774, nextRange last = 178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,669] DEBUG Next = 178775..178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:20,811] DEBUG Finished fetching range: 178535..178790 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,026] DEBUG Total blocks fetched = 179303, fetch count = 1024. Elapsed = 4582 ms, target fetch = 3000 ms, % of target fetch = 1.5273334, trace count = 11730, avg trace count = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,026] DEBUG Reorg size = 5, records size = 5121 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,026] DEBUG Polled 5126 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask) [2019-06-11 22:30:21,039] ERROR WorkerSinkTask{id=postgres-block-sink-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask) org.apache.kafka.connect.errors.ConnectException: Tolerance exceeded in error handler at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:178) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execute(RetryWithToleranceOperator.java:104) at org.apache.kafka.connect.runtime.TransformationChain.apply(TransformationChain.java:50) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertAndTransformRecord(WorkerSinkTask.java:506) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertMessages(WorkerSinkTask.java:464) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:320) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:224) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:192) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:175) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:219) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.IllegalArgumentException: Only map objects are supported when there is no schema at com.ethvm.kafka.connect.transforms.SnakeCase.applySchemaless(SnakeCase.kt:62) at com.ethvm.kafka.connect.transforms.SnakeCase.apply(SnakeCase.kt:35) at org.apache.kafka.connect.runtime.TransformationChain.lambda$apply$0(TransformationChain.java:50) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndRetry(RetryWithToleranceOperator.java:128) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:162) ... 14 more [2019-06-11 22:30:21,044] ERROR WorkerSinkTask{id=postgres-block-sink-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask) [2019-06-11 22:30:21,970] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask) [2019-06-11 22:30:21,973] DEBUG Next range. Current tail = 179303, current head = 5775865, maxSize = 512 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker) [2019-06-11 22:30:21,973] DEBUG Next range from chain tracker: 179303..179814 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker) [2019-06-11 22:30:21,973] DEBUG Range = 179303..179814, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource) [2019-06-11 22:30:21,973] DEBUG Fetching range: 179303..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,973] DEBUG Next = 179303..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179431..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179431..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179687..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179687..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Fetching range: 179559..179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:21,974] DEBUG Next = 179559..179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,788] DEBUG Latest block number = 179765, nextRange last = 179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,788] DEBUG Next = 179766..179814 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,903] DEBUG Latest block number = 179398, nextRange last = 179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:22,905] DEBUG Next = 179399..179430 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,435] DEBUG Latest block number = 179516, nextRange last = 179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,435] DEBUG Next = 179517..179558 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) [2019-06-11 22:30:23,455] ERROR WorkerSinkTask{id=postgres-block-sink-1} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask) org.apache.kafka.connect.errors.ConnectException: Tolerance exceeded in error handler at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:178) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execute(RetryWithToleranceOperator.java:104) at org.apache.kafka.connect.runtime.TransformationChain.apply(TransformationChain.java:50) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertAndTransformRecord(WorkerSinkTask.java:506) at org.apache.kafka.connect.runtime.WorkerSinkTask.convertMessages(WorkerSinkTask.java:464) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:320) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:224) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:192) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:175) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:219) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.IllegalArgumentException: Only map objects are supported when there is no schema at com.ethvm.kafka.connect.transforms.SnakeCase.applySchemaless(SnakeCase.kt:62) at com.ethvm.kafka.connect.transforms.SnakeCase.apply(SnakeCase.kt:35) at org.apache.kafka.connect.runtime.TransformationChain.lambda$apply$0(TransformationChain.java:50) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndRetry(RetryWithToleranceOperator.java:128) at org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperator.execAndHandleError(RetryWithToleranceOperator.java:162) ... 14 more [2019-06-11 22:30:23,466] ERROR WorkerSinkTask{id=postgres-block-sink-1} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask) [2019-06-11 22:30:23,501] DEBUG Latest block number = 179635, nextRange last = 179686 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource) ````
process
only map objects are supported when there is no schema i m submitting a feature request bug report bug report while deploying processing has thrown this exception debug finished fetching range com ethvm kafka connect sources sources parityfullblocksource debug latest block number nextrange last com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug finished fetching range com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask error workersinktask id postgres block sink task threw an uncaught and unrecoverable exception org apache kafka connect runtime workertask org apache kafka connect errors connectexception tolerance exceeded in error handler at org apache kafka connect runtime errors retrywithtoleranceoperator execandhandleerror retrywithtoleranceoperator java at org apache kafka connect runtime errors retrywithtoleranceoperator execute retrywithtoleranceoperator java at org apache kafka connect runtime transformationchain apply transformationchain java at org apache kafka connect runtime workersinktask convertandtransformrecord workersinktask java at org apache kafka connect runtime workersinktask convertmessages workersinktask java at org apache kafka connect runtime workersinktask poll workersinktask java at org apache kafka connect runtime workersinktask iteration workersinktask java at org apache kafka connect runtime workersinktask execute workersinktask java at org apache kafka connect runtime workertask dorun workertask java at org apache kafka connect runtime workertask run workertask java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java lang illegalargumentexception only map objects are supported when there is no schema at com ethvm kafka connect transforms snakecase applyschemaless snakecase kt at com ethvm kafka connect transforms snakecase apply snakecase kt at org apache kafka connect runtime transformationchain lambda apply transformationchain java at org apache kafka connect runtime errors retrywithtoleranceoperator execandretry retrywithtoleranceoperator java at org apache kafka connect runtime errors retrywithtoleranceoperator execandhandleerror retrywithtoleranceoperator java more error workersinktask id postgres block sink task is being killed and will not recover until manually restarted org apache kafka connect runtime workertask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head maxsize com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug fetching range com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug fetching range com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug fetching range com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug latest block number nextrange last com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug latest block number nextrange last com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource debug latest block number nextrange last com ethvm kafka connect sources sources parityfullblocksource debug next com ethvm kafka connect sources sources parityfullblocksource error workersinktask id postgres block sink task threw an uncaught and unrecoverable exception org apache kafka connect runtime workertask org apache kafka connect errors connectexception tolerance exceeded in error handler at org apache kafka connect runtime errors retrywithtoleranceoperator execandhandleerror retrywithtoleranceoperator java at org apache kafka connect runtime errors retrywithtoleranceoperator execute retrywithtoleranceoperator java at org apache kafka connect runtime transformationchain apply transformationchain java at org apache kafka connect runtime workersinktask convertandtransformrecord workersinktask java at org apache kafka connect runtime workersinktask convertmessages workersinktask java at org apache kafka connect runtime workersinktask poll workersinktask java at org apache kafka connect runtime workersinktask iteration workersinktask java at org apache kafka connect runtime workersinktask execute workersinktask java at org apache kafka connect runtime workertask dorun workertask java at org apache kafka connect runtime workertask run workertask java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java lang illegalargumentexception only map objects are supported when there is no schema at com ethvm kafka connect transforms snakecase applyschemaless snakecase kt at com ethvm kafka connect transforms snakecase apply snakecase kt at org apache kafka connect runtime transformationchain lambda apply transformationchain java at org apache kafka connect runtime errors retrywithtoleranceoperator execandretry retrywithtoleranceoperator java at org apache kafka connect runtime errors retrywithtoleranceoperator execandhandleerror retrywithtoleranceoperator java more error workersinktask id postgres block sink task is being killed and will not recover until manually restarted org apache kafka connect runtime workertask debug latest block number nextrange last com ethvm kafka connect sources sources parityfullblocksource
1
14,228
17,148,046,978
IssuesEvent
2021-07-13 16:44:33
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
closed
Unit tests emit deprecation warnings
api: storage type: process
``` $ .nox/unit-3-8/bin/py.test tests/unit/ ============================= test session starts ============================== platform linux -- Python 3.8.1, pytest-6.2.4, py-1.10.0, pluggy-0.13.1 rootdir: /home/tseaver/projects/agendaless/Google/src/python-storage plugins: cov-2.12.1 collected 1028 items tests/unit/test__helpers.py .................................... [ 3%] tests/unit/test__http.py ......... [ 4%] tests/unit/test__signing.py .s.......................................... [ 8%] ...................................... [ 12%] tests/unit/test_acl.py ................................................. [ 17%] ...................... [ 19%] tests/unit/test_batch.py ........................ [ 21%] tests/unit/test_blob.py ................................................ [ 26%] ........................................................................ [ 33%] ........................................................................ [ 40%] ........................................................................ [ 47%] ........................................................................ [ 54%] ............................... [ 57%] tests/unit/test_bucket.py .............................................. [ 61%] ........................................................................ [ 68%] ........................................................................ [ 75%] .............................. [ 78%] tests/unit/test_client.py .............................................. [ 83%] ....................................................... [ 88%] tests/unit/test_fileio.py ................................. [ 91%] tests/unit/test_hmac_key.py .............................. [ 94%] tests/unit/test_notification.py ............................. [ 97%] tests/unit/test_retry.py .......................... [100%] =============================== warnings summary =============================== tests/unit/test_blob.py::Test_Blob::test_compose_w_if_generation_match_and_if_s_generation_match /home/tseaver/projects/agendaless/Google/src/python-storage/tests/unit/test_blob.py:4319: DeprecationWarning: ("'if_generation_match: type list' is deprecated and supported for backwards-compatability reasons only. Use 'if_source_generation_match' instead' to match source objects' generations.",) destination.compose( tests/unit/test_blob.py::Test_Blob::test_download_as_string_no_retry /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:1516: PendingDeprecationWarning: Blob.download_as_string() is deprecated and will be removed in future.Use Blob.download_as_bytes() instead. warnings.warn( tests/unit/test_blob.py::Test_Blob::test_upload_from_filename_with_num_retries /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:2712: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self.upload_from_file( tests/unit/test_blob.py::Test_Blob::test_upload_from_string_w_text_w_num_retries /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:2844: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self.upload_from_file( tests/unit/test_fileio.py::TestBlobWriterBinary::test_num_retries_and_retry_conflict tests/unit/test_fileio.py::TestBlobWriterBinary::test_write tests/unit/test_fileio.py::TestBlobWriterText::test_write /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/fileio.py:382: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self._initiate_upload() -- Docs: https://docs.pytest.org/en/stable/warnings.html ================= 1027 passed, 1 skipped, 7 warnings in 11.15s ================= ```
1.0
Unit tests emit deprecation warnings - ``` $ .nox/unit-3-8/bin/py.test tests/unit/ ============================= test session starts ============================== platform linux -- Python 3.8.1, pytest-6.2.4, py-1.10.0, pluggy-0.13.1 rootdir: /home/tseaver/projects/agendaless/Google/src/python-storage plugins: cov-2.12.1 collected 1028 items tests/unit/test__helpers.py .................................... [ 3%] tests/unit/test__http.py ......... [ 4%] tests/unit/test__signing.py .s.......................................... [ 8%] ...................................... [ 12%] tests/unit/test_acl.py ................................................. [ 17%] ...................... [ 19%] tests/unit/test_batch.py ........................ [ 21%] tests/unit/test_blob.py ................................................ [ 26%] ........................................................................ [ 33%] ........................................................................ [ 40%] ........................................................................ [ 47%] ........................................................................ [ 54%] ............................... [ 57%] tests/unit/test_bucket.py .............................................. [ 61%] ........................................................................ [ 68%] ........................................................................ [ 75%] .............................. [ 78%] tests/unit/test_client.py .............................................. [ 83%] ....................................................... [ 88%] tests/unit/test_fileio.py ................................. [ 91%] tests/unit/test_hmac_key.py .............................. [ 94%] tests/unit/test_notification.py ............................. [ 97%] tests/unit/test_retry.py .......................... [100%] =============================== warnings summary =============================== tests/unit/test_blob.py::Test_Blob::test_compose_w_if_generation_match_and_if_s_generation_match /home/tseaver/projects/agendaless/Google/src/python-storage/tests/unit/test_blob.py:4319: DeprecationWarning: ("'if_generation_match: type list' is deprecated and supported for backwards-compatability reasons only. Use 'if_source_generation_match' instead' to match source objects' generations.",) destination.compose( tests/unit/test_blob.py::Test_Blob::test_download_as_string_no_retry /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:1516: PendingDeprecationWarning: Blob.download_as_string() is deprecated and will be removed in future.Use Blob.download_as_bytes() instead. warnings.warn( tests/unit/test_blob.py::Test_Blob::test_upload_from_filename_with_num_retries /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:2712: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self.upload_from_file( tests/unit/test_blob.py::Test_Blob::test_upload_from_string_w_text_w_num_retries /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/blob.py:2844: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self.upload_from_file( tests/unit/test_fileio.py::TestBlobWriterBinary::test_num_retries_and_retry_conflict tests/unit/test_fileio.py::TestBlobWriterBinary::test_write tests/unit/test_fileio.py::TestBlobWriterText::test_write /home/tseaver/projects/agendaless/Google/src/python-storage/google/cloud/storage/fileio.py:382: DeprecationWarning: `num_retries` has been deprecated and will be removed in a future release. Use the `retry` argument with a Retry or ConditionalRetryPolicy object, or None, instead. self._initiate_upload() -- Docs: https://docs.pytest.org/en/stable/warnings.html ================= 1027 passed, 1 skipped, 7 warnings in 11.15s ================= ```
process
unit tests emit deprecation warnings nox unit bin py test tests unit test session starts platform linux python pytest py pluggy rootdir home tseaver projects agendaless google src python storage plugins cov collected items tests unit test helpers py tests unit test http py tests unit test signing py s tests unit test acl py tests unit test batch py tests unit test blob py tests unit test bucket py tests unit test client py tests unit test fileio py tests unit test hmac key py tests unit test notification py tests unit test retry py warnings summary tests unit test blob py test blob test compose w if generation match and if s generation match home tseaver projects agendaless google src python storage tests unit test blob py deprecationwarning if generation match type list is deprecated and supported for backwards compatability reasons only use if source generation match instead to match source objects generations destination compose tests unit test blob py test blob test download as string no retry home tseaver projects agendaless google src python storage google cloud storage blob py pendingdeprecationwarning blob download as string is deprecated and will be removed in future use blob download as bytes instead warnings warn tests unit test blob py test blob test upload from filename with num retries home tseaver projects agendaless google src python storage google cloud storage blob py deprecationwarning num retries has been deprecated and will be removed in a future release use the retry argument with a retry or conditionalretrypolicy object or none instead self upload from file tests unit test blob py test blob test upload from string w text w num retries home tseaver projects agendaless google src python storage google cloud storage blob py deprecationwarning num retries has been deprecated and will be removed in a future release use the retry argument with a retry or conditionalretrypolicy object or none instead self upload from file tests unit test fileio py testblobwriterbinary test num retries and retry conflict tests unit test fileio py testblobwriterbinary test write tests unit test fileio py testblobwritertext test write home tseaver projects agendaless google src python storage google cloud storage fileio py deprecationwarning num retries has been deprecated and will be removed in a future release use the retry argument with a retry or conditionalretrypolicy object or none instead self initiate upload docs passed skipped warnings in
1
22,744
32,060,646,451
IssuesEvent
2023-09-24 16:12:02
serai-dex/serai
https://api.github.com/repos/serai-dex/serai
opened
Processor database cleaning up
improvement processor
Right now, the processor keeps a large database which only grows overtime. It keeps a full history of every received outputs, even when we'll never read them again. We should add pruning of no-longer-will-be-read data. On the coordinator side, the main note is https://github.com/serai-dex/serai/issues/352#issuecomment-1724708016. Accordingly, I'd like to keep this scoped to the processor.
1.0
Processor database cleaning up - Right now, the processor keeps a large database which only grows overtime. It keeps a full history of every received outputs, even when we'll never read them again. We should add pruning of no-longer-will-be-read data. On the coordinator side, the main note is https://github.com/serai-dex/serai/issues/352#issuecomment-1724708016. Accordingly, I'd like to keep this scoped to the processor.
process
processor database cleaning up right now the processor keeps a large database which only grows overtime it keeps a full history of every received outputs even when we ll never read them again we should add pruning of no longer will be read data on the coordinator side the main note is accordingly i d like to keep this scoped to the processor
1
173,146
13,388,753,309
IssuesEvent
2020-09-02 17:50:04
objectos/comuns
https://api.github.com/repos/objectos/comuns
closed
ConcreteProcessor: interface support
c:latest t:feature
Either add interface generation support to ConcreteProcessor or create a new processor/annotation pair to support it. Initial use case is the FunctionalInterface annotation: ```java @Concrete( modifiers = "public", simpleName = "Function" ) interface AbstractFunction<R, T> { R apply(T t); } interface AbstractFunctionJava6<R, T> extends AbstractFunction<R, T> {} @Concrete.Annotated("@FunctionalInterface") interface AbstractFunctionJava8<R, T> extends AbstractFunction<R, T> {} ``` generates: ```java public interface Function<R, T> extends AbstractFunctionJava6<R, T> {} // or @FunctionalInterface public interface Function<R, T> extends AbstractFunctionJava8<R, T> {} ```
1.0
ConcreteProcessor: interface support - Either add interface generation support to ConcreteProcessor or create a new processor/annotation pair to support it. Initial use case is the FunctionalInterface annotation: ```java @Concrete( modifiers = "public", simpleName = "Function" ) interface AbstractFunction<R, T> { R apply(T t); } interface AbstractFunctionJava6<R, T> extends AbstractFunction<R, T> {} @Concrete.Annotated("@FunctionalInterface") interface AbstractFunctionJava8<R, T> extends AbstractFunction<R, T> {} ``` generates: ```java public interface Function<R, T> extends AbstractFunctionJava6<R, T> {} // or @FunctionalInterface public interface Function<R, T> extends AbstractFunctionJava8<R, T> {} ```
non_process
concreteprocessor interface support either add interface generation support to concreteprocessor or create a new processor annotation pair to support it initial use case is the functionalinterface annotation java concrete modifiers public simplename function interface abstractfunction r apply t t interface extends abstractfunction concrete annotated functionalinterface interface extends abstractfunction generates java public interface function extends or functionalinterface public interface function extends
0
65,501
14,727,876,837
IssuesEvent
2021-01-06 09:11:09
Seagate/cortx-s3server
https://api.github.com/repos/Seagate/cortx-s3server
closed
CVE-2016-0753 (Medium) detected in activemodel-4.2.2.gem
needs-attention needs-triage security vulnerability
## CVE-2016-0753 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activemodel-4.2.2.gem</b></p></summary> <p>A toolkit for building modeling frameworks like Active Record. Rich support for attributes, callbacks, validations, serialization, internationalization, and testing.</p> <p>Library home page: <a href="https://rubygems.org/gems/activemodel-4.2.2.gem">https://rubygems.org/gems/activemodel-4.2.2.gem</a></p> <p> Dependency Hierarchy: - web-console-2.2.1.gem (Root Library) - :x: **activemodel-4.2.2.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/fde64200b4f94603ae17220b98da6422a531445e">fde64200b4f94603ae17220b98da6422a531445e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Active Model in Ruby on Rails 4.1.x before 4.1.14.1, 4.2.x before 4.2.5.1, and 5.x before 5.0.0.beta1.1 supports the use of instance-level writers for class accessors, which allows remote attackers to bypass intended validation steps via crafted parameters. <p>Publish Date: 2016-02-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-0753>CVE-2016-0753</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-0753">https://nvd.nist.gov/vuln/detail/CVE-2016-0753</a></p> <p>Release Date: 2016-02-16</p> <p>Fix Resolution: 4.1.14.1,4.2.5.1,5.0.0.beta1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-0753 (Medium) detected in activemodel-4.2.2.gem - ## CVE-2016-0753 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activemodel-4.2.2.gem</b></p></summary> <p>A toolkit for building modeling frameworks like Active Record. Rich support for attributes, callbacks, validations, serialization, internationalization, and testing.</p> <p>Library home page: <a href="https://rubygems.org/gems/activemodel-4.2.2.gem">https://rubygems.org/gems/activemodel-4.2.2.gem</a></p> <p> Dependency Hierarchy: - web-console-2.2.1.gem (Root Library) - :x: **activemodel-4.2.2.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/fde64200b4f94603ae17220b98da6422a531445e">fde64200b4f94603ae17220b98da6422a531445e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Active Model in Ruby on Rails 4.1.x before 4.1.14.1, 4.2.x before 4.2.5.1, and 5.x before 5.0.0.beta1.1 supports the use of instance-level writers for class accessors, which allows remote attackers to bypass intended validation steps via crafted parameters. <p>Publish Date: 2016-02-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-0753>CVE-2016-0753</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-0753">https://nvd.nist.gov/vuln/detail/CVE-2016-0753</a></p> <p>Release Date: 2016-02-16</p> <p>Fix Resolution: 4.1.14.1,4.2.5.1,5.0.0.beta1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in activemodel gem cve medium severity vulnerability vulnerable library activemodel gem a toolkit for building modeling frameworks like active record rich support for attributes callbacks validations serialization internationalization and testing library home page a href dependency hierarchy web console gem root library x activemodel gem vulnerable library found in head commit a href found in base branch main vulnerability details active model in ruby on rails x before x before and x before supports the use of instance level writers for class accessors which allows remote attackers to bypass intended validation steps via crafted parameters publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
438,319
30,636,604,448
IssuesEvent
2023-07-24 18:16:35
refinedev/refine
https://api.github.com/repos/refinedev/refine
closed
[DOC] Wrong symbol used in the date picker doc
documentation
### Documentation issue ![image](https://github.com/refinedev/refine/assets/91601706/ca9105a9-3817-471f-aa67-43172a86da91) this is not the correct symbol used here and copying this throws error in the code editor. ### Describe the thing to improve Instead proper single quote must be used. ### Describe the solution (optional) ![image](https://github.com/refinedev/refine/assets/91601706/6ced1bd5-170c-43a2-a00d-8ae1dc84d9eb) I have already made the change by forking the repo as you can see in the image Can I create a PR ?
1.0
[DOC] Wrong symbol used in the date picker doc - ### Documentation issue ![image](https://github.com/refinedev/refine/assets/91601706/ca9105a9-3817-471f-aa67-43172a86da91) this is not the correct symbol used here and copying this throws error in the code editor. ### Describe the thing to improve Instead proper single quote must be used. ### Describe the solution (optional) ![image](https://github.com/refinedev/refine/assets/91601706/6ced1bd5-170c-43a2-a00d-8ae1dc84d9eb) I have already made the change by forking the repo as you can see in the image Can I create a PR ?
non_process
wrong symbol used in the date picker doc documentation issue this is not the correct symbol used here and copying this throws error in the code editor describe the thing to improve instead proper single quote must be used describe the solution optional i have already made the change by forking the repo as you can see in the image can i create a pr
0
355,097
25,175,568,703
IssuesEvent
2022-11-11 08:57:04
waynezsy/pe
https://api.github.com/repos/waynezsy/pe
opened
UG does not mention date format
type.DocumentationBug severity.Low
As per title. Users will have to infer or get it wrong and depend on the application error message to figure it out. <!--session: 1668153144028-f3f2b80a-0039-49d4-8459-d034f6b371c6--> <!--Version: Web v3.4.4-->
1.0
UG does not mention date format - As per title. Users will have to infer or get it wrong and depend on the application error message to figure it out. <!--session: 1668153144028-f3f2b80a-0039-49d4-8459-d034f6b371c6--> <!--Version: Web v3.4.4-->
non_process
ug does not mention date format as per title users will have to infer or get it wrong and depend on the application error message to figure it out
0
3,967
6,900,429,572
IssuesEvent
2017-11-24 18:40:08
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
[sct_register] manage dimensionality for 2D
card:WORK_IN_PROCESS help wanted priority:LOW sct_concat_transfo
Register ````` isct_antsRegistration --dimensionality 2 --transform Affine[0.5] --metric MeanSquares[dest.nii.gz,src_regaffine_smooth.nii.gz,1,4] --convergence 100x100 --shrink-factors 2x2 --smoothing-sigmas 1x1vox --transform BSplineSyN[0.5,3] --metric MeanSquares[dest.nii.gz,src_regaffine_smooth.nii.gz,1,4] --convergence 500x500 --shrink-factors 2x2 --smoothing-sigmas 1x1vox --output [warp_,src_regaffine_smooth_reg.nii.gz] fslview dest.nii.gz src_regaffine_smooth_reg.nii.gz & ````` ![screen shot 2017-08-01 at 11 34 10 am](https://user-images.githubusercontent.com/7785316/28833450-5d12d03a-76ad-11e7-8ab1-f6aa4faa9639.png) Then ````` sct_concat_transfo -d dest.nii.gz -w warp_1Warp.nii.gz,warp_0GenericAffine.mat ````` and ````` sct_apply_transfo -i src_regaffine_smooth.nii.gz -d dest.nii.gz -w warp_final.nii.gz fslview dest.nii.gz src_regaffine_smooth_reg.nii.gz & ````` ![screen shot 2017-08-01 at 11 35 00 am](https://user-images.githubusercontent.com/7785316/28833488-7a065586-76ad-11e7-9926-0e50481ece14.png) data under: ```` /Volumes/folder_shared/sct_issues/20170801_tanguy ````
1.0
[sct_register] manage dimensionality for 2D - Register ````` isct_antsRegistration --dimensionality 2 --transform Affine[0.5] --metric MeanSquares[dest.nii.gz,src_regaffine_smooth.nii.gz,1,4] --convergence 100x100 --shrink-factors 2x2 --smoothing-sigmas 1x1vox --transform BSplineSyN[0.5,3] --metric MeanSquares[dest.nii.gz,src_regaffine_smooth.nii.gz,1,4] --convergence 500x500 --shrink-factors 2x2 --smoothing-sigmas 1x1vox --output [warp_,src_regaffine_smooth_reg.nii.gz] fslview dest.nii.gz src_regaffine_smooth_reg.nii.gz & ````` ![screen shot 2017-08-01 at 11 34 10 am](https://user-images.githubusercontent.com/7785316/28833450-5d12d03a-76ad-11e7-8ab1-f6aa4faa9639.png) Then ````` sct_concat_transfo -d dest.nii.gz -w warp_1Warp.nii.gz,warp_0GenericAffine.mat ````` and ````` sct_apply_transfo -i src_regaffine_smooth.nii.gz -d dest.nii.gz -w warp_final.nii.gz fslview dest.nii.gz src_regaffine_smooth_reg.nii.gz & ````` ![screen shot 2017-08-01 at 11 35 00 am](https://user-images.githubusercontent.com/7785316/28833488-7a065586-76ad-11e7-9926-0e50481ece14.png) data under: ```` /Volumes/folder_shared/sct_issues/20170801_tanguy ````
process
manage dimensionality for register isct antsregistration dimensionality transform affine metric meansquares convergence shrink factors smoothing sigmas transform bsplinesyn metric meansquares convergence shrink factors smoothing sigmas output fslview dest nii gz src regaffine smooth reg nii gz then sct concat transfo d dest nii gz w warp nii gz warp mat and sct apply transfo i src regaffine smooth nii gz d dest nii gz w warp final nii gz fslview dest nii gz src regaffine smooth reg nii gz data under volumes folder shared sct issues tanguy
1
4,781
7,655,563,723
IssuesEvent
2018-05-10 13:40:17
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
ess-eval-region in package directory failing until package is available
process:eval projects:package
Perhaps related to #334, but this time in R-3.2.5 and _not_ R-3.3.0. Behavior in R-3.2.5: 1. Edit a source file in an R-package directory. Ensure the package is neither _installed_ nor _loaded_ (via `dplyr::load_all`, for instance). Completely unavailable to R (at the moment). 2. Highlight _any code_ (that does not load the package). Eval it (e.g., `C-c C-r`). ``` Error in library(package, character.only = TRUE) : there is no package called 'mypackage' ``` The highlighted code is not executed. 3. `ess-eval-line-and-step` works without this problem. 4. If I load the local package into R with any method other than `ess-eval-region` (e.g., `ess-eval-line-and-step` or manually), all subsequent uses of `ess-eval-region` work without a problem. Behavior in R-3.3.0: no error, works as expected. (I believe the elisp is silently trying to load the package and its dependent namespaces into to the current R session. I make this inference on the fact that the first time I use `ess-eval-region` (in R-3.3.0 for this example), it lags a small moment and I get package-loading messages (e.g., masked functions). The second and subsequent evals do not lag.)
1.0
ess-eval-region in package directory failing until package is available - Perhaps related to #334, but this time in R-3.2.5 and _not_ R-3.3.0. Behavior in R-3.2.5: 1. Edit a source file in an R-package directory. Ensure the package is neither _installed_ nor _loaded_ (via `dplyr::load_all`, for instance). Completely unavailable to R (at the moment). 2. Highlight _any code_ (that does not load the package). Eval it (e.g., `C-c C-r`). ``` Error in library(package, character.only = TRUE) : there is no package called 'mypackage' ``` The highlighted code is not executed. 3. `ess-eval-line-and-step` works without this problem. 4. If I load the local package into R with any method other than `ess-eval-region` (e.g., `ess-eval-line-and-step` or manually), all subsequent uses of `ess-eval-region` work without a problem. Behavior in R-3.3.0: no error, works as expected. (I believe the elisp is silently trying to load the package and its dependent namespaces into to the current R session. I make this inference on the fact that the first time I use `ess-eval-region` (in R-3.3.0 for this example), it lags a small moment and I get package-loading messages (e.g., masked functions). The second and subsequent evals do not lag.)
process
ess eval region in package directory failing until package is available perhaps related to but this time in r and not r behavior in r edit a source file in an r package directory ensure the package is neither installed nor loaded via dplyr load all for instance completely unavailable to r at the moment highlight any code that does not load the package eval it e g c c c r error in library package character only true there is no package called mypackage the highlighted code is not executed ess eval line and step works without this problem if i load the local package into r with any method other than ess eval region e g ess eval line and step or manually all subsequent uses of ess eval region work without a problem behavior in r no error works as expected i believe the elisp is silently trying to load the package and its dependent namespaces into to the current r session i make this inference on the fact that the first time i use ess eval region in r for this example it lags a small moment and i get package loading messages e g masked functions the second and subsequent evals do not lag
1
816,931
30,617,419,972
IssuesEvent
2023-07-24 05:14:08
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
Proposal: Cache spliced bitstreams consistently in GCP bitstream cache
Priority:P2 Type:Cleanup Type:Question
We should figure out what we're doing with https://github.com/lowRISC/opentitan/issues/16619 In addition to the synthesized bitstream, we currently cache 1 spliced bitstream in GCP, we should consider caching all or none of them. This is related to and depends upon https://github.com/lowRISC/opentitan/issues/13807 but will let users testing software rely on updatemem more consistently.
1.0
Proposal: Cache spliced bitstreams consistently in GCP bitstream cache - We should figure out what we're doing with https://github.com/lowRISC/opentitan/issues/16619 In addition to the synthesized bitstream, we currently cache 1 spliced bitstream in GCP, we should consider caching all or none of them. This is related to and depends upon https://github.com/lowRISC/opentitan/issues/13807 but will let users testing software rely on updatemem more consistently.
non_process
proposal cache spliced bitstreams consistently in gcp bitstream cache we should figure out what we re doing with in addition to the synthesized bitstream we currently cache spliced bitstream in gcp we should consider caching all or none of them this is related to and depends upon but will let users testing software rely on updatemem more consistently
0
18,235
24,302,019,509
IssuesEvent
2022-09-29 14:30:50
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
taxon constraint incorrectly applied in P2GO?
taxon constraints multi-species process
P2GO gives an error "Extra Tax Id can only be used with terms that are descendants of GO:0044419 (interspecies interaction between organisms) or GO:0043903 (regulation of symbiosis, encompassing mutualism through parasitism) or GO:0018995 (host cellular component)" for an entity annotated with GO:0050829 defense response to Gram-negative bacterium - which, according to QuickGO is a descendant of GO:0044419 - is this a P2GO problem, or a GO problem? I've also added taxon Id for the microbe involved, which is 666 Vibrio cholerae.
1.0
taxon constraint incorrectly applied in P2GO? - P2GO gives an error "Extra Tax Id can only be used with terms that are descendants of GO:0044419 (interspecies interaction between organisms) or GO:0043903 (regulation of symbiosis, encompassing mutualism through parasitism) or GO:0018995 (host cellular component)" for an entity annotated with GO:0050829 defense response to Gram-negative bacterium - which, according to QuickGO is a descendant of GO:0044419 - is this a P2GO problem, or a GO problem? I've also added taxon Id for the microbe involved, which is 666 Vibrio cholerae.
process
taxon constraint incorrectly applied in gives an error extra tax id can only be used with terms that are descendants of go interspecies interaction between organisms or go regulation of symbiosis encompassing mutualism through parasitism or go host cellular component for an entity annotated with go defense response to gram negative bacterium which according to quickgo is a descendant of go is this a problem or a go problem i ve also added taxon id for the microbe involved which is vibrio cholerae
1
222,383
24,702,598,551
IssuesEvent
2022-10-19 16:21:22
jyom-wss-test/aspnet_accord
https://api.github.com/repos/jyom-wss-test/aspnet_accord
opened
htmlsanitizer.4.0.217.nupkg: 2 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>htmlsanitizer.4.0.217.nupkg</b></p></summary> <p>Cleans HTML from constructs that can be used for cross site scripting (XSS)</p> <p>Library home page: <a href="https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg">https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg,/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (htmlsanitizer.4.0.217.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.net.http.4.1.0.nupkg | Transitive | N/A* | &#10060; | | [CVE-2020-26293](https://www.mend.io/vulnerability-database/CVE-2020-26293) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | htmlsanitizer.4.0.217.nupkg | Direct | v5.0.372 | &#9989; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-8292</summary> ### Vulnerable Library - <b>system.net.http.4.1.0.nupkg</b></p> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.1.0.nupkg">https://api.nuget.org/packages/system.net.http.4.1.0.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.1.0/system.net.http.4.1.0.nupkg</p> <p> Dependency Hierarchy: - htmlsanitizer.4.0.217.nupkg (Root Library) - anglesharp.0.9.11.nupkg - system.net.requests.4.0.11.nupkg - :x: **system.net.http.4.1.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: Oct 10, 2018 1:29:01 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: Oct 10, 2018 1:29:01 PM</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-26293</summary> ### Vulnerable Library - <b>htmlsanitizer.4.0.217.nupkg</b></p> <p>Cleans HTML from constructs that can be used for cross site scripting (XSS)</p> <p>Library home page: <a href="https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg">https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg,/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg</p> <p> Dependency Hierarchy: - :x: **htmlsanitizer.4.0.217.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> HtmlSanitizer is a .NET library for cleaning HTML fragments and documents from constructs that can lead to XSS attacks. In HtmlSanitizer before version 5.0.372, there is a possible XSS bypass if style tag is allowed. If you have explicitly allowed the `<style>` tag, an attacker could craft HTML that includes script after passing through the sanitizer. The default settings disallow the `<style>` tag so there is no risk if you have not explicitly allowed the `<style>` tag. The problem has been fixed in version 5.0.372. <p>Publish Date: Jan 4, 2021 7:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26293>CVE-2020-26293</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26293">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26293</a></p> <p>Release Date: Jan 4, 2021 7:15:00 PM</p> <p>Fix Resolution: v5.0.372</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
htmlsanitizer.4.0.217.nupkg: 2 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>htmlsanitizer.4.0.217.nupkg</b></p></summary> <p>Cleans HTML from constructs that can be used for cross site scripting (XSS)</p> <p>Library home page: <a href="https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg">https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg,/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (htmlsanitizer.4.0.217.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.net.http.4.1.0.nupkg | Transitive | N/A* | &#10060; | | [CVE-2020-26293](https://www.mend.io/vulnerability-database/CVE-2020-26293) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | htmlsanitizer.4.0.217.nupkg | Direct | v5.0.372 | &#9989; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-8292</summary> ### Vulnerable Library - <b>system.net.http.4.1.0.nupkg</b></p> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.1.0.nupkg">https://api.nuget.org/packages/system.net.http.4.1.0.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.1.0/system.net.http.4.1.0.nupkg</p> <p> Dependency Hierarchy: - htmlsanitizer.4.0.217.nupkg (Root Library) - anglesharp.0.9.11.nupkg - system.net.requests.4.0.11.nupkg - :x: **system.net.http.4.1.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: Oct 10, 2018 1:29:01 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: Oct 10, 2018 1:29:01 PM</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-26293</summary> ### Vulnerable Library - <b>htmlsanitizer.4.0.217.nupkg</b></p> <p>Cleans HTML from constructs that can be used for cross site scripting (XSS)</p> <p>Library home page: <a href="https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg">https://api.nuget.org/packages/htmlsanitizer.4.0.217.nupkg</a></p> <p>Path to dependency file: /src/Conduit/Conduit.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg,/packages/htmlsanitizer/4.0.217/htmlsanitizer.4.0.217.nupkg</p> <p> Dependency Hierarchy: - :x: **htmlsanitizer.4.0.217.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jyom-wss-test/aspnet_accord/commit/5825ae1fe43e3c5a34179dc3c414b41c693cff82">5825ae1fe43e3c5a34179dc3c414b41c693cff82</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> HtmlSanitizer is a .NET library for cleaning HTML fragments and documents from constructs that can lead to XSS attacks. In HtmlSanitizer before version 5.0.372, there is a possible XSS bypass if style tag is allowed. If you have explicitly allowed the `<style>` tag, an attacker could craft HTML that includes script after passing through the sanitizer. The default settings disallow the `<style>` tag so there is no risk if you have not explicitly allowed the `<style>` tag. The problem has been fixed in version 5.0.372. <p>Publish Date: Jan 4, 2021 7:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26293>CVE-2020-26293</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26293">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26293</a></p> <p>Release Date: Jan 4, 2021 7:15:00 PM</p> <p>Fix Resolution: v5.0.372</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_process
htmlsanitizer nupkg vulnerabilities highest severity is vulnerable library htmlsanitizer nupkg cleans html from constructs that can be used for cross site scripting xss library home page a href path to dependency file src conduit conduit csproj path to vulnerable library home wss scanner nuget packages htmlsanitizer htmlsanitizer nupkg packages htmlsanitizer htmlsanitizer nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in htmlsanitizer nupkg version remediation available high system net http nupkg transitive n a medium htmlsanitizer nupkg direct for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file src conduit conduit csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy htmlsanitizer nupkg root library anglesharp nupkg system net requests nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch main vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date oct pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date oct pm fix resolution system net http microsoft powershell commands utility rc cve vulnerable library htmlsanitizer nupkg cleans html from constructs that can be used for cross site scripting xss library home page a href path to dependency file src conduit conduit csproj path to vulnerable library home wss scanner nuget packages htmlsanitizer htmlsanitizer nupkg packages htmlsanitizer htmlsanitizer nupkg dependency hierarchy x htmlsanitizer nupkg vulnerable library found in head commit a href found in base branch main vulnerability details htmlsanitizer is a net library for cleaning html fragments and documents from constructs that can lead to xss attacks in htmlsanitizer before version there is a possible xss bypass if style tag is allowed if you have explicitly allowed the tag an attacker could craft html that includes script after passing through the sanitizer the default settings disallow the tag so there is no risk if you have not explicitly allowed the tag the problem has been fixed in version publish date jan pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date jan pm fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
96,847
16,167,857,674
IssuesEvent
2021-05-01 21:36:08
hapifhir/hapi-fhir
https://api.github.com/repos/hapifhir/hapi-fhir
closed
CVE-2019-10086 (High) detected in commons-beanutils-1.7.0.jar, commons-beanutils-1.9.3.jar - autoclosed
security vulnerability
## CVE-2019-10086 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-beanutils-1.7.0.jar</b>, <b>commons-beanutils-1.9.3.jar</b></p></summary> <p> <details><summary><b>commons-beanutils-1.7.0.jar</b></p></summary> <p></p> <p>Path to dependency file: hapi-fhir/hapi-tinder-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) </details> <details><summary><b>commons-beanutils-1.9.3.jar</b></p></summary> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: hapi-fhir/hapi-fhir-jpaserver-cql/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - engine-1.5.1.jar (Root Library) - jaxb2-basics-0.12.0.jar - jaxb2-basics-tools-0.12.0.jar - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/983d811af6c6902f1a9bb98c2fa7b9e6e7e7b191">983d811af6c6902f1a9bb98c2fa7b9e6e7e7b191</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10086 (High) detected in commons-beanutils-1.7.0.jar, commons-beanutils-1.9.3.jar - autoclosed - ## CVE-2019-10086 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-beanutils-1.7.0.jar</b>, <b>commons-beanutils-1.9.3.jar</b></p></summary> <p> <details><summary><b>commons-beanutils-1.7.0.jar</b></p></summary> <p></p> <p>Path to dependency file: hapi-fhir/hapi-tinder-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) </details> <details><summary><b>commons-beanutils-1.9.3.jar</b></p></summary> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: hapi-fhir/hapi-fhir-jpaserver-cql/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - engine-1.5.1.jar (Root Library) - jaxb2-basics-0.12.0.jar - jaxb2-basics-tools-0.12.0.jar - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/983d811af6c6902f1a9bb98c2fa7b9e6e7e7b191">983d811af6c6902f1a9bb98c2fa7b9e6e7e7b191</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in commons beanutils jar commons beanutils jar autoclosed cve high severity vulnerability vulnerable libraries commons beanutils jar commons beanutils jar commons beanutils jar path to dependency file hapi fhir hapi tinder plugin pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy velocity tools jar root library x commons beanutils jar vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file hapi fhir hapi fhir jpaserver cql pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy engine jar root library basics jar basics tools jar x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils step up your open source security game with whitesource
0
5,126
7,894,637,666
IssuesEvent
2018-06-28 22:19:42
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
support for sparse archives in atlas
enhancement post-processor/atlas
once https://github.com/golang/go/issues/13548 is merged in, depending on if there are any API changes, we should support sparse archives. This may mean a change to atlas-go as well
1.0
support for sparse archives in atlas - once https://github.com/golang/go/issues/13548 is merged in, depending on if there are any API changes, we should support sparse archives. This may mean a change to atlas-go as well
process
support for sparse archives in atlas once is merged in depending on if there are any api changes we should support sparse archives this may mean a change to atlas go as well
1
8,948
12,058,865,173
IssuesEvent
2020-04-15 18:13:34
fablabbcn/fablabs.io
https://api.github.com/repos/fablabbcn/fablabs.io
opened
NO Ctl+Z options on backstage
Approval Process
**Describe the bug** Under the approval process, there are 4 options: <img width="332" alt="Screen Shot 2020-04-15 at 1 08 21 PM" src="https://user-images.githubusercontent.com/24419466/79371568-33040480-7f1a-11ea-8084-d5d30c908b71.png"> if anyone by accident selects the remove bottom, there is no way to go back. **To Reproduce** Steps to reproduce the behavior: 1. Go to Backstage on Labs 2. Click on Remove by error 3. There is no way to correct the error 4. See error <img width="777" alt="Screen Shot 2020-04-15 at 1 10 37 PM" src="https://user-images.githubusercontent.com/24419466/79371873-b0c81000-7f1a-11ea-99d6-8e1ca8ded9ad.png"> **Expected behavior** There should be a way to go back if you make a mistake. **Desktop (please complete the following information):** - OS: IOS - Browser: Chrome **Additional context** This means if by mistakes someone removes a lab, the lab admin will need to create a new lab application.
1.0
NO Ctl+Z options on backstage - **Describe the bug** Under the approval process, there are 4 options: <img width="332" alt="Screen Shot 2020-04-15 at 1 08 21 PM" src="https://user-images.githubusercontent.com/24419466/79371568-33040480-7f1a-11ea-8084-d5d30c908b71.png"> if anyone by accident selects the remove bottom, there is no way to go back. **To Reproduce** Steps to reproduce the behavior: 1. Go to Backstage on Labs 2. Click on Remove by error 3. There is no way to correct the error 4. See error <img width="777" alt="Screen Shot 2020-04-15 at 1 10 37 PM" src="https://user-images.githubusercontent.com/24419466/79371873-b0c81000-7f1a-11ea-99d6-8e1ca8ded9ad.png"> **Expected behavior** There should be a way to go back if you make a mistake. **Desktop (please complete the following information):** - OS: IOS - Browser: Chrome **Additional context** This means if by mistakes someone removes a lab, the lab admin will need to create a new lab application.
process
no ctl z options on backstage describe the bug under the approval process there are options img width alt screen shot at pm src if anyone by accident selects the remove bottom there is no way to go back to reproduce steps to reproduce the behavior go to backstage on labs click on remove by error there is no way to correct the error see error img width alt screen shot at pm src expected behavior there should be a way to go back if you make a mistake desktop please complete the following information os ios browser chrome additional context this means if by mistakes someone removes a lab the lab admin will need to create a new lab application
1
421,496
28,321,676,280
IssuesEvent
2023-04-11 02:04:28
Azure/aztfexport
https://api.github.com/repos/Azure/aztfexport
closed
Document update for new env variable regarding issue https://github.com/Azure/aztfexport/issues/288
documentation
Can the document please be updated to reflect this change by adding a small section for case of using AzureGov subscription?
1.0
Document update for new env variable regarding issue https://github.com/Azure/aztfexport/issues/288 - Can the document please be updated to reflect this change by adding a small section for case of using AzureGov subscription?
non_process
document update for new env variable regarding issue can the document please be updated to reflect this change by adding a small section for case of using azuregov subscription
0
13,154
15,573,365,380
IssuesEvent
2021-03-17 08:29:45
plazi/arcadia-project
https://api.github.com/repos/plazi/arcadia-project
opened
access to journals: SIB
Article processing
Hi Patrick Thanks for looking into this. Here are the closed access journals with the most treatments: - zootaxa - phytotaxa - Palaeoentomology - Annales Zoologici - Zoological Journal of the Linnean Society - Journal of Natural history Let me know, if you need more information about the journals, or whether this is enough All the best Donat ............. March 9 Hi Donat, A question regarding access to journals: Do you think, it might be possible to get a subscription at your library to the journals Zootaxa and Phytotaxa which are still an important source for treatments, but I rely on my source in the US only. Since this is a critical source, it would be good to get a second source. Sure. I am going to ask my library what are the best deal options. Only these two journals: Zootaxa and Phytotaxa ? As far as I know, access is at ETHZ or UniZ but only if you are onsite as visitor or member of the institutions via VPN. Unfortunately, we did not manage to get either. Uni Berne and the Natural History Museum Bern has neither access. Ok - fine for me. Best, Patrick
1.0
access to journals: SIB - Hi Patrick Thanks for looking into this. Here are the closed access journals with the most treatments: - zootaxa - phytotaxa - Palaeoentomology - Annales Zoologici - Zoological Journal of the Linnean Society - Journal of Natural history Let me know, if you need more information about the journals, or whether this is enough All the best Donat ............. March 9 Hi Donat, A question regarding access to journals: Do you think, it might be possible to get a subscription at your library to the journals Zootaxa and Phytotaxa which are still an important source for treatments, but I rely on my source in the US only. Since this is a critical source, it would be good to get a second source. Sure. I am going to ask my library what are the best deal options. Only these two journals: Zootaxa and Phytotaxa ? As far as I know, access is at ETHZ or UniZ but only if you are onsite as visitor or member of the institutions via VPN. Unfortunately, we did not manage to get either. Uni Berne and the Natural History Museum Bern has neither access. Ok - fine for me. Best, Patrick
process
access to journals sib hi patrick thanks for looking into this here are the closed access journals with the most treatments zootaxa phytotaxa palaeoentomology annales zoologici zoological journal of the linnean society journal of natural history let me know if you need more information about the journals or whether this is enough all the best donat march hi donat a question regarding access to journals do you think it might be possible to get a subscription at your library to the journals zootaxa and phytotaxa which are still an important source for treatments but i rely on my source in the us only since this is a critical source it would be good to get a second source sure i am going to ask my library what are the best deal options only these two journals zootaxa and phytotaxa as far as i know access is at ethz or uniz but only if you are onsite as visitor or member of the institutions via vpn unfortunately we did not manage to get either uni berne and the natural history museum bern has neither access ok fine for me best patrick
1
69,759
9,331,901,126
IssuesEvent
2019-03-28 10:50:04
visual-framework/vf-core
https://api.github.com/repos/visual-framework/vf-core
closed
DOCS - components ≠ patterns
Changelog Type: Documentation
This one comes from a conversation on slack: We've been using "components" and "patterns" fairly interchangeably. That's probably not helpful. (for example: our component generator makes files with lots of pattern text) Footnotes on how we're seeing the distinction: - _everything_ is a component, and several components can be put together to make a design(ed) pattern. - patterns would (typically) not have new code associated with it - a pattern is the recipe and components the ingredients - patterns are both collections of components _and_ usability - For organisations that might need to use different components (say: buttons), pattern function is still be relevant - tbd: patterns live largely outside of vf-core - further reading: - [Patterns ≠ Components](https://medium.com/eightshapes-llc/patterns-components-2ce778cbe4e8) - https://design-system.service.gov.uk/patterns/ and https://design-system.service.gov.uk/components/ With that in mind; two task: - [x] Go through our docs and replace many references of "patterns" with "components" (as appropriate). - [x] Add some text to `docs/guidelines/index.njk` on this; suggestion: > ## Components vs patterns > `vf-core` is a collection of modular component "ingredients" and templates that can be assembled into patterns of richer templates with usability guidance. > > For further guidance on the distinctions, see: [Patterns ≠ Components](https://medium.com/eightshapes-llc/patterns-components-2ce778cbe4e8) and UK.gov's information on [patterns](https://design-system.service.gov.uk/patterns) vs [components](https://design-system.service.gov.uk/components) I've been in the docs recently, so I'll have a first go at making a PR for this ... but please shout if you think I'm off on any of the above.
1.0
DOCS - components ≠ patterns - This one comes from a conversation on slack: We've been using "components" and "patterns" fairly interchangeably. That's probably not helpful. (for example: our component generator makes files with lots of pattern text) Footnotes on how we're seeing the distinction: - _everything_ is a component, and several components can be put together to make a design(ed) pattern. - patterns would (typically) not have new code associated with it - a pattern is the recipe and components the ingredients - patterns are both collections of components _and_ usability - For organisations that might need to use different components (say: buttons), pattern function is still be relevant - tbd: patterns live largely outside of vf-core - further reading: - [Patterns ≠ Components](https://medium.com/eightshapes-llc/patterns-components-2ce778cbe4e8) - https://design-system.service.gov.uk/patterns/ and https://design-system.service.gov.uk/components/ With that in mind; two task: - [x] Go through our docs and replace many references of "patterns" with "components" (as appropriate). - [x] Add some text to `docs/guidelines/index.njk` on this; suggestion: > ## Components vs patterns > `vf-core` is a collection of modular component "ingredients" and templates that can be assembled into patterns of richer templates with usability guidance. > > For further guidance on the distinctions, see: [Patterns ≠ Components](https://medium.com/eightshapes-llc/patterns-components-2ce778cbe4e8) and UK.gov's information on [patterns](https://design-system.service.gov.uk/patterns) vs [components](https://design-system.service.gov.uk/components) I've been in the docs recently, so I'll have a first go at making a PR for this ... but please shout if you think I'm off on any of the above.
non_process
docs components ≠ patterns this one comes from a conversation on slack we ve been using components and patterns fairly interchangeably that s probably not helpful for example our component generator makes files with lots of pattern text footnotes on how we re seeing the distinction everything is a component and several components can be put together to make a design ed pattern patterns would typically not have new code associated with it a pattern is the recipe and components the ingredients patterns are both collections of components and usability for organisations that might need to use different components say buttons pattern function is still be relevant tbd patterns live largely outside of vf core further reading and with that in mind two task go through our docs and replace many references of patterns with components as appropriate add some text to docs guidelines index njk on this suggestion components vs patterns vf core is a collection of modular component ingredients and templates that can be assembled into patterns of richer templates with usability guidance for further guidance on the distinctions see and uk gov s information on vs i ve been in the docs recently so i ll have a first go at making a pr for this but please shout if you think i m off on any of the above
0
27,414
6,848,193,039
IssuesEvent
2017-11-13 17:39:32
Daniel-Svensson/issues_before_code
https://api.github.com/repos/Daniel-Svensson/issues_before_code
closed
Duplicate 'DatabaseGenerated' attribute
CodePlexMigrationInitiated Impact: Unassigned
http://social.msdn.microsoft.com/Forums/silverlight/en-US/8106b161-a01e-4372-81ee-1a87ede9fa81/duplicate-databasegenerated-attribute?forum=silverlightwcf vs2010, silverlight 5, EF5.0 code first, WCF RiaService 1.0 SP2 public partial class Employee { [Key] public string Emp_no { get; set; } [DatabaseGenerated(DatabaseGeneratedOption.Computed)] public string CalculatedField { get; set; } public string Emp_name { get; set; } [Timestamp()] public byte[] LockID { get; set; } } Compiled EmployeeModel.Web success Compiler EmployeeModel error in directory Generated_Code, file EmployeeModel.Web.g.cs it generated Duplicate 'DatabaseGenerated' attribute error. EmployeeModel.Web.g.cs [DatabaseGenerated(DatabaseGeneratedOption.Computed)] [DatabaseGenerated(DatabaseGeneratedOption.Computed)] [DataMember()] public string CalculatedField If I remove DatabaseGenerated, then on update I get an error The column "CalculatedField" cannot be modified because it is either a computed column or is the result of a UNION operator. Any suggestion how can I resolve this issue? thanks in advance. #### This work item was migrated from CodePlex CodePlex work item ID: '13' Vote count: '2'
1.0
Duplicate 'DatabaseGenerated' attribute - http://social.msdn.microsoft.com/Forums/silverlight/en-US/8106b161-a01e-4372-81ee-1a87ede9fa81/duplicate-databasegenerated-attribute?forum=silverlightwcf vs2010, silverlight 5, EF5.0 code first, WCF RiaService 1.0 SP2 public partial class Employee { [Key] public string Emp_no { get; set; } [DatabaseGenerated(DatabaseGeneratedOption.Computed)] public string CalculatedField { get; set; } public string Emp_name { get; set; } [Timestamp()] public byte[] LockID { get; set; } } Compiled EmployeeModel.Web success Compiler EmployeeModel error in directory Generated_Code, file EmployeeModel.Web.g.cs it generated Duplicate 'DatabaseGenerated' attribute error. EmployeeModel.Web.g.cs [DatabaseGenerated(DatabaseGeneratedOption.Computed)] [DatabaseGenerated(DatabaseGeneratedOption.Computed)] [DataMember()] public string CalculatedField If I remove DatabaseGenerated, then on update I get an error The column "CalculatedField" cannot be modified because it is either a computed column or is the result of a UNION operator. Any suggestion how can I resolve this issue? thanks in advance. #### This work item was migrated from CodePlex CodePlex work item ID: '13' Vote count: '2'
non_process
duplicate databasegenerated attribute silverlight code first wcf riaservice public partial class employee public string emp no get set public string calculatedfield get set public string emp name get set public byte lockid get set compiled employeemodel web success compiler employeemodel error in directory generated code file employeemodel web g cs it generated duplicate databasegenerated attribute error employeemodel web g cs public string calculatedfield if i remove databasegenerated then on update i get an error the column calculatedfield cannot be modified because it is either a computed column or is the result of a union operator any suggestion how can i resolve this issue thanks in advance this work item was migrated from codeplex codeplex work item id vote count
0
17,841
23,779,683,890
IssuesEvent
2022-09-02 02:23:33
googleapis/nodejs-asset
https://api.github.com/repos/googleapis/nodejs-asset
closed
feat: add batchGetEffectiveIamPolicies sample code.
type: process api: cloudasset priority: p3 samples
Thanks for stopping by to let us know something could be better! Add batchGetEffectiveIamPolicies sample code and also lint the protobuf imports.
1.0
feat: add batchGetEffectiveIamPolicies sample code. - Thanks for stopping by to let us know something could be better! Add batchGetEffectiveIamPolicies sample code and also lint the protobuf imports.
process
feat add batchgeteffectiveiampolicies sample code thanks for stopping by to let us know something could be better add batchgeteffectiveiampolicies sample code and also lint the protobuf imports
1
15,910
20,115,874,430
IssuesEvent
2022-02-07 19:27:28
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
`system-tests` jobs are hanging for external contributors
process: tests process: contributing stage: needs review
See: * https://github.com/cypress-io/cypress/pull/20019 * https://github.com/cypress-io/cypress/pull/19837 * https://github.com/cypress-io/cypress/pull/19792 Something is causing these `system-tests-*` jobs to hang for external contributors, possibly a test is relying on a `RECORD_KEY` or something.
2.0
`system-tests` jobs are hanging for external contributors - See: * https://github.com/cypress-io/cypress/pull/20019 * https://github.com/cypress-io/cypress/pull/19837 * https://github.com/cypress-io/cypress/pull/19792 Something is causing these `system-tests-*` jobs to hang for external contributors, possibly a test is relying on a `RECORD_KEY` or something.
process
system tests jobs are hanging for external contributors see something is causing these system tests jobs to hang for external contributors possibly a test is relying on a record key or something
1
103,424
11,356,188,293
IssuesEvent
2020-01-24 21:59:35
trisyoungs/dissolve
https://api.github.com/repos/trisyoungs/dissolve
closed
SiO2 (Tutorial Material)
Scope: Documentation Status: Accepted
Silica tutorial - Neutron only. - Calculation of..... local tetrahedrality, angles around O atoms (dynamic sites)
1.0
SiO2 (Tutorial Material) - Silica tutorial - Neutron only. - Calculation of..... local tetrahedrality, angles around O atoms (dynamic sites)
non_process
tutorial material silica tutorial neutron only calculation of local tetrahedrality angles around o atoms dynamic sites
0
15,511
2,858,504,634
IssuesEvent
2015-06-03 03:10:25
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Problem with Inflector
Defect On hold utility
Since 2.6.5 I have problems con multi-word irregular nouns (with 3 or more words). In 2.6.4 this works fine. Example: I define (in spanish) with the Inflector::rules the plural "categorias_preguntas_frecuentes" for the singular "categoria_pregunta_frecuente"... But in the model the application looks for the table "faq_categoria_preguntas_frecuentes" instead the correct table "faq_categorias_preguntas_frecuentes". My model is into a plugin named Faq (tables start with "faq_"). I have obseved that now I need to prepend the name of the plugin into the expressions like "faq_categorias_preguntas_frecuentes" for the plural and "faq_categoria_pregunta_frecuente" for the singular in order to keep the application working... In previous versions this was not required. ¿Is this behaviour correct or is an error...? Thank you so much for your attention.
1.0
Problem with Inflector - Since 2.6.5 I have problems con multi-word irregular nouns (with 3 or more words). In 2.6.4 this works fine. Example: I define (in spanish) with the Inflector::rules the plural "categorias_preguntas_frecuentes" for the singular "categoria_pregunta_frecuente"... But in the model the application looks for the table "faq_categoria_preguntas_frecuentes" instead the correct table "faq_categorias_preguntas_frecuentes". My model is into a plugin named Faq (tables start with "faq_"). I have obseved that now I need to prepend the name of the plugin into the expressions like "faq_categorias_preguntas_frecuentes" for the plural and "faq_categoria_pregunta_frecuente" for the singular in order to keep the application working... In previous versions this was not required. ¿Is this behaviour correct or is an error...? Thank you so much for your attention.
non_process
problem with inflector since i have problems con multi word irregular nouns with or more words in this works fine example i define in spanish with the inflector rules the plural categorias preguntas frecuentes for the singular categoria pregunta frecuente but in the model the application looks for the table faq categoria preguntas frecuentes instead the correct table faq categorias preguntas frecuentes my model is into a plugin named faq tables start with faq i have obseved that now i need to prepend the name of the plugin into the expressions like faq categorias preguntas frecuentes for the plural and faq categoria pregunta frecuente for the singular in order to keep the application working in previous versions this was not required ¿is this behaviour correct or is an error thank you so much for your attention
0
48,089
13,300,929,834
IssuesEvent
2020-08-25 12:11:29
loftwah/beatsmiff.com
https://api.github.com/repos/loftwah/beatsmiff.com
opened
CVE-2012-6708 (Medium) detected in multiple libraries
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>phpdocumentor/phpdocumentor-v2.9.1</b>, <b>jquery-1.4.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary> <p> <details><summary><b>phpdocumentor/phpdocumentor-v2.9.1</b></p></summary> <p>Documentation Generator for PHP</p> <p>Library home page: <a href="https://api.github.com/repos/phpDocumentor/phpDocumentor/zipball/2e4f981a55ebe6f5db592d7da892d13d5b3c7816">https://api.github.com/repos/phpDocumentor/phpDocumentor/zipball/2e4f981a55ebe6f5db592d7da892d13d5b3c7816</a></p> <p> Dependency Hierarchy: - :x: **phpdocumentor/phpdocumentor-v2.9.1** (Vulnerable Library) </details> <details><summary><b>jquery-1.4.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.2/jquery.min.js</a></p> <p>Path to vulnerable library: /beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/old-ocean/js/jquery-1.4.2.min.js,/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/js/jquery-1.4.2.min.js,/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/abstract/js/jquery-1.4.2.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/graph.html</p> <p>Path to vulnerable library: /beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/graph.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/loftwah/beatsmiff.com/commit/c0b4db35946373bc1bf5154f2643acd2851b6129">c0b4db35946373bc1bf5154f2643acd2851b6129</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-6708 (Medium) detected in multiple libraries - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>phpdocumentor/phpdocumentor-v2.9.1</b>, <b>jquery-1.4.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary> <p> <details><summary><b>phpdocumentor/phpdocumentor-v2.9.1</b></p></summary> <p>Documentation Generator for PHP</p> <p>Library home page: <a href="https://api.github.com/repos/phpDocumentor/phpDocumentor/zipball/2e4f981a55ebe6f5db592d7da892d13d5b3c7816">https://api.github.com/repos/phpDocumentor/phpDocumentor/zipball/2e4f981a55ebe6f5db592d7da892d13d5b3c7816</a></p> <p> Dependency Hierarchy: - :x: **phpdocumentor/phpdocumentor-v2.9.1** (Vulnerable Library) </details> <details><summary><b>jquery-1.4.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.2/jquery.min.js</a></p> <p>Path to vulnerable library: /beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/old-ocean/js/jquery-1.4.2.min.js,/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/js/jquery-1.4.2.min.js,/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/abstract/js/jquery-1.4.2.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/graph.html</p> <p>Path to vulnerable library: /beatsmiff.com/wp-content/plugins/ameliabooking/vendor/microsoft/microsoft-graph/vendor/phpdocumentor/phpdocumentor/data/templates/responsive/graph.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/loftwah/beatsmiff.com/commit/c0b4db35946373bc1bf5154f2643acd2851b6129">c0b4db35946373bc1bf5154f2643acd2851b6129</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries phpdocumentor phpdocumentor jquery min js jquery min js phpdocumentor phpdocumentor documentation generator for php library home page a href dependency hierarchy x phpdocumentor phpdocumentor vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library beatsmiff com wp content plugins ameliabooking vendor microsoft microsoft graph vendor phpdocumentor phpdocumentor data templates old ocean js jquery min js beatsmiff com wp content plugins ameliabooking vendor microsoft microsoft graph vendor phpdocumentor phpdocumentor data templates responsive js jquery min js beatsmiff com wp content plugins ameliabooking vendor microsoft microsoft graph vendor phpdocumentor phpdocumentor data templates abstract js jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm beatsmiff com wp content plugins ameliabooking vendor microsoft microsoft graph vendor phpdocumentor phpdocumentor data templates responsive graph html path to vulnerable library beatsmiff com wp content plugins ameliabooking vendor microsoft microsoft graph vendor phpdocumentor phpdocumentor data templates responsive graph html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
471,704
13,608,614,939
IssuesEvent
2020-09-23 02:52:06
sopel-irc/sopel
https://api.github.com/repos/sopel-irc/sopel
closed
Quit after timeout from server
Feature Low Priority
I've run into an issue where the bot times out from the server (freenode) and on reconnecting, the bot will quit because it sees that the nick is still in use. A couple possible solutions would be to ghost the existing nick, or retry a finite number of times before disconnecting to give the server a chance to disconnect the nick.
1.0
Quit after timeout from server - I've run into an issue where the bot times out from the server (freenode) and on reconnecting, the bot will quit because it sees that the nick is still in use. A couple possible solutions would be to ghost the existing nick, or retry a finite number of times before disconnecting to give the server a chance to disconnect the nick.
non_process
quit after timeout from server i ve run into an issue where the bot times out from the server freenode and on reconnecting the bot will quit because it sees that the nick is still in use a couple possible solutions would be to ghost the existing nick or retry a finite number of times before disconnecting to give the server a chance to disconnect the nick
0
94,157
27,130,906,129
IssuesEvent
2023-02-16 09:43:07
OpenNMT/CTranslate2
https://api.github.com/repos/OpenNMT/CTranslate2
closed
Conv1D on GPU currently requires the cuDNN library which is not integrated in this build (Windows)
enhancement build gpu
When using CTranslate2 on windows, the error, ```Conv1D on GPU currently requires the cuDNN library which is not integrated in this build``` occurs. The windows system has both Cuda 11.x and cuDNN 8.x.x installed.
1.0
Conv1D on GPU currently requires the cuDNN library which is not integrated in this build (Windows) - When using CTranslate2 on windows, the error, ```Conv1D on GPU currently requires the cuDNN library which is not integrated in this build``` occurs. The windows system has both Cuda 11.x and cuDNN 8.x.x installed.
non_process
on gpu currently requires the cudnn library which is not integrated in this build windows when using on windows the error on gpu currently requires the cudnn library which is not integrated in this build occurs the windows system has both cuda x and cudnn x x installed
0
133,919
12,557,567,127
IssuesEvent
2020-06-07 13:22:19
kookmin-sw/capstone-2020-4
https://api.github.com/repos/kookmin-sw/capstone-2020-4
opened
최종 발표 자료 제작
documentation
판넬 및 책자 제작 - 이주형, 김성수 발표 영상 및 PPT 제작 - 이인평, 이태훈 수행결과보고서 작성 - 김성수, 이주형, 김민재 최종 보고서 작성 - 이인평, 이태훈
1.0
최종 발표 자료 제작 - 판넬 및 책자 제작 - 이주형, 김성수 발표 영상 및 PPT 제작 - 이인평, 이태훈 수행결과보고서 작성 - 김성수, 이주형, 김민재 최종 보고서 작성 - 이인평, 이태훈
non_process
최종 발표 자료 제작 판넬 및 책자 제작 이주형 김성수 발표 영상 및 ppt 제작 이인평 이태훈 수행결과보고서 작성 김성수 이주형 김민재 최종 보고서 작성 이인평 이태훈
0
134,282
5,223,070,610
IssuesEvent
2017-01-27 10:49:16
antoinecarme/pyaf
https://api.github.com/repos/antoinecarme/pyaf
closed
Add a Jupyter Notebook to demonstrate the use of Hierarchical Forecasting
class:Doc priority:high topic:hierarchical topic:user_support
need a similar doc with hierarchical and grouped time series detailed examples.
1.0
Add a Jupyter Notebook to demonstrate the use of Hierarchical Forecasting - need a similar doc with hierarchical and grouped time series detailed examples.
non_process
add a jupyter notebook to demonstrate the use of hierarchical forecasting need a similar doc with hierarchical and grouped time series detailed examples
0
12,234
9,659,623,871
IssuesEvent
2019-05-20 13:51:14
maidsafe/safe-build-infrastructure
https://api.github.com/repos/maidsafe/safe-build-infrastructure
opened
External Disk for Windows Slaves
enhancement infrastructure
The default 8GB disk most likely won't be enough for these slaves. Need to extend this to add another disk and put the Jenkins slave workspace on this disk. The difficulty here will be the automation for formatting the disk, if that doesn't happen automatically.
1.0
External Disk for Windows Slaves - The default 8GB disk most likely won't be enough for these slaves. Need to extend this to add another disk and put the Jenkins slave workspace on this disk. The difficulty here will be the automation for formatting the disk, if that doesn't happen automatically.
non_process
external disk for windows slaves the default disk most likely won t be enough for these slaves need to extend this to add another disk and put the jenkins slave workspace on this disk the difficulty here will be the automation for formatting the disk if that doesn t happen automatically
0
95,955
10,914,399,612
IssuesEvent
2019-11-21 09:06:51
GSG-G7/far-nearer
https://api.github.com/repos/GSG-G7/far-nearer
closed
Receive newsletter and contact the project owners
documentation user-story
#11 As a user, I can subscribe to the mailing list and be able to contact the project owners
1.0
Receive newsletter and contact the project owners - #11 As a user, I can subscribe to the mailing list and be able to contact the project owners
non_process
receive newsletter and contact the project owners as a user i can subscribe to the mailing list and be able to contact the project owners
0
10,856
13,630,429,746
IssuesEvent
2020-09-24 16:25:50
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Node must de-duplicate environment variables before calling CreateProcess (and upon env.extend())
duplicate process windows
* **Version**: 12.9.1 * **Platform**: x64 * **Subsystem**: Windows 10 ### What steps will reproduce the bug? yarn sets up some environment variables (like `npm_config_cache`) and calls execSync. This calls the Win32 API `CreateProcess` and passes an environment block which is a merge of the existing environment block (i.e. from the calling process) with some additional stuff that includes `npm_config_cache` among others. Environment variables in Windows are case-insensitive, but CreateProcess does not seem to sanitize the environment block. This is a problem because if the calling process had set up `NPM_CONFIG_CACHE` (all caps), then yarn will just set the lowercase variant of the variable, and node will call `CreateProcess` which will contain the variable twice). This breaks some tools that create dictionaries out of the environment block because they expect to run into each variable exactly once (regardless of casing). In my case we have a nodejs script that we invoke via yarn, and the nodejs script calls a Windows build utility (msbuild) which crashes when trying to set up options for the compiler when it finds this option twice. We run into this when building in our CI in Azure DevOps pipeline, which sets up NPM_CONFIG_CACHE in the environment. Related: - https://github.com/actions/virtual-environments/pull/1566/files - https://github.com/dotnet/msbuild/issues/5726 ### How often does it reproduce? Is there a required condition? 100% ### What is the expected behavior? node should de-duplicate variables before passing them to CreateProcess ### What do you see instead? CreateProcess gets an environment block with both variables: ![image](https://user-images.githubusercontent.com/22989529/92638930-f1d90c00-f28f-11ea-821f-8dc5804357c3.png) ### Additional information <!-- Tell us anything else you think we should know. -->
1.0
Node must de-duplicate environment variables before calling CreateProcess (and upon env.extend()) - * **Version**: 12.9.1 * **Platform**: x64 * **Subsystem**: Windows 10 ### What steps will reproduce the bug? yarn sets up some environment variables (like `npm_config_cache`) and calls execSync. This calls the Win32 API `CreateProcess` and passes an environment block which is a merge of the existing environment block (i.e. from the calling process) with some additional stuff that includes `npm_config_cache` among others. Environment variables in Windows are case-insensitive, but CreateProcess does not seem to sanitize the environment block. This is a problem because if the calling process had set up `NPM_CONFIG_CACHE` (all caps), then yarn will just set the lowercase variant of the variable, and node will call `CreateProcess` which will contain the variable twice). This breaks some tools that create dictionaries out of the environment block because they expect to run into each variable exactly once (regardless of casing). In my case we have a nodejs script that we invoke via yarn, and the nodejs script calls a Windows build utility (msbuild) which crashes when trying to set up options for the compiler when it finds this option twice. We run into this when building in our CI in Azure DevOps pipeline, which sets up NPM_CONFIG_CACHE in the environment. Related: - https://github.com/actions/virtual-environments/pull/1566/files - https://github.com/dotnet/msbuild/issues/5726 ### How often does it reproduce? Is there a required condition? 100% ### What is the expected behavior? node should de-duplicate variables before passing them to CreateProcess ### What do you see instead? CreateProcess gets an environment block with both variables: ![image](https://user-images.githubusercontent.com/22989529/92638930-f1d90c00-f28f-11ea-821f-8dc5804357c3.png) ### Additional information <!-- Tell us anything else you think we should know. -->
process
node must de duplicate environment variables before calling createprocess and upon env extend version platform subsystem windows what steps will reproduce the bug yarn sets up some environment variables like npm config cache and calls execsync this calls the api createprocess and passes an environment block which is a merge of the existing environment block i e from the calling process with some additional stuff that includes npm config cache among others environment variables in windows are case insensitive but createprocess does not seem to sanitize the environment block this is a problem because if the calling process had set up npm config cache all caps then yarn will just set the lowercase variant of the variable and node will call createprocess which will contain the variable twice this breaks some tools that create dictionaries out of the environment block because they expect to run into each variable exactly once regardless of casing in my case we have a nodejs script that we invoke via yarn and the nodejs script calls a windows build utility msbuild which crashes when trying to set up options for the compiler when it finds this option twice we run into this when building in our ci in azure devops pipeline which sets up npm config cache in the environment related how often does it reproduce is there a required condition what is the expected behavior node should de duplicate variables before passing them to createprocess what do you see instead createprocess gets an environment block with both variables additional information tell us anything else you think we should know
1
17,840
23,776,778,234
IssuesEvent
2022-09-01 21:54:05
googleapis/java-storage
https://api.github.com/repos/googleapis/java-storage
closed
refactor integration tests
type: process api: storage
Integration tests need to be split out into respective Test files with only what's required to run them. We have the issue that all integration tests are within one file and coupled. This will also lay the ground work for multi-transport testing.
1.0
refactor integration tests - Integration tests need to be split out into respective Test files with only what's required to run them. We have the issue that all integration tests are within one file and coupled. This will also lay the ground work for multi-transport testing.
process
refactor integration tests integration tests need to be split out into respective test files with only what s required to run them we have the issue that all integration tests are within one file and coupled this will also lay the ground work for multi transport testing
1
19,978
26,458,726,899
IssuesEvent
2023-01-16 15:55:18
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
`InstrumentedProcessError` with RdFault when generating tests with `ContestEstimator` for `spoon-core-7.0.0`
ctg-bug comp-contest-estimator comp-instrumented-process
**Description** Errors in instrumented process caused by RdFault for a `spoon` libarary method. **To Reproduce** 1. Run the 'utbot' project in IntelliJ Idea 2022.2.4 2. Open `utbot-junit-contest/src/main/kotlin/org/utbot/contest/ContestEstimator.kt` 3. Change `javaHome` property to full local path to JDK 8 4. In `fun main` change: `projectFilter = listOf("spoon-core-7.0.0")` 5. Run `fun main` - make sure to set JRE in Edit Configurations to JDK 8. 6. Check the logs **Expected behavior** Test generation should not fail with Instrumented process errors. **Actual behavior** Errors in instrumented process caused by RdFault for `spoon.support.reflect.code.CtAssignmentImpl.setType` method. **Visual proofs (screenshots, logs, images)** The following exceptions are present: ```java 13:16:36.074 | INFO | Started: method spoon.support.reflect.code.CtAssignmentImpl.setType 13:16:36.074 | INFO | -- Remaining time budget: 33422 ms, #remaining_methods: 13, budget for method: 2570 ms, solver timeout: 1000 ms, budget for symbolic execution: 1370 ms, budget for concrete execution: 1200 ms, -- 13:16:37.447 | INFO | |> Starting concrete execution for remaining state: spoon.support.reflect.code.CtAssignmentImpl.setType 13:16:37.671 | ERROR | Symbolic execution FAILED org.utbot.instrumentation.util.InstrumentedProcessError: Error in the instrumented process |> com.jetbrains.rd.util.reactive.RdFault: spoon.support.reflect.declaration.CtElementImpl, reason: java.lang.NoClassDefFoundError: spoon.support.reflect.declaration.CtElementImpl at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2729) at java.lang.Class.getDeclaredMethods(Class.java:2003) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:79) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:78) at kotlin.sequences.TransformingSequence$iterator$1.next(Sequences.kt:210) at kotlin.sequences.FilteringSequence$iterator$1.calcNext(Sequences.kt:170) at kotlin.sequences.FilteringSequence$iterator$1.hasNext(Sequences.kt:194) at kotlin.sequences.SequencesKt___SequencesKt.firstOrNull(_Sequences.kt:168) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleMethodOrNull(SignatureUtil.kt:80) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableIdOrNull(SignatureUtil.kt:102) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableId(SignatureUtil.kt:94) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:126) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:96) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:128) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:124) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1$1.invoke(ClientProcessUtil.kt:96) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:83) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1.invoke(ClientProcessUtil.kt:95) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at org.utbot.instrumentation.ConcreteExecutor.withProcess(ConcreteExecutor.kt:227) ~[main/:?] at org.utbot.instrumentation.ConcreteExecutor$withProcess$1.invokeSuspend(ConcreteExecutor.kt) ~[main/:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) [kotlin-stdlib-1.7.20.jar:1.7.20-release-201(1.7.20)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:104) [kotlinx-coroutines-core-jvm-1.6.3.jar:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_332] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_332] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) [?:1.8.0_332] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) [?:1.8.0_332] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_332] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_332] at java.lang.Thread.run(Thread.java:750) [?:1.8.0_332] Caused by: com.jetbrains.rd.util.reactive.RdFault: spoon.support.reflect.declaration.CtElementImpl, reason: java.lang.NoClassDefFoundError: spoon.support.reflect.declaration.CtElementImpl at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2729) at java.lang.Class.getDeclaredMethods(Class.java:2003) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:79) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:78) at kotlin.sequences.TransformingSequence$iterator$1.next(Sequences.kt:210) at kotlin.sequences.FilteringSequence$iterator$1.calcNext(Sequences.kt:170) at kotlin.sequences.FilteringSequence$iterator$1.hasNext(Sequences.kt:194) at kotlin.sequences.SequencesKt___SequencesKt.firstOrNull(_Sequences.kt:168) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleMethodOrNull(SignatureUtil.kt:80) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableIdOrNull(SignatureUtil.kt:102) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableId(SignatureUtil.kt:94) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:126) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:96) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:128) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:124) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1$1.invoke(ClientProcessUtil.kt:96) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:83) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1.invoke(ClientProcessUtil.kt:95) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) ~[rd-framework-2022.2.1.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) ~[kotlin-stdlib-1.7.20.jar:1.7.20-release-201(1.7.20)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) ~[kotlinx-coroutines-core-jvm-1.6.3.jar:?] ... 7 more 13:16:37.673 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.674 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.676 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.678 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.745 | INFO | Finished (in 1.670 sec): method spoon.support.reflect.code.CtAssignmentImpl.setType <StatsForMethod> : #generatedTC=0 FAILED 5 time(s) with 1 different exception(s)" ``` **Environment** Windows 10 Pro IntelliJ IDEA 2022.2.4 JDK 8
1.0
`InstrumentedProcessError` with RdFault when generating tests with `ContestEstimator` for `spoon-core-7.0.0` - **Description** Errors in instrumented process caused by RdFault for a `spoon` libarary method. **To Reproduce** 1. Run the 'utbot' project in IntelliJ Idea 2022.2.4 2. Open `utbot-junit-contest/src/main/kotlin/org/utbot/contest/ContestEstimator.kt` 3. Change `javaHome` property to full local path to JDK 8 4. In `fun main` change: `projectFilter = listOf("spoon-core-7.0.0")` 5. Run `fun main` - make sure to set JRE in Edit Configurations to JDK 8. 6. Check the logs **Expected behavior** Test generation should not fail with Instrumented process errors. **Actual behavior** Errors in instrumented process caused by RdFault for `spoon.support.reflect.code.CtAssignmentImpl.setType` method. **Visual proofs (screenshots, logs, images)** The following exceptions are present: ```java 13:16:36.074 | INFO | Started: method spoon.support.reflect.code.CtAssignmentImpl.setType 13:16:36.074 | INFO | -- Remaining time budget: 33422 ms, #remaining_methods: 13, budget for method: 2570 ms, solver timeout: 1000 ms, budget for symbolic execution: 1370 ms, budget for concrete execution: 1200 ms, -- 13:16:37.447 | INFO | |> Starting concrete execution for remaining state: spoon.support.reflect.code.CtAssignmentImpl.setType 13:16:37.671 | ERROR | Symbolic execution FAILED org.utbot.instrumentation.util.InstrumentedProcessError: Error in the instrumented process |> com.jetbrains.rd.util.reactive.RdFault: spoon.support.reflect.declaration.CtElementImpl, reason: java.lang.NoClassDefFoundError: spoon.support.reflect.declaration.CtElementImpl at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2729) at java.lang.Class.getDeclaredMethods(Class.java:2003) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:79) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:78) at kotlin.sequences.TransformingSequence$iterator$1.next(Sequences.kt:210) at kotlin.sequences.FilteringSequence$iterator$1.calcNext(Sequences.kt:170) at kotlin.sequences.FilteringSequence$iterator$1.hasNext(Sequences.kt:194) at kotlin.sequences.SequencesKt___SequencesKt.firstOrNull(_Sequences.kt:168) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleMethodOrNull(SignatureUtil.kt:80) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableIdOrNull(SignatureUtil.kt:102) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableId(SignatureUtil.kt:94) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:126) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:96) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:128) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:124) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1$1.invoke(ClientProcessUtil.kt:96) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:83) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1.invoke(ClientProcessUtil.kt:95) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at org.utbot.instrumentation.ConcreteExecutor.withProcess(ConcreteExecutor.kt:227) ~[main/:?] at org.utbot.instrumentation.ConcreteExecutor$withProcess$1.invokeSuspend(ConcreteExecutor.kt) ~[main/:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) [kotlin-stdlib-1.7.20.jar:1.7.20-release-201(1.7.20)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:104) [kotlinx-coroutines-core-jvm-1.6.3.jar:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_332] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_332] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) [?:1.8.0_332] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) [?:1.8.0_332] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_332] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_332] at java.lang.Thread.run(Thread.java:750) [?:1.8.0_332] Caused by: com.jetbrains.rd.util.reactive.RdFault: spoon.support.reflect.declaration.CtElementImpl, reason: java.lang.NoClassDefFoundError: spoon.support.reflect.declaration.CtElementImpl at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2729) at java.lang.Class.getDeclaredMethods(Class.java:2003) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:79) at org.utbot.framework.plugin.api.util.SignatureUtilKt$singleMethodOrNull$2.invoke(SignatureUtil.kt:78) at kotlin.sequences.TransformingSequence$iterator$1.next(Sequences.kt:210) at kotlin.sequences.FilteringSequence$iterator$1.calcNext(Sequences.kt:170) at kotlin.sequences.FilteringSequence$iterator$1.hasNext(Sequences.kt:194) at kotlin.sequences.SequencesKt___SequencesKt.firstOrNull(_Sequences.kt:168) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleMethodOrNull(SignatureUtil.kt:80) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableIdOrNull(SignatureUtil.kt:102) at org.utbot.framework.plugin.api.util.SignatureUtilKt.singleExecutableId(SignatureUtil.kt:94) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:126) at org.utbot.framework.concrete.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:96) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:128) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:124) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1$1.invoke(ClientProcessUtil.kt:96) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:83) at org.utbot.rd.IdleWatchdog$wrapActiveCall$1.invoke(ClientProcessUtil.kt:95) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) ~[rd-framework-2022.2.1.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) ~[kotlin-stdlib-1.7.20.jar:1.7.20-release-201(1.7.20)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) ~[kotlinx-coroutines-core-jvm-1.6.3.jar:?] ... 7 more 13:16:37.673 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.674 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.676 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.678 | ERROR | Symbolic execution FAILED ... <<stack trace duplicated>> 13:16:37.745 | INFO | Finished (in 1.670 sec): method spoon.support.reflect.code.CtAssignmentImpl.setType <StatsForMethod> : #generatedTC=0 FAILED 5 time(s) with 1 different exception(s)" ``` **Environment** Windows 10 Pro IntelliJ IDEA 2022.2.4 JDK 8
process
instrumentedprocesserror with rdfault when generating tests with contestestimator for spoon core description errors in instrumented process caused by rdfault for a spoon libarary method to reproduce run the utbot project in intellij idea open utbot junit contest src main kotlin org utbot contest contestestimator kt change javahome property to full local path to jdk in fun main change projectfilter listof spoon core run fun main make sure to set jre in edit configurations to jdk check the logs expected behavior test generation should not fail with instrumented process errors actual behavior errors in instrumented process caused by rdfault for spoon support reflect code ctassignmentimpl settype method visual proofs screenshots logs images the following exceptions are present java info started method spoon support reflect code ctassignmentimpl settype info remaining time budget ms remaining methods budget for method ms solver timeout ms budget for symbolic execution ms budget for concrete execution ms info starting concrete execution for remaining state spoon support reflect code ctassignmentimpl settype error symbolic execution failed org utbot instrumentation util instrumentedprocesserror error in the instrumented process com jetbrains rd util reactive rdfault spoon support reflect declaration ctelementimpl reason java lang noclassdeffounderror spoon support reflect declaration ctelementimpl at java lang class native method at java lang class privategetdeclaredmethods class java at java lang class getdeclaredmethods class java at org utbot framework plugin api util signatureutilkt singlemethodornull invoke signatureutil kt at org utbot framework plugin api util signatureutilkt singlemethodornull invoke signatureutil kt at kotlin sequences transformingsequence iterator next sequences kt at kotlin sequences filteringsequence iterator calcnext sequences kt at kotlin sequences filteringsequence iterator hasnext sequences kt at kotlin sequences sequenceskt sequenceskt firstornull sequences kt at org utbot framework plugin api util signatureutilkt singlemethodornull signatureutil kt at org utbot framework plugin api util signatureutilkt singleexecutableidornull signatureutil kt at org utbot framework plugin api util signatureutilkt singleexecutableid signatureutil kt at org utbot framework concrete utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot framework concrete utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot rd idlewatchdog wrapactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog wrapactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com jetbrains rd framework rdtaskresult companion read taskinterfaces kt at com jetbrains rd framework impl callsitewiredrdtask onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invoke rdtask kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invokesuspend rdtask kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at org utbot instrumentation concreteexecutor withprocess concreteexecutor kt at org utbot instrumentation concreteexecutor withprocess invokesuspend concreteexecutor kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by com jetbrains rd util reactive rdfault spoon support reflect declaration ctelementimpl reason java lang noclassdeffounderror spoon support reflect declaration ctelementimpl at java lang class native method at java lang class privategetdeclaredmethods class java at java lang class getdeclaredmethods class java at org utbot framework plugin api util signatureutilkt singlemethodornull invoke signatureutil kt at org utbot framework plugin api util signatureutilkt singlemethodornull invoke signatureutil kt at kotlin sequences transformingsequence iterator next sequences kt at kotlin sequences filteringsequence iterator calcnext sequences kt at kotlin sequences filteringsequence iterator hasnext sequences kt at kotlin sequences sequenceskt sequenceskt firstornull sequences kt at org utbot framework plugin api util signatureutilkt singlemethodornull signatureutil kt at org utbot framework plugin api util signatureutilkt singleexecutableidornull signatureutil kt at org utbot framework plugin api util signatureutilkt singleexecutableid signatureutil kt at org utbot framework concrete utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot framework concrete utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot rd idlewatchdog wrapactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog wrapactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com jetbrains rd framework rdtaskresult companion read taskinterfaces kt at com jetbrains rd framework impl callsitewiredrdtask onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invoke rdtask kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invokesuspend rdtask kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt more error symbolic execution failed error symbolic execution failed error symbolic execution failed error symbolic execution failed info finished in sec method spoon support reflect code ctassignmentimpl settype generatedtc failed time s with different exception s environment windows pro intellij idea jdk
1
7,801
10,959,607,085
IssuesEvent
2019-11-27 11:47:22
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Will this also work for VM Scale Sets?
Pri1 automation/svc cxp process-automation/subsvc product-issue triaged
i see that the vm scale set is being targeted but it fails when the automation occurs. et-AutomationConnection : Connections asset not found. To create this Connections asset, navigate to the Assets blade and create a Connections asset named: AzureClassicRunAsConnection.At line:183 char:27+ ... $connection = Get-AutomationConnection -Name $connectionAssetName+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (:) [Get-AutomationConnection], AssetManagementClientException + FullyQualifiedErrorId : 3,Orchestrator.AssetManagement.Cmdlets.GetAutomationConnectionCmdlet --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#feedback) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
Will this also work for VM Scale Sets? - i see that the vm scale set is being targeted but it fails when the automation occurs. et-AutomationConnection : Connections asset not found. To create this Connections asset, navigate to the Assets blade and create a Connections asset named: AzureClassicRunAsConnection.At line:183 char:27+ ... $connection = Get-AutomationConnection -Name $connectionAssetName+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (:) [Get-AutomationConnection], AssetManagementClientException + FullyQualifiedErrorId : 3,Orchestrator.AssetManagement.Cmdlets.GetAutomationConnectionCmdlet --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#feedback) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
will this also work for vm scale sets i see that the vm scale set is being targeted but it fails when the automation occurs et automationconnection connections asset not found to create this connections asset navigate to the assets blade and create a connections asset named azureclassicrunasconnection at line char connection get automationconnection name connectionassetname categoryinfo objectnotfound assetmanagementclientexception fullyqualifiederrorid orchestrator assetmanagement cmdlets getautomationconnectioncmdlet document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
68,134
7,088,260,134
IssuesEvent
2018-01-11 20:50:46
Vizzuality/half-earth
https://api.github.com/repos/Vizzuality/half-earth
closed
New styles for barcharts
difficulty:easy ready-to-test-staging
Several spider-charts have been replaced by bar-charts, the bar=chart designs are slightly different from what we already had, this should be adjusted.
1.0
New styles for barcharts - Several spider-charts have been replaced by bar-charts, the bar=chart designs are slightly different from what we already had, this should be adjusted.
non_process
new styles for barcharts several spider charts have been replaced by bar charts the bar chart designs are slightly different from what we already had this should be adjusted
0
117,831
4,728,255,568
IssuesEvent
2016-10-18 15:30:54
default1406/PhyLab
https://api.github.com/repos/default1406/PhyLab
closed
协商预估PhyLab后端开发时长
priority 3 size 2
- 对象:冯炜韬 - 预计时长:1.5h - 详情:接手PhyLab项目后,需要进行二次开发,在确定相应的功能与改进方向后,对其进行初步的模块开发时长预估,并与邓楚云达成一致。为编码阶段任务布置做好初步规划。 - 提交:各模块开发时长信息文档
1.0
协商预估PhyLab后端开发时长 - - 对象:冯炜韬 - 预计时长:1.5h - 详情:接手PhyLab项目后,需要进行二次开发,在确定相应的功能与改进方向后,对其进行初步的模块开发时长预估,并与邓楚云达成一致。为编码阶段任务布置做好初步规划。 - 提交:各模块开发时长信息文档
non_process
协商预估phylab后端开发时长 对象:冯炜韬 预计时长: 详情:接手phylab项目后,需要进行二次开发,在确定相应的功能与改进方向后,对其进行初步的模块开发时长预估,并与邓楚云达成一致。为编码阶段任务布置做好初步规划。 提交:各模块开发时长信息文档
0
13,811
16,574,668,340
IssuesEvent
2021-05-31 01:22:45
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Merge ui-configuration and recording scripts [3]
Data Processing Sprint 2 Week 1
Combine already created scripts for setting camera/ui on league replay client and recording.
1.0
Merge ui-configuration and recording scripts [3] - Combine already created scripts for setting camera/ui on league replay client and recording.
process
merge ui configuration and recording scripts combine already created scripts for setting camera ui on league replay client and recording
1
10,949
13,756,468,115
IssuesEvent
2020-10-06 19:59:13
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Spatial join summary processing alg report a TypeError (feedback) if no spatial index in batch process
Bug Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Spatial join summary report a type Error if no spatial index in batch process but not in single process. Batch process > Algorithme de traitement 1/1… > Démarrage de l'algorithme Joindre les attributs par localisation (résumé)… > Paramètres en entrée: > {'DISCARD_NONMATCHING': True, > 'INPUT': 'Estimation_zones_protegees_2762776f_1221_4a37_ab7e_455e2c9d3d66', > 'JOIN': 'Base_Entreprises_4877a359_5fbc_444a_b161_76a2ae26d31d', > 'JOIN_FIELDS': ['EffSIRENE'], > 'OUTPUT': <QgsProcessingOutputLayerDefinition {'sink':[...]/entzpp.shp, 'createOptions': {}}>, > 'PREDICATE': [0, 1, 2, 3, 4, 5, 6], > 'SUMMARIES': [5]} > > Traceback (most recent call last): > File "D:/Qgis/apps/qgis/./python/plugins\processing\algs\qgis\SpatialJoinSummary.py", line 165, in processAlgorithm > feedback.reportError(self.tr("No spatial index exists for join layer, performance will be severely degraded")) > TypeError: reportError() missing 1 required positional argument: 'fatalError' > > L'algorithme Joindre les attributs par localisation (résumé) a échoué… > Execution failed after 0.01 seconds > Batch execution completed in 0.03 seconds > 1 l'exécution a échoué. Consultez le journal pour plus de détails. Single process > Traitement de l'algorithme… > Démarrage de l'algorithme 'Joindre les attributs par localisation (résumé)'… > Paramètres en entrée: > { 'DISCARD_NONMATCHING' : True, 'INPUT' : '[...]/Estimation_zones_protegees.shp', 'JOIN' : '[...]/Base_Entreprises.shp', 'JOIN_FIELDS' : ['EffSIRENE'], 'OUTPUT' : '[...]/entzpp.shp', 'PREDICATE' : [0,1,2,3,4,5,6], 'SUMMARIES' : [5] } > > No spatial index exists for join layer, performance will be severely degraded > Exécution terminée en 0.29 secondes > Résultats: > {'OUTPUT': '[...]/entzpp.shp'} > > Chargement des couches de résultat > Algorithme 'Joindre les attributs par localisation (résumé)' terminé **QGIS and OS versions** Qgis 3.14.16-2 Windows 10 <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. -->
1.0
Spatial join summary processing alg report a TypeError (feedback) if no spatial index in batch process - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Spatial join summary report a type Error if no spatial index in batch process but not in single process. Batch process > Algorithme de traitement 1/1… > Démarrage de l'algorithme Joindre les attributs par localisation (résumé)… > Paramètres en entrée: > {'DISCARD_NONMATCHING': True, > 'INPUT': 'Estimation_zones_protegees_2762776f_1221_4a37_ab7e_455e2c9d3d66', > 'JOIN': 'Base_Entreprises_4877a359_5fbc_444a_b161_76a2ae26d31d', > 'JOIN_FIELDS': ['EffSIRENE'], > 'OUTPUT': <QgsProcessingOutputLayerDefinition {'sink':[...]/entzpp.shp, 'createOptions': {}}>, > 'PREDICATE': [0, 1, 2, 3, 4, 5, 6], > 'SUMMARIES': [5]} > > Traceback (most recent call last): > File "D:/Qgis/apps/qgis/./python/plugins\processing\algs\qgis\SpatialJoinSummary.py", line 165, in processAlgorithm > feedback.reportError(self.tr("No spatial index exists for join layer, performance will be severely degraded")) > TypeError: reportError() missing 1 required positional argument: 'fatalError' > > L'algorithme Joindre les attributs par localisation (résumé) a échoué… > Execution failed after 0.01 seconds > Batch execution completed in 0.03 seconds > 1 l'exécution a échoué. Consultez le journal pour plus de détails. Single process > Traitement de l'algorithme… > Démarrage de l'algorithme 'Joindre les attributs par localisation (résumé)'… > Paramètres en entrée: > { 'DISCARD_NONMATCHING' : True, 'INPUT' : '[...]/Estimation_zones_protegees.shp', 'JOIN' : '[...]/Base_Entreprises.shp', 'JOIN_FIELDS' : ['EffSIRENE'], 'OUTPUT' : '[...]/entzpp.shp', 'PREDICATE' : [0,1,2,3,4,5,6], 'SUMMARIES' : [5] } > > No spatial index exists for join layer, performance will be severely degraded > Exécution terminée en 0.29 secondes > Résultats: > {'OUTPUT': '[...]/entzpp.shp'} > > Chargement des couches de résultat > Algorithme 'Joindre les attributs par localisation (résumé)' terminé **QGIS and OS versions** Qgis 3.14.16-2 Windows 10 <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. -->
process
spatial join summary processing alg report a typeerror feedback if no spatial index in batch process bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug spatial join summary report a type error if no spatial index in batch process but not in single process batch process algorithme de traitement … démarrage de l algorithme joindre les attributs par localisation résumé … paramètres en entrée discard nonmatching true input estimation zones protegees join base entreprises join fields output predicate summaries traceback most recent call last file d qgis apps qgis python plugins processing algs qgis spatialjoinsummary py line in processalgorithm feedback reporterror self tr no spatial index exists for join layer performance will be severely degraded typeerror reporterror missing required positional argument fatalerror l algorithme joindre les attributs par localisation résumé a échoué… execution failed after seconds batch execution completed in seconds l exécution a échoué consultez le journal pour plus de détails single process traitement de l algorithme… démarrage de l algorithme joindre les attributs par localisation résumé … paramètres en entrée discard nonmatching true input estimation zones protegees shp join base entreprises shp join fields output entzpp shp predicate summaries no spatial index exists for join layer performance will be severely degraded exécution terminée en secondes résultats output entzpp shp chargement des couches de résultat algorithme joindre les attributs par localisation résumé terminé qgis and os versions qgis windows about click in the table ctrl a and then ctrl c finally paste here additional context
1
331,142
24,293,279,445
IssuesEvent
2022-09-29 08:04:28
Oxford-Dynamics-Repository/mindglass
https://api.github.com/repos/Oxford-Dynamics-Repository/mindglass
opened
Add all documentation on Sharepoint to the corresponding READMEs.
documentation
We have a lot of presentation and documents related to architectures and installation setups. Find them all and put a link on the corresponding README.
1.0
Add all documentation on Sharepoint to the corresponding READMEs. - We have a lot of presentation and documents related to architectures and installation setups. Find them all and put a link on the corresponding README.
non_process
add all documentation on sharepoint to the corresponding readmes we have a lot of presentation and documents related to architectures and installation setups find them all and put a link on the corresponding readme
0
28,319
13,637,835,233
IssuesEvent
2020-09-25 08:28:03
eclipse/vorto
https://api.github.com/repos/eclipse/vorto
opened
Optimize model mapping retrieval
performance
This is a spin off of #2599. Similar situation but different UI. Querying model mappings can take an extravagant amount of time. I suppose some fork/join implementation could help in `ModelRepository#getMappingModelsForTargetPlatform`, keeping in mind the usual requirements (injection of security context and request attributes, bypassing JRC session helper until fixed for concurrent usage...).
True
Optimize model mapping retrieval - This is a spin off of #2599. Similar situation but different UI. Querying model mappings can take an extravagant amount of time. I suppose some fork/join implementation could help in `ModelRepository#getMappingModelsForTargetPlatform`, keeping in mind the usual requirements (injection of security context and request attributes, bypassing JRC session helper until fixed for concurrent usage...).
non_process
optimize model mapping retrieval this is a spin off of similar situation but different ui querying model mappings can take an extravagant amount of time i suppose some fork join implementation could help in modelrepository getmappingmodelsfortargetplatform keeping in mind the usual requirements injection of security context and request attributes bypassing jrc session helper until fixed for concurrent usage
0
21,215
28,293,363,014
IssuesEvent
2023-04-09 13:46:41
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
reopened
[C++] Nightly Integration Testing Report
type: process nightly-testing
Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)** *** <hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sun Apr 9 03:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4649723451)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 11:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)** <hidden value="integration-test-status-comment"></hidden>
1.0
[C++] Nightly Integration Testing Report - Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)** *** <hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sun Apr 9 03:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4649723451)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 11:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)** <hidden value="integration-test-status-comment"></hidden>
process
nightly integration testing report note this report excludes firestore please also check ✅ nbsp integration test succeeded requested by dellabitta on commit last updated sun apr pdt ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated sat apr pdt
1
14,992
18,670,805,929
IssuesEvent
2021-10-30 17:15:17
bgyori/indra_cogex
https://api.github.com/repos/bgyori/indra_cogex
closed
ChEMBL Indication Processor
Processor
This SQL statement can be used with the `chembl_downloader` to get links between chemicals and their indications (either as mesh or EFO/HP) and their maximum clinical phase: ```sql SELECT MOLECULE_DICTIONARY.chembl_id, MOLECULE_DICTIONARY.pref_name, DRUG_INDICATION.mesh_id, DRUG_INDICATION.mesh_heading, DRUG_INDICATION.efo_id, DRUG_INDICATION.efo_term, DRUG_INDICATION.max_phase_for_ind FROM MOLECULE_DICTIONARY JOIN DRUG_INDICATION ON MOLECULE_DICTIONARY.molregno == DRUG_INDICATION.molregno ``` There's another column in `MOLECULE_DICTIONARY` called `chebi_par_id` but a spot check of the first 10 rows showed this doesn't have good coverage
1.0
ChEMBL Indication Processor - This SQL statement can be used with the `chembl_downloader` to get links between chemicals and their indications (either as mesh or EFO/HP) and their maximum clinical phase: ```sql SELECT MOLECULE_DICTIONARY.chembl_id, MOLECULE_DICTIONARY.pref_name, DRUG_INDICATION.mesh_id, DRUG_INDICATION.mesh_heading, DRUG_INDICATION.efo_id, DRUG_INDICATION.efo_term, DRUG_INDICATION.max_phase_for_ind FROM MOLECULE_DICTIONARY JOIN DRUG_INDICATION ON MOLECULE_DICTIONARY.molregno == DRUG_INDICATION.molregno ``` There's another column in `MOLECULE_DICTIONARY` called `chebi_par_id` but a spot check of the first 10 rows showed this doesn't have good coverage
process
chembl indication processor this sql statement can be used with the chembl downloader to get links between chemicals and their indications either as mesh or efo hp and their maximum clinical phase sql select molecule dictionary chembl id molecule dictionary pref name drug indication mesh id drug indication mesh heading drug indication efo id drug indication efo term drug indication max phase for ind from molecule dictionary join drug indication on molecule dictionary molregno drug indication molregno there s another column in molecule dictionary called chebi par id but a spot check of the first rows showed this doesn t have good coverage
1
22,195
30,750,995,381
IssuesEvent
2023-07-28 19:13:50
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
opened
Update `ndbc_stations.xml` after 7-character buoy id's are introduced in Aug/Sept 2023
requestor: NOAA/EMC type: task priority: medium alert: NEED ACCOUNT KEY MET: PreProcessing Tools (Point)
## Describe the Task ## The `ndbc_stations.xml` file is a static file defining the (lat, lon) buoy locations. Unfortunately, the NDBC file format is not self-describing and ascii2nc performs a lookup from a static table file to retrieve the (lat, lon) location for each buoy based on its ID. The issue is that these locations do change relatively frequently. For example, in the 18 days between July 10, 2023 and July 28, 2023, differences emerged for 30 of the 2626 lines in that file. The huge problem here is that NDBC buoy data can only be trusted if... 1. The `ndbc_stations.xml` locations file has been updated very recently. 2. You are processing near realtime buoy data. Processing "old" buoy data creates low confidence in the accuracy of the (lat, lon) locations. This issue is to update the locations in that file after a relatively large change occurs in mid to late August 2023. Some of the stations ids will be modified from 5 characters to 7 characters. Since ascii2nc parses this data based on whitespace rather than using a fixed-width-format, no code changes should be required. But we should test with 7 character buoy id's to confirm they are processed as expected. Please see this change description: https://www.weather.gov/media/notification/pdf_2023_24/scn23-85_ndbc_web_file_format_change.pdf FYI, here's how I've run this in the past: ``` cd MET/scripts/python/utility ./build_ndbc_stations_from_web.py mv merged.txt ndbc_stations.xml ``` Please coordinate with alicia.bentley@noaa.gov on these updates. ### Time Estimate ### 2 hours. ### Sub-Issues ### Consider breaking the task down into sub-issues. None needed. ### Relevant Deadlines ### Do this between mid-August and mid-September, 2023 after the changes have taken effect. ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED CYCLE ASSIGNMENT** label - [x] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) No impacts. ## Task Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Update `ndbc_stations.xml` after 7-character buoy id's are introduced in Aug/Sept 2023 - ## Describe the Task ## The `ndbc_stations.xml` file is a static file defining the (lat, lon) buoy locations. Unfortunately, the NDBC file format is not self-describing and ascii2nc performs a lookup from a static table file to retrieve the (lat, lon) location for each buoy based on its ID. The issue is that these locations do change relatively frequently. For example, in the 18 days between July 10, 2023 and July 28, 2023, differences emerged for 30 of the 2626 lines in that file. The huge problem here is that NDBC buoy data can only be trusted if... 1. The `ndbc_stations.xml` locations file has been updated very recently. 2. You are processing near realtime buoy data. Processing "old" buoy data creates low confidence in the accuracy of the (lat, lon) locations. This issue is to update the locations in that file after a relatively large change occurs in mid to late August 2023. Some of the stations ids will be modified from 5 characters to 7 characters. Since ascii2nc parses this data based on whitespace rather than using a fixed-width-format, no code changes should be required. But we should test with 7 character buoy id's to confirm they are processed as expected. Please see this change description: https://www.weather.gov/media/notification/pdf_2023_24/scn23-85_ndbc_web_file_format_change.pdf FYI, here's how I've run this in the past: ``` cd MET/scripts/python/utility ./build_ndbc_stations_from_web.py mv merged.txt ndbc_stations.xml ``` Please coordinate with alicia.bentley@noaa.gov on these updates. ### Time Estimate ### 2 hours. ### Sub-Issues ### Consider breaking the task down into sub-issues. None needed. ### Relevant Deadlines ### Do this between mid-August and mid-September, 2023 after the changes have taken effect. ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED CYCLE ASSIGNMENT** label - [x] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) No impacts. ## Task Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
process
update ndbc stations xml after character buoy id s are introduced in aug sept describe the task the ndbc stations xml file is a static file defining the lat lon buoy locations unfortunately the ndbc file format is not self describing and performs a lookup from a static table file to retrieve the lat lon location for each buoy based on its id the issue is that these locations do change relatively frequently for example in the days between july and july differences emerged for of the lines in that file the huge problem here is that ndbc buoy data can only be trusted if the ndbc stations xml locations file has been updated very recently you are processing near realtime buoy data processing old buoy data creates low confidence in the accuracy of the lat lon locations this issue is to update the locations in that file after a relatively large change occurs in mid to late august some of the stations ids will be modified from characters to characters since parses this data based on whitespace rather than using a fixed width format no code changes should be required but we should test with character buoy id s to confirm they are processed as expected please see this change description fyi here s how i ve run this in the past cd met scripts python utility build ndbc stations from web py mv merged txt ndbc stations xml please coordinate with alicia bentley noaa gov on these updates time estimate hours sub issues consider breaking the task down into sub issues none needed relevant deadlines do this between mid august and mid september after the changes have taken effect funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need cycle assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components no impacts task checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and development issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
13,244
15,715,605,837
IssuesEvent
2021-03-28 02:17:49
tdwg/chrono
https://api.github.com/repos/tdwg/chrono
closed
Update Term List Abstract
Process - under public review
Following recommendation by @gkampmeier in issue #15, update abstract on [term list document](https://tdwg.github.io/chrono/list/) to say: "The Chronometric Age Vocabulary is a standard for transmitting information about chronometric ages - the processes and results of an assay or other analysis used to determine the age of a sample. This document lists all terms in namespaces currently used in the vocabulary."
1.0
Update Term List Abstract - Following recommendation by @gkampmeier in issue #15, update abstract on [term list document](https://tdwg.github.io/chrono/list/) to say: "The Chronometric Age Vocabulary is a standard for transmitting information about chronometric ages - the processes and results of an assay or other analysis used to determine the age of a sample. This document lists all terms in namespaces currently used in the vocabulary."
process
update term list abstract following recommendation by gkampmeier in issue update abstract on to say the chronometric age vocabulary is a standard for transmitting information about chronometric ages the processes and results of an assay or other analysis used to determine the age of a sample this document lists all terms in namespaces currently used in the vocabulary
1
15,913
20,118,959,742
IssuesEvent
2022-02-07 23:00:02
googleapis/cloud-bigtable-cbt-cli
https://api.github.com/repos/googleapis/cloud-bigtable-cbt-cli
closed
Security Policy violation Branch Protection
allstar type: process
Allstar has detected that this repository’s Branch Protection security policy is out of compliance. Status: Branch Protection enforcement is configured in Allstar, however Branch Protection is not available on this repository. Upgrade to GitHub Pro or make this repository public to enable this feature. See: https://docs.github.com/en/repositories/configuring-branches-and-merges-in-your-repository/defining-the-mergeability-of-pull-requests/about-protected-branches for more information. If this is not feasible, then disable Branch Protection policy enforcement for this repository in Allstar configuration. This issue will auto resolve when the policy is in compliance. Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
1.0
Security Policy violation Branch Protection - Allstar has detected that this repository’s Branch Protection security policy is out of compliance. Status: Branch Protection enforcement is configured in Allstar, however Branch Protection is not available on this repository. Upgrade to GitHub Pro or make this repository public to enable this feature. See: https://docs.github.com/en/repositories/configuring-branches-and-merges-in-your-repository/defining-the-mergeability-of-pull-requests/about-protected-branches for more information. If this is not feasible, then disable Branch Protection policy enforcement for this repository in Allstar configuration. This issue will auto resolve when the policy is in compliance. Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
process
security policy violation branch protection allstar has detected that this repository’s branch protection security policy is out of compliance status branch protection enforcement is configured in allstar however branch protection is not available on this repository upgrade to github pro or make this repository public to enable this feature see for more information if this is not feasible then disable branch protection policy enforcement for this repository in allstar configuration this issue will auto resolve when the policy is in compliance issue created by allstar see for more information for questions specific to the repository please contact the owner or maintainer
1
1,532
4,119,246,916
IssuesEvent
2016-06-08 14:21:50
gratipay/inside.gratipay.com
https://api.github.com/repos/gratipay/inside.gratipay.com
opened
appraise our CMMI maturity level
Governance & Process
I noticed a quick [appraisal](https://www.loomio.org/d/j5yu8acP/does-loomio-provide-any-reliable-support) of Loomio/Enspiral as having CMMI Level 1 (cf. #421), and it got me thinking: what's _our_ CMMI maturity level? > CMMI models provide guidance for developing or improving processes that meet the business goals of an organization. A CMMI model may also be used as a framework for appraising the process maturity of the organization. https://en.wikipedia.org/wiki/Capability_Maturity_Model_Integration > An organization cannot be certified in CMMI; instead, an organization is _appraised_. Depending on the type of appraisal, the organization can be awarded a maturity level rating (1-5) or a capability level achievement profile. https://en.wikipedia.org/wiki/Capability_Maturity_Model_Integration#Appraisal ![screen shot 2016-06-08 at 10 00 28 am](https://cloud.githubusercontent.com/assets/134455/15896817/ddbcb40a-2d5f-11e6-90f1-b140c1016606.png)
1.0
appraise our CMMI maturity level - I noticed a quick [appraisal](https://www.loomio.org/d/j5yu8acP/does-loomio-provide-any-reliable-support) of Loomio/Enspiral as having CMMI Level 1 (cf. #421), and it got me thinking: what's _our_ CMMI maturity level? > CMMI models provide guidance for developing or improving processes that meet the business goals of an organization. A CMMI model may also be used as a framework for appraising the process maturity of the organization. https://en.wikipedia.org/wiki/Capability_Maturity_Model_Integration > An organization cannot be certified in CMMI; instead, an organization is _appraised_. Depending on the type of appraisal, the organization can be awarded a maturity level rating (1-5) or a capability level achievement profile. https://en.wikipedia.org/wiki/Capability_Maturity_Model_Integration#Appraisal ![screen shot 2016-06-08 at 10 00 28 am](https://cloud.githubusercontent.com/assets/134455/15896817/ddbcb40a-2d5f-11e6-90f1-b140c1016606.png)
process
appraise our cmmi maturity level i noticed a quick of loomio enspiral as having cmmi level cf and it got me thinking what s our cmmi maturity level cmmi models provide guidance for developing or improving processes that meet the business goals of an organization a cmmi model may also be used as a framework for appraising the process maturity of the organization an organization cannot be certified in cmmi instead an organization is appraised depending on the type of appraisal the organization can be awarded a maturity level rating or a capability level achievement profile
1
272,417
20,745,090,850
IssuesEvent
2022-03-14 21:51:58
kubernetes-sigs/gateway-api
https://api.github.com/repos/kubernetes-sigs/gateway-api
opened
Document Path Redirects and Rewrites (GEP 726)
kind/documentation
This GEP was implemented but has not been documented yet.
1.0
Document Path Redirects and Rewrites (GEP 726) - This GEP was implemented but has not been documented yet.
non_process
document path redirects and rewrites gep this gep was implemented but has not been documented yet
0
132,078
12,497,823,725
IssuesEvent
2020-06-01 17:10:22
OpenLiberty/ci.docker
https://api.github.com/repos/OpenLiberty/ci.docker
closed
Update for sso provider "oauth2" in https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md
documentation
@meiaus commented on [Sun May 17 2020](https://github.com/OpenLiberty/open-liberty-operator/issues/159) Not sure if the following SECURITY.md page is handled via "open-liberyt-operator" repo, but since the following docker build problem is experienced from running with Open Liberty Operator, tracking this ticket here: https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md Recently when running docker build with Dockerfile which was tested in April, e.g. ``` FROM openliberty/open-liberty:full-java8-openj9-ubi # Optional functionality ARG TLS=true ARG SEC_SSO_PROVIDERS="oauth oidc facebook google github twitter linkedin" #ARG OPENJ9_SCC=false ARG VERBOSE=true COPY --chown=1001:0 server.xml /config/ COPY --chown=1001:0 AcmeWebClientEar.ear /config/apps # This script will add the requested XML snippets and gow image to be fit-for-purpose RUN configure.sh ``` the following failure appeared: ``` ++ grep oauth + [[ -n sso-oauth2.xml ]] + cp /opt/ol/helpers/build/configuration_snippets/sso-oauth.xml /config/configDropins/defaults cp: cannot stat '/opt/ol/helpers/build/configuration_snippets/sso-oauth.xml': No such file or directory ``` After consulting with @brutif and @leochr, it's realized, the sso provider "oauth" has been changed to "oauth2". After updating Dockerfile to reflect the new provider "oauth2", ``` ARG SEC_SSO_PROVIDERS="oauth2 oidc facebook google github twitter linkedin" ``` the problem is the resolved. Please update the readme accoringly: https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md ![image](https://user-images.githubusercontent.com/27024803/82160585-08300780-985c-11ea-8a09-d395f8caed23.png)
1.0
Update for sso provider "oauth2" in https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md - @meiaus commented on [Sun May 17 2020](https://github.com/OpenLiberty/open-liberty-operator/issues/159) Not sure if the following SECURITY.md page is handled via "open-liberyt-operator" repo, but since the following docker build problem is experienced from running with Open Liberty Operator, tracking this ticket here: https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md Recently when running docker build with Dockerfile which was tested in April, e.g. ``` FROM openliberty/open-liberty:full-java8-openj9-ubi # Optional functionality ARG TLS=true ARG SEC_SSO_PROVIDERS="oauth oidc facebook google github twitter linkedin" #ARG OPENJ9_SCC=false ARG VERBOSE=true COPY --chown=1001:0 server.xml /config/ COPY --chown=1001:0 AcmeWebClientEar.ear /config/apps # This script will add the requested XML snippets and gow image to be fit-for-purpose RUN configure.sh ``` the following failure appeared: ``` ++ grep oauth + [[ -n sso-oauth2.xml ]] + cp /opt/ol/helpers/build/configuration_snippets/sso-oauth.xml /config/configDropins/defaults cp: cannot stat '/opt/ol/helpers/build/configuration_snippets/sso-oauth.xml': No such file or directory ``` After consulting with @brutif and @leochr, it's realized, the sso provider "oauth" has been changed to "oauth2". After updating Dockerfile to reflect the new provider "oauth2", ``` ARG SEC_SSO_PROVIDERS="oauth2 oidc facebook google github twitter linkedin" ``` the problem is the resolved. Please update the readme accoringly: https://github.com/OpenLiberty/ci.docker/blob/master/SECURITY.md ![image](https://user-images.githubusercontent.com/27024803/82160585-08300780-985c-11ea-8a09-d395f8caed23.png)
non_process
update for sso provider in meiaus commented on not sure if the following security md page is handled via open liberyt operator repo but since the following docker build problem is experienced from running with open liberty operator tracking this ticket here recently when running docker build with dockerfile which was tested in april e g from openliberty open liberty full ubi optional functionality arg tls true arg sec sso providers oauth oidc facebook google github twitter linkedin arg scc false arg verbose true copy chown server xml config copy chown acmewebclientear ear config apps this script will add the requested xml snippets and gow image to be fit for purpose run configure sh the following failure appeared grep oauth cp opt ol helpers build configuration snippets sso oauth xml config configdropins defaults cp cannot stat opt ol helpers build configuration snippets sso oauth xml no such file or directory after consulting with brutif and leochr it s realized the sso provider oauth has been changed to after updating dockerfile to reflect the new provider arg sec sso providers oidc facebook google github twitter linkedin the problem is the resolved please update the readme accoringly
0
15,714
19,848,805,557
IssuesEvent
2022-01-21 09:56:48
ooi-data/CE02SHSM-RID26-01-ADCPTA000-telemetered-adcp_velocity_earth
https://api.github.com/repos/ooi-data/CE02SHSM-RID26-01-ADCPTA000-telemetered-adcp_velocity_earth
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:47.948285. ## Details Flow name: `CE02SHSM-RID26-01-ADCPTA000-telemetered-adcp_velocity_earth` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:47.948285. ## Details Flow name: `CE02SHSM-RID26-01-ADCPTA000-telemetered-adcp_velocity_earth` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name telemetered adcp velocity earth task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
9,203
12,238,626,901
IssuesEvent
2020-05-04 20:09:24
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Conditionally skip a stage / job with a manual approval
Pri1 cba devops-cicd-process/tech devops/prod support-request
For our Terraform Deployment, we use an Azure DevOps pipeline that has 3 stages: 1. plan 2. apply (manual approval) 3. test For the apply stage we use a deployment job with an environment that has a manual approval (check). What we would like to have is "skipping" the apply and test stage, if the plan stage has shows no changes. Therefore we try to use the following yaml configuration for the apply stage: ```yaml - stage: ApplyShared dependsOn: PlanShared jobs: - job: CheckSharedChanges steps: - task: DownloadPipelineArtifact@2 inputs: artifactName: TerraformBuild downloadPath: $(System.DefaultWorkingDirectory) - bash: | # using a file for indicating changes in TF plan, since # you cannot pass variables between stages in Azure DevOps if [ -f ".shared-changes" ]; then echo '##vso[task.setvariable variable=shared_changes]yes' fi name: Check - deployment: ApplyShared dependsOn: CheckSharedChanges # this condition seems to be ignored, if there is a manual # approval on the stage condition: eq(dependencies.CheckSharedChanges.outputs['Check.shared_env'], 'yes') displayName: 'Apply - shared' # we configured a manual approval (check) for this environment, # so the pipeline stops and asks for an operator to approve the deployment environment: 'infra-shared' ``` According to this [issue on the MS Developer Community](https://developercommunity.visualstudio.com/content/problem/910855/multi-stage-pipelines-with-stage-conditions-and-ap.html), a condition on a stage with an approval is not checked before the approval, so the approach does not work. My question is: do you know any other way to implement this? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: b067a175-f640-7503-9c1e-f0130c6dbeda * Version Independent ID: ff743c7b-a103-eae6-4478-62ba995a4b36 * Content: [Pipeline deployment approvals - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/approvals.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @azooinmyluggage * Microsoft Alias: **shashban**
1.0
Conditionally skip a stage / job with a manual approval - For our Terraform Deployment, we use an Azure DevOps pipeline that has 3 stages: 1. plan 2. apply (manual approval) 3. test For the apply stage we use a deployment job with an environment that has a manual approval (check). What we would like to have is "skipping" the apply and test stage, if the plan stage has shows no changes. Therefore we try to use the following yaml configuration for the apply stage: ```yaml - stage: ApplyShared dependsOn: PlanShared jobs: - job: CheckSharedChanges steps: - task: DownloadPipelineArtifact@2 inputs: artifactName: TerraformBuild downloadPath: $(System.DefaultWorkingDirectory) - bash: | # using a file for indicating changes in TF plan, since # you cannot pass variables between stages in Azure DevOps if [ -f ".shared-changes" ]; then echo '##vso[task.setvariable variable=shared_changes]yes' fi name: Check - deployment: ApplyShared dependsOn: CheckSharedChanges # this condition seems to be ignored, if there is a manual # approval on the stage condition: eq(dependencies.CheckSharedChanges.outputs['Check.shared_env'], 'yes') displayName: 'Apply - shared' # we configured a manual approval (check) for this environment, # so the pipeline stops and asks for an operator to approve the deployment environment: 'infra-shared' ``` According to this [issue on the MS Developer Community](https://developercommunity.visualstudio.com/content/problem/910855/multi-stage-pipelines-with-stage-conditions-and-ap.html), a condition on a stage with an approval is not checked before the approval, so the approach does not work. My question is: do you know any other way to implement this? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: b067a175-f640-7503-9c1e-f0130c6dbeda * Version Independent ID: ff743c7b-a103-eae6-4478-62ba995a4b36 * Content: [Pipeline deployment approvals - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/approvals.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @azooinmyluggage * Microsoft Alias: **shashban**
process
conditionally skip a stage job with a manual approval for our terraform deployment we use an azure devops pipeline that has stages plan apply manual approval test for the apply stage we use a deployment job with an environment that has a manual approval check what we would like to have is skipping the apply and test stage if the plan stage has shows no changes therefore we try to use the following yaml configuration for the apply stage yaml stage applyshared dependson planshared jobs job checksharedchanges steps task downloadpipelineartifact inputs artifactname terraformbuild downloadpath system defaultworkingdirectory bash using a file for indicating changes in tf plan since you cannot pass variables between stages in azure devops if then echo vso yes fi name check deployment applyshared dependson checksharedchanges this condition seems to be ignored if there is a manual approval on the stage condition eq dependencies checksharedchanges outputs yes displayname apply shared we configured a manual approval check for this environment so the pipeline stops and asks for an operator to approve the deployment environment infra shared according to this a condition on a stage with an approval is not checked before the approval so the approach does not work my question is do you know any other way to implement this document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login azooinmyluggage microsoft alias shashban
1
74,272
15,325,450,975
IssuesEvent
2021-02-26 01:20:31
idonthaveafifaaddiction/MapLoom
https://api.github.com/repos/idonthaveafifaaddiction/MapLoom
closed
CVE-2016-10541 (High) detected in shell-quote-0.0.1.tgz - autoclosed
security vulnerability
## CVE-2016-10541 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-0.0.1.tgz</b></p></summary> <p>quote and parse shell commands</p> <p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-0.0.1.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-0.0.1.tgz</a></p> <p>Path to dependency file: MapLoom/vendor/mgrs/package.json</p> <p>Path to vulnerable library: MapLoom/vendor/mgrs/node_modules/shell-quote/package.json</p> <p> Dependency Hierarchy: - browserify-8.1.3.tgz (Root Library) - :x: **shell-quote-0.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/MapLoom/commit/58d4532ed41534d625ef81d1ac77b1b561e8870b">58d4532ed41534d625ef81d1ac77b1b561e8870b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm module "shell-quote" 1.6.0 and earlier cannot correctly escape ">" and "<" operator used for redirection in shell. Applications that depend on shell-quote may also be vulnerable. A malicious user could perform code injection. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10541>CVE-2016-10541</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10541">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10541</a></p> <p>Release Date: 2018-12-15</p> <p>Fix Resolution: 1.6.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"shell-quote","packageVersion":"0.0.1","packageFilePaths":["/vendor/mgrs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browserify:8.1.3;shell-quote:0.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10541","vulnerabilityDetails":"The npm module \"shell-quote\" 1.6.0 and earlier cannot correctly escape \"\u003e\" and \"\u003c\" operator used for redirection in shell. Applications that depend on shell-quote may also be vulnerable. A malicious user could perform code injection.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10541","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-10541 (High) detected in shell-quote-0.0.1.tgz - autoclosed - ## CVE-2016-10541 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-0.0.1.tgz</b></p></summary> <p>quote and parse shell commands</p> <p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-0.0.1.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-0.0.1.tgz</a></p> <p>Path to dependency file: MapLoom/vendor/mgrs/package.json</p> <p>Path to vulnerable library: MapLoom/vendor/mgrs/node_modules/shell-quote/package.json</p> <p> Dependency Hierarchy: - browserify-8.1.3.tgz (Root Library) - :x: **shell-quote-0.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/MapLoom/commit/58d4532ed41534d625ef81d1ac77b1b561e8870b">58d4532ed41534d625ef81d1ac77b1b561e8870b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm module "shell-quote" 1.6.0 and earlier cannot correctly escape ">" and "<" operator used for redirection in shell. Applications that depend on shell-quote may also be vulnerable. A malicious user could perform code injection. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10541>CVE-2016-10541</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10541">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10541</a></p> <p>Release Date: 2018-12-15</p> <p>Fix Resolution: 1.6.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"shell-quote","packageVersion":"0.0.1","packageFilePaths":["/vendor/mgrs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browserify:8.1.3;shell-quote:0.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10541","vulnerabilityDetails":"The npm module \"shell-quote\" 1.6.0 and earlier cannot correctly escape \"\u003e\" and \"\u003c\" operator used for redirection in shell. Applications that depend on shell-quote may also be vulnerable. A malicious user could perform code injection.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10541","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in shell quote tgz autoclosed cve high severity vulnerability vulnerable library shell quote tgz quote and parse shell commands library home page a href path to dependency file maploom vendor mgrs package json path to vulnerable library maploom vendor mgrs node modules shell quote package json dependency hierarchy browserify tgz root library x shell quote tgz vulnerable library found in head commit a href found in base branch master vulnerability details the npm module shell quote and earlier cannot correctly escape and operator used for redirection in shell applications that depend on shell quote may also be vulnerable a malicious user could perform code injection publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree browserify shell quote isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm module shell quote and earlier cannot correctly escape and operator used for redirection in shell applications that depend on shell quote may also be vulnerable a malicious user could perform code injection vulnerabilityurl
0
18,901
24,839,965,164
IssuesEvent
2022-10-26 11:58:19
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Release 5.x - 'rules_jvm_external' not found in repo
more data needed type: support / not a bug (process) team-OSS
INFO: Repository rules_jvm_external instantiated at: E:/opensource/lyra/WORKSPACE:143:13: in <toplevel> Repository rule http_archive defined at: C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl:370:31: in <toplevel> DEBUG: Rule 'com_google_protobuf' indicated that a canonical reproducible form can be obtained by modifying arguments commit = "fd8aabf32d1d1ae3319e280fed07aa6eb24d150d", shallow_since = "1614800268 -0800" and dropping ["tag"] DEBUG: Repository com_google_protobuf instantiated at: E:/opensource/lyra/WORKSPACE:26:15: in <toplevel> Repository rule git_repository defined at: C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/git.bzl:176:33: in <toplevel> WARNING: Download from https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip failed: class java.io.FileNotFoundException GET returned 404 Not Found ERROR: An error occurred during the fetch of repository 'rules_jvm_external': Traceback (most recent call last): File "C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl", line 130, column 45, in _http_archive_impl download_info = ctx.download_and_extract( Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found ERROR: E:/opensource/lyra/WORKSPACE:143:13: fetching http_archive rule //external:rules_jvm_external: Traceback (most recent call last): File "C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl", line 130, column 45, in _http_archive_impl download_info = ctx.download_and_extract( Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found ERROR: Error computing the main repository mapping: no such package '@rules_jvm_external//': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found
1.0
Release 5.x - 'rules_jvm_external' not found in repo - INFO: Repository rules_jvm_external instantiated at: E:/opensource/lyra/WORKSPACE:143:13: in <toplevel> Repository rule http_archive defined at: C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl:370:31: in <toplevel> DEBUG: Rule 'com_google_protobuf' indicated that a canonical reproducible form can be obtained by modifying arguments commit = "fd8aabf32d1d1ae3319e280fed07aa6eb24d150d", shallow_since = "1614800268 -0800" and dropping ["tag"] DEBUG: Repository com_google_protobuf instantiated at: E:/opensource/lyra/WORKSPACE:26:15: in <toplevel> Repository rule git_repository defined at: C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/git.bzl:176:33: in <toplevel> WARNING: Download from https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip failed: class java.io.FileNotFoundException GET returned 404 Not Found ERROR: An error occurred during the fetch of repository 'rules_jvm_external': Traceback (most recent call last): File "C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl", line 130, column 45, in _http_archive_impl download_info = ctx.download_and_extract( Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found ERROR: E:/opensource/lyra/WORKSPACE:143:13: fetching http_archive rule //external:rules_jvm_external: Traceback (most recent call last): File "C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/bazel_tools/tools/build_defs/repo/http.bzl", line 130, column 45, in _http_archive_impl download_info = ctx.download_and_extract( Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found ERROR: Error computing the main repository mapping: no such package '@rules_jvm_external//': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_jvm_external/archive/4.0.zip] to C:/users/cshsoft/_bazel_cshsoft/xcvtuzjs/external/rules_jvm_external/temp8625613266728458292/4.0.zip: GET returned 404 Not Found
process
release x rules jvm external not found in repo info repository rules jvm external instantiated at e opensource lyra workspace in repository rule http archive defined at c users cshsoft bazel cshsoft xcvtuzjs external bazel tools tools build defs repo http bzl in debug rule com google protobuf indicated that a canonical reproducible form can be obtained by modifying arguments commit shallow since and dropping debug repository com google protobuf instantiated at e opensource lyra workspace in repository rule git repository defined at c users cshsoft bazel cshsoft xcvtuzjs external bazel tools tools build defs repo git bzl in warning download from failed class java io filenotfoundexception get returned not found error an error occurred during the fetch of repository rules jvm external traceback most recent call last file c users cshsoft bazel cshsoft xcvtuzjs external bazel tools tools build defs repo http bzl line column in http archive impl download info ctx download and extract error in download and extract java io ioexception error downloading to c users cshsoft bazel cshsoft xcvtuzjs external rules jvm external zip get returned not found error e opensource lyra workspace fetching http archive rule external rules jvm external traceback most recent call last file c users cshsoft bazel cshsoft xcvtuzjs external bazel tools tools build defs repo http bzl line column in http archive impl download info ctx download and extract error in download and extract java io ioexception error downloading to c users cshsoft bazel cshsoft xcvtuzjs external rules jvm external zip get returned not found error error computing the main repository mapping no such package rules jvm external java io ioexception error downloading to c users cshsoft bazel cshsoft xcvtuzjs external rules jvm external zip get returned not found
1
11,800
7,714,284,970
IssuesEvent
2018-05-23 01:07:10
JuliaLang/julia
https://api.github.com/repos/JuliaLang/julia
closed
skipmissing performance regression
iteration missing data performance regression
The new iteration protocol appears to have introduced a large performance regression for `skipmissing`: Before: ```julia julia> x = Vector{Union{Float64,Missing}}(rand(10_000)); julia> @btime sum(skipmissing(x)); 65.523 μs (5 allocations: 80 bytes) ``` After: ```julia julia> x = Vector{Union{Float64,Missing}}(rand(10_000)); julia> @btime sum(skipmissing(x)); 2.211 ms (48983 allocations: 921.61 KiB) ``` https://github.com/JuliaCI/BaseBenchmarks.jl/pull/200 adds benchmarks for this. Cc: @Keno
True
skipmissing performance regression - The new iteration protocol appears to have introduced a large performance regression for `skipmissing`: Before: ```julia julia> x = Vector{Union{Float64,Missing}}(rand(10_000)); julia> @btime sum(skipmissing(x)); 65.523 μs (5 allocations: 80 bytes) ``` After: ```julia julia> x = Vector{Union{Float64,Missing}}(rand(10_000)); julia> @btime sum(skipmissing(x)); 2.211 ms (48983 allocations: 921.61 KiB) ``` https://github.com/JuliaCI/BaseBenchmarks.jl/pull/200 adds benchmarks for this. Cc: @Keno
non_process
skipmissing performance regression the new iteration protocol appears to have introduced a large performance regression for skipmissing before julia julia x vector union missing rand julia btime sum skipmissing x μs allocations bytes after julia julia x vector union missing rand julia btime sum skipmissing x ms allocations kib adds benchmarks for this cc keno
0
9,916
12,955,326,163
IssuesEvent
2020-07-20 06:04:26
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Implement missing unexecutable migration warnings
component: migration engine process/candidate
The following checks have been planned for a while, but were never implemented: - Unimplementable unique constraints: added a unique constraint to a column that contains duplicates - Deleted used enum value: one variant of an enum was removed in the prisma schema, but it is being used in the database This issue is about issuing warnings without actually checking database state, for now, to make implementation simple. We can later actually check for duplicates and used enum variants. Proposed messages: - "A unique constraint was added to the `...` field. If there are existing duplicate values in that column, the migration will fail." - "The `...` value on the `...` enum was removed. If this variant is still used in the database, the migration will fail."
1.0
Implement missing unexecutable migration warnings - The following checks have been planned for a while, but were never implemented: - Unimplementable unique constraints: added a unique constraint to a column that contains duplicates - Deleted used enum value: one variant of an enum was removed in the prisma schema, but it is being used in the database This issue is about issuing warnings without actually checking database state, for now, to make implementation simple. We can later actually check for duplicates and used enum variants. Proposed messages: - "A unique constraint was added to the `...` field. If there are existing duplicate values in that column, the migration will fail." - "The `...` value on the `...` enum was removed. If this variant is still used in the database, the migration will fail."
process
implement missing unexecutable migration warnings the following checks have been planned for a while but were never implemented unimplementable unique constraints added a unique constraint to a column that contains duplicates deleted used enum value one variant of an enum was removed in the prisma schema but it is being used in the database this issue is about issuing warnings without actually checking database state for now to make implementation simple we can later actually check for duplicates and used enum variants proposed messages a unique constraint was added to the field if there are existing duplicate values in that column the migration will fail the value on the enum was removed if this variant is still used in the database the migration will fail
1
4,293
7,192,425,134
IssuesEvent
2018-02-03 03:25:59
amaster507/ifbmt
https://api.github.com/repos/amaster507/ifbmt
closed
Calendar and timezones
GUI calendar idea process question
The issue of timezones was brought up by @andrewwippler when tagging the calendar Orion of the app. Here is my thoughts for this process: - Enable users in the setup condiguration and profile settings to select and change their timezone. - possibly allow a quick change of the timezone setting when traveling. This change will not be automattic bit like a quick link to change on the fly when needed. - store all bookings, notes and events in UTC and then display according to the setting chosen at the time. - when viewing church details there can be a notice that they are in a different timezone. - when booking the meeting the time will be saved as UTC offset by the churches local time. Let me know if your thoughts and suggestions for this.process.
1.0
Calendar and timezones - The issue of timezones was brought up by @andrewwippler when tagging the calendar Orion of the app. Here is my thoughts for this process: - Enable users in the setup condiguration and profile settings to select and change their timezone. - possibly allow a quick change of the timezone setting when traveling. This change will not be automattic bit like a quick link to change on the fly when needed. - store all bookings, notes and events in UTC and then display according to the setting chosen at the time. - when viewing church details there can be a notice that they are in a different timezone. - when booking the meeting the time will be saved as UTC offset by the churches local time. Let me know if your thoughts and suggestions for this.process.
process
calendar and timezones the issue of timezones was brought up by andrewwippler when tagging the calendar orion of the app here is my thoughts for this process enable users in the setup condiguration and profile settings to select and change their timezone possibly allow a quick change of the timezone setting when traveling this change will not be automattic bit like a quick link to change on the fly when needed store all bookings notes and events in utc and then display according to the setting chosen at the time when viewing church details there can be a notice that they are in a different timezone when booking the meeting the time will be saved as utc offset by the churches local time let me know if your thoughts and suggestions for this process
1
8,429
11,595,796,634
IssuesEvent
2020-02-24 17:41:04
MHRA/products
https://api.github.com/repos/MHRA/products
closed
Stub API needs to implement Authentication
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up: STORY :book:
**Is your feature request related to a problem? Please describe.** In order to unblock integration from Sentinel, the Batch API will need to implement Authorization (basic auth with a static header and a static 'API key'. **Acceptance Criteria** - [ ] If the user doesn't provide the correct username & key in an Authorization header when calling the stub API, the stub should return a `401 Unauthorized` response. **Describe alternatives you've considered** We considered only locking down by network, but adding this layer of security is trivial and allows for a more secure upload workflow. **Exit Criteria met** - [x] Backlog - [x] Discovery - [x] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
Stub API needs to implement Authentication - **Is your feature request related to a problem? Please describe.** In order to unblock integration from Sentinel, the Batch API will need to implement Authorization (basic auth with a static header and a static 'API key'. **Acceptance Criteria** - [ ] If the user doesn't provide the correct username & key in an Authorization header when calling the stub API, the stub should return a `401 Unauthorized` response. **Describe alternatives you've considered** We considered only locking down by network, but adding this layer of security is trivial and allows for a more secure upload workflow. **Exit Criteria met** - [x] Backlog - [x] Discovery - [x] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
stub api needs to implement authentication is your feature request related to a problem please describe in order to unblock integration from sentinel the batch api will need to implement authorization basic auth with a static header and a static api key acceptance criteria if the user doesn t provide the correct username key in an authorization header when calling the stub api the stub should return a unauthorized response describe alternatives you ve considered we considered only locking down by network but adding this layer of security is trivial and allows for a more secure upload workflow exit criteria met backlog discovery duxd development quality assurance release and validate
1
22,103
3,933,833,423
IssuesEvent
2016-04-25 20:30:04
GCSO2/D3S-INVENTORY
https://api.github.com/repos/GCSO2/D3S-INVENTORY
closed
Cloud: PO Details: Add Part - Make Supplier catalog records available for selection
bug ready for gio to test
- Add Catalog # field after Part # ![image](https://cloud.githubusercontent.com/assets/5666297/13961273/48148014-f032-11e5-8a7c-42000623e7db.png) - Add Catalog # after Part # on selection grid below ![image](https://cloud.githubusercontent.com/assets/5666297/13961319/7a5873f0-f032-11e5-8ff4-5e4527787d71.png) - Change underlining query ADD union with part catalog table. Only part catalog for supplier selected on the PO Details should show up for selection: barcode show from catalog table (inv_part_catalogues_wv) part number show from part table catalog number show from catalog table description show from catalog table. group based on join from part table.
1.0
Cloud: PO Details: Add Part - Make Supplier catalog records available for selection - - Add Catalog # field after Part # ![image](https://cloud.githubusercontent.com/assets/5666297/13961273/48148014-f032-11e5-8a7c-42000623e7db.png) - Add Catalog # after Part # on selection grid below ![image](https://cloud.githubusercontent.com/assets/5666297/13961319/7a5873f0-f032-11e5-8ff4-5e4527787d71.png) - Change underlining query ADD union with part catalog table. Only part catalog for supplier selected on the PO Details should show up for selection: barcode show from catalog table (inv_part_catalogues_wv) part number show from part table catalog number show from catalog table description show from catalog table. group based on join from part table.
non_process
cloud po details add part make supplier catalog records available for selection add catalog field after part add catalog after part on selection grid below change underlining query add union with part catalog table only part catalog for supplier selected on the po details should show up for selection barcode show from catalog table inv part catalogues wv part number show from part table catalog number show from catalog table description show from catalog table group based on join from part table
0
22,483
31,395,271,417
IssuesEvent
2023-08-26 21:21:13
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add The Splattering from Locke & Key
suggested title in process
Title: The Splatting Type (film/tv show): Film Film or show in which it appears: Locke & Key Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80241239) About when in the parent film/show does it appear? Season 2, episode 1, 31:15 Actual footage of the film/show can be seen (yes/no)? I don't think the entirety of _The Splattering_ exists inside the containing show, but a much cleaner edit of it exists in an official promo for the TV show: https://www.youtube.com/watch?v=etreLzo-i0s
1.0
Add The Splattering from Locke & Key - Title: The Splatting Type (film/tv show): Film Film or show in which it appears: Locke & Key Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80241239) About when in the parent film/show does it appear? Season 2, episode 1, 31:15 Actual footage of the film/show can be seen (yes/no)? I don't think the entirety of _The Splattering_ exists inside the containing show, but a much cleaner edit of it exists in an official promo for the TV show: https://www.youtube.com/watch?v=etreLzo-i0s
process
add the splattering from locke key title the splatting type film tv show film film or show in which it appears locke key is the parent film show streaming anywhere about when in the parent film show does it appear season episode actual footage of the film show can be seen yes no i don t think the entirety of the splattering exists inside the containing show but a much cleaner edit of it exists in an official promo for the tv show
1
7,598
10,707,498,689
IssuesEvent
2019-10-24 17:36:22
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
Spanner: 'test_transaction_batch_update_and_execute_dml' systest flakes with 'None' status code.
api: spanner flaky testing type: process
Similar to #7504, but this time the status code is `None`. From [this CI failure](https://source.cloud.google.com/results/invocations/ff8f8339-1a05-41e6-ac3a-6ed36f0061e1/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fspanner/log): ```python _________ TestSessionAPI.test_transaction_batch_update_and_execute_dml _________ self = <tests.system.test_system.TestSessionAPI testMethod=test_transaction_batch_update_and_execute_dml> def test_transaction_batch_update_and_execute_dml(self): retry = RetryInstanceState(_has_all_ddl) retry(self._db.reload)() session = self._db.session() session.create() self.to_delete.append(session) with session.batch() as batch: batch.delete(self.TABLE, self.ALL) insert_statements = list(self._generate_insert_statements()) update_statements = [ ( "UPDATE contacts SET email = @email " "WHERE contact_id = @contact_id;", {"contact_id": 1, "email": "phreddy@example.com"}, {"contact_id": Type(code=INT64), "email": Type(code=STRING)}, ) ] delete_statement = "DELETE contacts WHERE TRUE;" def unit_of_work(transaction, self): rows = list(transaction.read(self.TABLE, self.COLUMNS, self.ALL)) self.assertEqual(rows, []) status, row_counts = transaction.batch_update( insert_statements + update_statements ) self._check_batch_status(status.code) self.assertEqual(len(row_counts), len(insert_statements) + 1) for row_count in row_counts: self.assertEqual(row_count, 1) row_count = transaction.execute_update(delete_statement) self.assertEqual(row_count, len(insert_statements)) > session.run_in_transaction(unit_of_work, self) tests/system/test_system.py:874: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/spanner_v1/session.py:299: in run_in_transaction return_value = func(txn, *args, **kw) tests/system/test_system.py:865: in unit_of_work self._check_batch_status(status.code) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ status_code = 10 @staticmethod def _check_batch_status(status_code): if status_code != code_pb2.OK: > raise exceptions.from_grpc_status(status_code, "batch_update failed") E google.api_core.exceptions.GoogleAPICallError: None batch_update failed ```
1.0
Spanner: 'test_transaction_batch_update_and_execute_dml' systest flakes with 'None' status code. - Similar to #7504, but this time the status code is `None`. From [this CI failure](https://source.cloud.google.com/results/invocations/ff8f8339-1a05-41e6-ac3a-6ed36f0061e1/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fspanner/log): ```python _________ TestSessionAPI.test_transaction_batch_update_and_execute_dml _________ self = <tests.system.test_system.TestSessionAPI testMethod=test_transaction_batch_update_and_execute_dml> def test_transaction_batch_update_and_execute_dml(self): retry = RetryInstanceState(_has_all_ddl) retry(self._db.reload)() session = self._db.session() session.create() self.to_delete.append(session) with session.batch() as batch: batch.delete(self.TABLE, self.ALL) insert_statements = list(self._generate_insert_statements()) update_statements = [ ( "UPDATE contacts SET email = @email " "WHERE contact_id = @contact_id;", {"contact_id": 1, "email": "phreddy@example.com"}, {"contact_id": Type(code=INT64), "email": Type(code=STRING)}, ) ] delete_statement = "DELETE contacts WHERE TRUE;" def unit_of_work(transaction, self): rows = list(transaction.read(self.TABLE, self.COLUMNS, self.ALL)) self.assertEqual(rows, []) status, row_counts = transaction.batch_update( insert_statements + update_statements ) self._check_batch_status(status.code) self.assertEqual(len(row_counts), len(insert_statements) + 1) for row_count in row_counts: self.assertEqual(row_count, 1) row_count = transaction.execute_update(delete_statement) self.assertEqual(row_count, len(insert_statements)) > session.run_in_transaction(unit_of_work, self) tests/system/test_system.py:874: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/spanner_v1/session.py:299: in run_in_transaction return_value = func(txn, *args, **kw) tests/system/test_system.py:865: in unit_of_work self._check_batch_status(status.code) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ status_code = 10 @staticmethod def _check_batch_status(status_code): if status_code != code_pb2.OK: > raise exceptions.from_grpc_status(status_code, "batch_update failed") E google.api_core.exceptions.GoogleAPICallError: None batch_update failed ```
process
spanner test transaction batch update and execute dml systest flakes with none status code similar to but this time the status code is none from python testsessionapi test transaction batch update and execute dml self def test transaction batch update and execute dml self retry retryinstancestate has all ddl retry self db reload session self db session session create self to delete append session with session batch as batch batch delete self table self all insert statements list self generate insert statements update statements update contacts set email email where contact id contact id contact id email phreddy example com contact id type code email type code string delete statement delete contacts where true def unit of work transaction self rows list transaction read self table self columns self all self assertequal rows status row counts transaction batch update insert statements update statements self check batch status status code self assertequal len row counts len insert statements for row count in row counts self assertequal row count row count transaction execute update delete statement self assertequal row count len insert statements session run in transaction unit of work self tests system test system py google cloud spanner session py in run in transaction return value func txn args kw tests system test system py in unit of work self check batch status status code status code staticmethod def check batch status status code if status code code ok raise exceptions from grpc status status code batch update failed e google api core exceptions googleapicallerror none batch update failed
1
10,072
13,044,161,910
IssuesEvent
2020-07-29 03:47:27
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `ToSeconds` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `ToSeconds` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `ToSeconds` from TiDB - ## Description Port the scalar function `ToSeconds` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function toseconds from tidb description port the scalar function toseconds from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
6,714
2,857,499,087
IssuesEvent
2015-06-02 20:01:52
alexrj/Slic3r
https://api.github.com/repos/alexrj/Slic3r
closed
Small Perimeter Speed Not Applied to Holes
Needs more info Needs test STL file
Small perimeter speeds are not being applied to holes. a 3.25mm hole is printing with the standard perimeter speed. I'm using 1.1.6, and have seen it in 1.1.5 as well.
1.0
Small Perimeter Speed Not Applied to Holes - Small perimeter speeds are not being applied to holes. a 3.25mm hole is printing with the standard perimeter speed. I'm using 1.1.6, and have seen it in 1.1.5 as well.
non_process
small perimeter speed not applied to holes small perimeter speeds are not being applied to holes a hole is printing with the standard perimeter speed i m using and have seen it in as well
0
5,416
8,249,183,082
IssuesEvent
2018-09-11 20:45:33
JDRF/spirit
https://api.github.com/repos/JDRF/spirit
closed
Getting Started
4 PUBLISH Process
Deliver * Publish guidelines doc For Designers: https://docs.google.com/document/d/1zGl8aIkrjJVgAnC2_9vfvPaOrli-XSkDq3WeKmW7HMo/edit#heading=h.tur6asjl97sy For Developers: https://docs.google.com/document/d/1izGQJbckd99tjsSO63f1ExUanFH7ox5gRJwG9V2bSRg/edit
1.0
Getting Started - Deliver * Publish guidelines doc For Designers: https://docs.google.com/document/d/1zGl8aIkrjJVgAnC2_9vfvPaOrli-XSkDq3WeKmW7HMo/edit#heading=h.tur6asjl97sy For Developers: https://docs.google.com/document/d/1izGQJbckd99tjsSO63f1ExUanFH7ox5gRJwG9V2bSRg/edit
process
getting started deliver publish guidelines doc for designers for developers
1
11,863
14,665,726,451
IssuesEvent
2020-12-29 14:51:00
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Use of keyword keyref in a DITA Map
bug preprocess/keyref priority/medium stale
Let's say my DITA Map has inside it something like: ``` <keydef keys="original"> <topicmeta><keywords><keyword>Original Content</keyword></keywords></topicmeta> </keydef> <keydef keys="product_name"> <topicmeta> <keywords><keyword keyref="original"/></keywords> </topicmeta> </keydef> ``` and in a DITA topic I have: ``` <ph keyref="product_name"/> ``` Should the "product_name" key be resolved to the "Original Content" text in the published output for the HTML topic? Right now it is not, with either DITA OT 1.8 or 2.x. A workaround would have been to use a conref instead of a keyref in the DITA Map: ``` <keydef keys="original"> <topicmeta><keywords><keyword id="originalID">Original Content</keyword></keywords></topicmeta> </keydef> <keydef keys="product_name"> <topicmeta> <keywords><keyword conref="#originalID"/></keywords> </topicmeta> </keydef> ``` but this no longer works with DITA OT 2.x because of this bug: https://github.com/dita-ot/dita-ot/issues/2420
1.0
Use of keyword keyref in a DITA Map - Let's say my DITA Map has inside it something like: ``` <keydef keys="original"> <topicmeta><keywords><keyword>Original Content</keyword></keywords></topicmeta> </keydef> <keydef keys="product_name"> <topicmeta> <keywords><keyword keyref="original"/></keywords> </topicmeta> </keydef> ``` and in a DITA topic I have: ``` <ph keyref="product_name"/> ``` Should the "product_name" key be resolved to the "Original Content" text in the published output for the HTML topic? Right now it is not, with either DITA OT 1.8 or 2.x. A workaround would have been to use a conref instead of a keyref in the DITA Map: ``` <keydef keys="original"> <topicmeta><keywords><keyword id="originalID">Original Content</keyword></keywords></topicmeta> </keydef> <keydef keys="product_name"> <topicmeta> <keywords><keyword conref="#originalID"/></keywords> </topicmeta> </keydef> ``` but this no longer works with DITA OT 2.x because of this bug: https://github.com/dita-ot/dita-ot/issues/2420
process
use of keyword keyref in a dita map let s say my dita map has inside it something like original content and in a dita topic i have should the product name key be resolved to the original content text in the published output for the html topic right now it is not with either dita ot or x a workaround would have been to use a conref instead of a keyref in the dita map original content but this no longer works with dita ot x because of this bug
1
56,043
14,078,378,250
IssuesEvent
2020-11-04 13:29:24
themagicalmammal/android_kernel_samsung_a5xelte
https://api.github.com/repos/themagicalmammal/android_kernel_samsung_a5xelte
opened
CVE-2019-14821 (High) detected in linux-yocto-4.10v3.17
security vulnerability
## CVE-2019-14821 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.10v3.17</b></p></summary> <p> <p>Linux 4.10 Embedded Kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.10>https://git.yoctoproject.org/git/linux-yocto-4.10</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p> <p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/virt/kvm/coalesced_mmio.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/virt/kvm/coalesced_mmio.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An out-of-bounds access issue was found in the Linux kernel, all versions through 5.3, in the way Linux kernel's KVM hypervisor implements the Coalesced MMIO write operation. It operates on an MMIO ring buffer 'struct kvm_coalesced_mmio' object, wherein write indices 'ring->first' and 'ring->last' value could be supplied by a host user-space process. An unprivileged host user or process with access to '/dev/kvm' device could use this flaw to crash the host kernel, resulting in a denial of service or potentially escalating privileges on the system. <p>Publish Date: 2019-09-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14821>CVE-2019-14821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-14821">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-14821</a></p> <p>Release Date: 2019-09-19</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14821 (High) detected in linux-yocto-4.10v3.17 - ## CVE-2019-14821 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.10v3.17</b></p></summary> <p> <p>Linux 4.10 Embedded Kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.10>https://git.yoctoproject.org/git/linux-yocto-4.10</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p> <p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/virt/kvm/coalesced_mmio.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/virt/kvm/coalesced_mmio.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An out-of-bounds access issue was found in the Linux kernel, all versions through 5.3, in the way Linux kernel's KVM hypervisor implements the Coalesced MMIO write operation. It operates on an MMIO ring buffer 'struct kvm_coalesced_mmio' object, wherein write indices 'ring->first' and 'ring->last' value could be supplied by a host user-space process. An unprivileged host user or process with access to '/dev/kvm' device could use this flaw to crash the host kernel, resulting in a denial of service or potentially escalating privileges on the system. <p>Publish Date: 2019-09-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14821>CVE-2019-14821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-14821">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-14821</a></p> <p>Release Date: 2019-09-19</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux yocto cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files android kernel samsung virt kvm coalesced mmio c android kernel samsung virt kvm coalesced mmio c vulnerability details an out of bounds access issue was found in the linux kernel all versions through in the way linux kernel s kvm hypervisor implements the coalesced mmio write operation it operates on an mmio ring buffer struct kvm coalesced mmio object wherein write indices ring first and ring last value could be supplied by a host user space process an unprivileged host user or process with access to dev kvm device could use this flaw to crash the host kernel resulting in a denial of service or potentially escalating privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
7,201
10,337,694,910
IssuesEvent
2019-09-03 15:20:39
ESMValGroup/ESMValTool
https://api.github.com/repos/ESMValGroup/ESMValTool
opened
preprocessor mask_landseaice not working anymore
preprocessor
When using the mask_landseaice preprocessor I get ValueError: Landsea-ice mask could not be found. Stopping. The last time I tried and used it is a while ago, and my recipe (recipe_collins13ipcc.yml) might be the only one using this. So far the recipe is only in ESMValTool-private, but it should go into public and AR6 (asap of course ;-)). Anyone able to help?
1.0
preprocessor mask_landseaice not working anymore - When using the mask_landseaice preprocessor I get ValueError: Landsea-ice mask could not be found. Stopping. The last time I tried and used it is a while ago, and my recipe (recipe_collins13ipcc.yml) might be the only one using this. So far the recipe is only in ESMValTool-private, but it should go into public and AR6 (asap of course ;-)). Anyone able to help?
process
preprocessor mask landseaice not working anymore when using the mask landseaice preprocessor i get valueerror landsea ice mask could not be found stopping the last time i tried and used it is a while ago and my recipe recipe yml might be the only one using this so far the recipe is only in esmvaltool private but it should go into public and asap of course anyone able to help
1
1,293
3,829,714,696
IssuesEvent
2016-03-31 11:58:37
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
SVG: xlink:href attribute value is not proxed
!IMPORTANT! AREA: client SYSTEM: resource processing TYPE: bug
see https://developer.mozilla.org/en-US/docs/Web/SVG/Attribute/xlink:href. If xlink attribute value contains external url we should to proxy its value. Example: ```javascript <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Main document</h1> <svg xmlns="http://www.w3.org/2000/svg" style="display: none"> <rect id="svg-rect" x="10" y="10" height="100" width="100" style="stroke:#ff0000; fill: #0000ff"/> </svg> <svg> <use style="width: 100px; height: 100px;" xlink:href="http://localhost:3000/#svg-rect"></use> </svg> </body> </html> ``` On example page there is the blue rectangle. If you open this page using playouground then blue rectangle don't display.
1.0
SVG: xlink:href attribute value is not proxed - see https://developer.mozilla.org/en-US/docs/Web/SVG/Attribute/xlink:href. If xlink attribute value contains external url we should to proxy its value. Example: ```javascript <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Main document</h1> <svg xmlns="http://www.w3.org/2000/svg" style="display: none"> <rect id="svg-rect" x="10" y="10" height="100" width="100" style="stroke:#ff0000; fill: #0000ff"/> </svg> <svg> <use style="width: 100px; height: 100px;" xlink:href="http://localhost:3000/#svg-rect"></use> </svg> </body> </html> ``` On example page there is the blue rectangle. If you open this page using playouground then blue rectangle don't display.
process
svg xlink href attribute value is not proxed see if xlink attribute value contains external url we should to proxy its value example javascript title main document use style width height xlink href on example page there is the blue rectangle if you open this page using playouground then blue rectangle don t display
1
424,324
29,048,413,301
IssuesEvent
2023-05-13 21:55:24
rustformers/llm
https://api.github.com/repos/rustformers/llm
opened
Start writing up developer documentation
documentation
We're getting more developers who are adding new implementations, which is great, but transformers and LLMs are complicated, and it'll require more than Rust knowledge to get to the bottom of why they aren't working. We should assist new devs by writing up docs for where they can find out more about LLMs and how they work, as well as defining any terms or concepts that they're likely to encounter. This is probably also a good place to discuss the idiosyncrasies of the ML ecosystem, including formats, conversions, tooling, and more. Here's my initial list of LLM resources: - <https://jalammar.github.io/illustrated-gpt2/> - <https://jalammar.github.io/illustrated-transformer/> - <https://www.youtube.com/watch?v=kCc8FmEb1nY> - <https://udlbook.github.io/udlbook/> (Chapter 12 specifically)
1.0
Start writing up developer documentation - We're getting more developers who are adding new implementations, which is great, but transformers and LLMs are complicated, and it'll require more than Rust knowledge to get to the bottom of why they aren't working. We should assist new devs by writing up docs for where they can find out more about LLMs and how they work, as well as defining any terms or concepts that they're likely to encounter. This is probably also a good place to discuss the idiosyncrasies of the ML ecosystem, including formats, conversions, tooling, and more. Here's my initial list of LLM resources: - <https://jalammar.github.io/illustrated-gpt2/> - <https://jalammar.github.io/illustrated-transformer/> - <https://www.youtube.com/watch?v=kCc8FmEb1nY> - <https://udlbook.github.io/udlbook/> (Chapter 12 specifically)
non_process
start writing up developer documentation we re getting more developers who are adding new implementations which is great but transformers and llms are complicated and it ll require more than rust knowledge to get to the bottom of why they aren t working we should assist new devs by writing up docs for where they can find out more about llms and how they work as well as defining any terms or concepts that they re likely to encounter this is probably also a good place to discuss the idiosyncrasies of the ml ecosystem including formats conversions tooling and more here s my initial list of llm resources chapter specifically
0
470,010
13,529,617,653
IssuesEvent
2020-09-15 18:35:09
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
"Attributes > Local Dialect > Local Attributes " modification does not affect in "Console > mange > user profile"
Priority/Highest Severity/Critical bug
**Describe the issue:** When doing any update "Show this attribute on user profile and user registration page" function in "Attributes > Local Dialect > Local Attributes " those modifications do not affect in "Console > mange > user profile" **How to reproduce:** Login to Console > Mange 1. Navigate to Attribute Dialects > Local Dialects > 2. select any Local Attribute 3. ex: Country and check > Show this attribute on user profile and user registration page option 4. Update 5. Verify that attribute displaying in the user profile, when new user creation ![image](https://user-images.githubusercontent.com/39120228/93249613-36dcd100-f7af-11ea-83bb-e156ac6d6275.png) Verify user profile after attribute update & while new user creation ![image](https://user-images.githubusercontent.com/39120228/93250139-fb8ed200-f7af-11ea-9c1e-272fe1d261ee.png) **Expected behavior:** The modification should be affected in the user profile as well as the User registration page. **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: 5.11 m36 snapshot - OS: Mac - Database: H2 - Userstore: LDAP ---
1.0
"Attributes > Local Dialect > Local Attributes " modification does not affect in "Console > mange > user profile" - **Describe the issue:** When doing any update "Show this attribute on user profile and user registration page" function in "Attributes > Local Dialect > Local Attributes " those modifications do not affect in "Console > mange > user profile" **How to reproduce:** Login to Console > Mange 1. Navigate to Attribute Dialects > Local Dialects > 2. select any Local Attribute 3. ex: Country and check > Show this attribute on user profile and user registration page option 4. Update 5. Verify that attribute displaying in the user profile, when new user creation ![image](https://user-images.githubusercontent.com/39120228/93249613-36dcd100-f7af-11ea-83bb-e156ac6d6275.png) Verify user profile after attribute update & while new user creation ![image](https://user-images.githubusercontent.com/39120228/93250139-fb8ed200-f7af-11ea-9c1e-272fe1d261ee.png) **Expected behavior:** The modification should be affected in the user profile as well as the User registration page. **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: 5.11 m36 snapshot - OS: Mac - Database: H2 - Userstore: LDAP ---
non_process
attributes local dialect local attributes modification does not affect in console mange user profile describe the issue when doing any update show this attribute on user profile and user registration page function in attributes local dialect local attributes those modifications do not affect in console mange user profile how to reproduce login to console mange navigate to attribute dialects local dialects select any local attribute ex country and check show this attribute on user profile and user registration page option update verify that attribute displaying in the user profile when new user creation verify user profile after attribute update while new user creation expected behavior the modification should be affected in the user profile as well as the user registration page environment information please complete the following information remove any unnecessary fields product version snapshot os mac database userstore ldap
0
9,126
3,022,769,555
IssuesEvent
2015-07-31 22:37:32
magnumripper/JohnTheRipper
https://api.github.com/repos/magnumripper/JohnTheRipper
opened
ntlmv2-opencl fails TS w/ UTF-8
testing
``` Invalid negative number (-1) in chr at ./jtrts.pl line 732.DONE [PASSED] Invalid negative number (-1) in chr at ./jtrts.pl line 732. .pot CHK:netntlmv2-opencl-utf8 guesses: 1442 -show=1442 0:00:00:08 DONE : Expected count(s) (1442)(1500) [!!!FAILED4!!!] (1440 val-pwd 2 inval-pwd) ```
1.0
ntlmv2-opencl fails TS w/ UTF-8 - ``` Invalid negative number (-1) in chr at ./jtrts.pl line 732.DONE [PASSED] Invalid negative number (-1) in chr at ./jtrts.pl line 732. .pot CHK:netntlmv2-opencl-utf8 guesses: 1442 -show=1442 0:00:00:08 DONE : Expected count(s) (1442)(1500) [!!!FAILED4!!!] (1440 val-pwd 2 inval-pwd) ```
non_process
opencl fails ts w utf invalid negative number in chr at jtrts pl line done invalid negative number in chr at jtrts pl line pot chk opencl guesses show done expected count s val pwd inval pwd
0
36,665
12,418,568,810
IssuesEvent
2020-05-23 01:00:06
wrbejar/wrbejar-Nova8Ksa
https://api.github.com/repos/wrbejar/wrbejar-Nova8Ksa
opened
CVE-2014-0054 (Medium) detected in spring-web-3.1.1.RELEASE.jar
security vulnerability
## CVE-2014-0054 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: /tmp/ws-scm/wrbejar-Nova8Ksa/ksa-web-core/pom.xml</p> <p>Path to vulnerable library: epository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/wrbejar-Nova8Ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Jaxb2RootElementHttpMessageConverter in Spring MVC in Spring Framework before 3.2.8 and 4.0.0 before 4.0.2 does not disable external entity resolution, which allows remote attackers to read arbitrary files, cause a denial of service, and conduct CSRF attacks via crafted XML, aka an XML External Entity (XXE) issue. NOTE: this vulnerability exists because of an incomplete fix for CVE-2013-4152, CVE-2013-7315, and CVE-2013-6429. <p>Publish Date: 2014-04-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0054>CVE-2014-0054</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-0054">https://nvd.nist.gov/vuln/detail/CVE-2014-0054</a></p> <p>Release Date: 2014-04-17</p> <p>Fix Resolution: org.springframework:spring-web:3.2.8.RELEASE,4.0.2.RELEASE,org.springframework:spring-oxm:4.0.2.RELEASE,3.2.8.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-web","packageVersion":"3.1.1.RELEASE","isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-web:3.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-web:3.2.8.RELEASE,4.0.2.RELEASE,org.springframework:spring-oxm:4.0.2.RELEASE,3.2.8.RELEASE"}],"vulnerabilityIdentifier":"CVE-2014-0054","vulnerabilityDetails":"The Jaxb2RootElementHttpMessageConverter in Spring MVC in Spring Framework before 3.2.8 and 4.0.0 before 4.0.2 does not disable external entity resolution, which allows remote attackers to read arbitrary files, cause a denial of service, and conduct CSRF attacks via crafted XML, aka an XML External Entity (XXE) issue. NOTE: this vulnerability exists because of an incomplete fix for CVE-2013-4152, CVE-2013-7315, and CVE-2013-6429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0054","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
True
CVE-2014-0054 (Medium) detected in spring-web-3.1.1.RELEASE.jar - ## CVE-2014-0054 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: /tmp/ws-scm/wrbejar-Nova8Ksa/ksa-web-core/pom.xml</p> <p>Path to vulnerable library: epository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/wrbejar-Nova8Ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Jaxb2RootElementHttpMessageConverter in Spring MVC in Spring Framework before 3.2.8 and 4.0.0 before 4.0.2 does not disable external entity resolution, which allows remote attackers to read arbitrary files, cause a denial of service, and conduct CSRF attacks via crafted XML, aka an XML External Entity (XXE) issue. NOTE: this vulnerability exists because of an incomplete fix for CVE-2013-4152, CVE-2013-7315, and CVE-2013-6429. <p>Publish Date: 2014-04-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0054>CVE-2014-0054</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-0054">https://nvd.nist.gov/vuln/detail/CVE-2014-0054</a></p> <p>Release Date: 2014-04-17</p> <p>Fix Resolution: org.springframework:spring-web:3.2.8.RELEASE,4.0.2.RELEASE,org.springframework:spring-oxm:4.0.2.RELEASE,3.2.8.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-web","packageVersion":"3.1.1.RELEASE","isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-web:3.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-web:3.2.8.RELEASE,4.0.2.RELEASE,org.springframework:spring-oxm:4.0.2.RELEASE,3.2.8.RELEASE"}],"vulnerabilityIdentifier":"CVE-2014-0054","vulnerabilityDetails":"The Jaxb2RootElementHttpMessageConverter in Spring MVC in Spring Framework before 3.2.8 and 4.0.0 before 4.0.2 does not disable external entity resolution, which allows remote attackers to read arbitrary files, cause a denial of service, and conduct CSRF attacks via crafted XML, aka an XML External Entity (XXE) issue. NOTE: this vulnerability exists because of an incomplete fix for CVE-2013-4152, CVE-2013-7315, and CVE-2013-6429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0054","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in spring web release jar cve medium severity vulnerability vulnerable library spring web release jar spring framework parent path to dependency file tmp ws scm wrbejar ksa web core pom xml path to vulnerable library epository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar wrbejar ksa web root ksa web target root web inf lib spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar dependency hierarchy x spring web release jar vulnerable library vulnerability details the in spring mvc in spring framework before and before does not disable external entity resolution which allows remote attackers to read arbitrary files cause a denial of service and conduct csrf attacks via crafted xml aka an xml external entity xxe issue note this vulnerability exists because of an incomplete fix for cve cve and cve publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution org springframework spring web release release org springframework spring oxm release release isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the in spring mvc in spring framework before and before does not disable external entity resolution which allows remote attackers to read arbitrary files cause a denial of service and conduct csrf attacks via crafted xml aka an xml external entity xxe issue note this vulnerability exists because of an incomplete fix for cve cve and cve vulnerabilityurl
0
31,986
12,054,236,803
IssuesEvent
2020-04-15 10:46:16
istio/istio
https://api.github.com/repos/istio/istio
closed
Example task throwing a SSL error: Plugging in existing CA Certificates
area/security kind/docs
**Bug description** I have been running the Security example tasks and all have worked perfectly but I keep seeing an error in the "Plugging in existing CA Certificates" task when executing step 2 [link](https://istio.io/docs/tasks/security/plugin-ca-cert/#verifying-the-certificates): ``` 2. Retrieve the certificate chain of httpbin: kubectl exec $(kubectl get pod -l app=sleep -n foo -o jsonpath={.items..metadata.name}) -c istio-proxy -n foo -- openssl s_client -showcerts -connect httpbin.foo:8000 > httpbin-proxy-cert.txt ``` The command returns an error: ``` depth=2 C = US, ST = California, L = Sunnyvale, O = Istio, OU = Test, CN = Root CA, emailAddress = testrootca@istio.io verify error:num=19:self signed certificate in certificate chain 139945834082752:error:1409445C:SSL routines:ssl3_read_bytes:tlsv13 alert certificate required:../ssl/record/rec_layer_s3.c:1528:SSL alert number 116 command terminated with exit code 1 ``` When I open `httpbin-proxy-cert.txt` I see the three certificates and I can continue the task without any problem, so that error does not impact the end result. **Expected behavior** Command executing without any error **Steps to reproduce the bug** `curl -L https://istio.io/downloadIstio | sh -` `cd istio-1.5.1` `export PATH=$PWD/bin:$PATH` `kubectl create namespace istio-system` `kubectl create secret generic cacerts -n istio-system --from-file=samples/certs/ca-cert.pem \ --from-file=samples/certs/ca-key.pem --from-file=samples/certs/root-cert.pem \ --from-file=samples/certs/cert-chain.pem` `istioctl manifest apply --set profile=demo --set values.global.mtls.enabled=true` `kubectl create ns foo` `kubectl apply -f <(istioctl kube-inject -f samples/httpbin/httpbin.yaml) -n foo` `kubectl apply -f <(istioctl kube-inject -f samples/sleep/sleep.yaml) -n foo` `kubectl exec $(kubectl get pod -l app=sleep -n foo -o jsonpath={.items..metadata.name}) -c istio-proxy -n foo -- openssl s_client -showcerts -connect httpbin.foo:8000 > httpbin-proxy-cert.txt` **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** ``` client version: 1.5.1 control plane version: 1.5.1 data plane version: 1.5.1 (5 proxies) ``` **How was Istio installed?** Through istioctl **Environment where bug was observed (cloud vendor, OS, etc)** Baremetal, openSUSE
True
Example task throwing a SSL error: Plugging in existing CA Certificates - **Bug description** I have been running the Security example tasks and all have worked perfectly but I keep seeing an error in the "Plugging in existing CA Certificates" task when executing step 2 [link](https://istio.io/docs/tasks/security/plugin-ca-cert/#verifying-the-certificates): ``` 2. Retrieve the certificate chain of httpbin: kubectl exec $(kubectl get pod -l app=sleep -n foo -o jsonpath={.items..metadata.name}) -c istio-proxy -n foo -- openssl s_client -showcerts -connect httpbin.foo:8000 > httpbin-proxy-cert.txt ``` The command returns an error: ``` depth=2 C = US, ST = California, L = Sunnyvale, O = Istio, OU = Test, CN = Root CA, emailAddress = testrootca@istio.io verify error:num=19:self signed certificate in certificate chain 139945834082752:error:1409445C:SSL routines:ssl3_read_bytes:tlsv13 alert certificate required:../ssl/record/rec_layer_s3.c:1528:SSL alert number 116 command terminated with exit code 1 ``` When I open `httpbin-proxy-cert.txt` I see the three certificates and I can continue the task without any problem, so that error does not impact the end result. **Expected behavior** Command executing without any error **Steps to reproduce the bug** `curl -L https://istio.io/downloadIstio | sh -` `cd istio-1.5.1` `export PATH=$PWD/bin:$PATH` `kubectl create namespace istio-system` `kubectl create secret generic cacerts -n istio-system --from-file=samples/certs/ca-cert.pem \ --from-file=samples/certs/ca-key.pem --from-file=samples/certs/root-cert.pem \ --from-file=samples/certs/cert-chain.pem` `istioctl manifest apply --set profile=demo --set values.global.mtls.enabled=true` `kubectl create ns foo` `kubectl apply -f <(istioctl kube-inject -f samples/httpbin/httpbin.yaml) -n foo` `kubectl apply -f <(istioctl kube-inject -f samples/sleep/sleep.yaml) -n foo` `kubectl exec $(kubectl get pod -l app=sleep -n foo -o jsonpath={.items..metadata.name}) -c istio-proxy -n foo -- openssl s_client -showcerts -connect httpbin.foo:8000 > httpbin-proxy-cert.txt` **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** ``` client version: 1.5.1 control plane version: 1.5.1 data plane version: 1.5.1 (5 proxies) ``` **How was Istio installed?** Through istioctl **Environment where bug was observed (cloud vendor, OS, etc)** Baremetal, openSUSE
non_process
example task throwing a ssl error plugging in existing ca certificates bug description i have been running the security example tasks and all have worked perfectly but i keep seeing an error in the plugging in existing ca certificates task when executing step retrieve the certificate chain of httpbin kubectl exec kubectl get pod l app sleep n foo o jsonpath items metadata name c istio proxy n foo openssl s client showcerts connect httpbin foo httpbin proxy cert txt the command returns an error depth c us st california l sunnyvale o istio ou test cn root ca emailaddress testrootca istio io verify error num self signed certificate in certificate chain error ssl routines read bytes alert certificate required ssl record rec layer c ssl alert number command terminated with exit code when i open httpbin proxy cert txt i see the three certificates and i can continue the task without any problem so that error does not impact the end result expected behavior command executing without any error steps to reproduce the bug curl l sh cd istio export path pwd bin path kubectl create namespace istio system kubectl create secret generic cacerts n istio system from file samples certs ca cert pem from file samples certs ca key pem from file samples certs root cert pem from file samples certs cert chain pem istioctl manifest apply set profile demo set values global mtls enabled true kubectl create ns foo kubectl apply f istioctl kube inject f samples httpbin httpbin yaml n foo kubectl apply f istioctl kube inject f samples sleep sleep yaml n foo kubectl exec kubectl get pod l app sleep n foo o jsonpath items metadata name c istio proxy n foo openssl s client showcerts connect httpbin foo httpbin proxy cert txt version include the output of istioctl version remote and kubectl version and helm version if you used helm client version control plane version data plane version proxies how was istio installed through istioctl environment where bug was observed cloud vendor os etc baremetal opensuse
0
101,286
21,640,683,855
IssuesEvent
2022-05-05 18:26:57
phetsims/geometric-optics
https://api.github.com/repos/phetsims/geometric-optics
closed
Code review
dev:code-review
# PhET Code-Review Checklist (a.k.a "CRC") * The responsible dev is responsible for removing the irrelevant parts (@pixelzoom: I've done this ~~using strikethrough text~~.) * A checked-off item doesn't mean "no problem here", it means "it was reviewed" * Problems can be noted in side issues that reference this issue, or through `// REVIEW` comments in the code ## GitHub Issues The following standard GitHub issues should exist. _If these issues are missing, or have not been completed, pause code review until the issues have been created and addressed by the responsible dev._ - [x] https://github.com/phetsims/geometric-optics/issues/134 -- Familiarize yourself with the model by reading model.md. Does it adequately describe the model, in terms appropriate for teachers? Has it been reviewed by the sim designer? - [x] https://github.com/phetsims/geometric-optics/issues/135 -- Familiarize yourself with the implementation by reading implementation-notes.md. Does it provide an overview that will be useful to future maintainers? - [x] https://github.com/phetsims/geometric-optics/issues/372 - [x] https://github.com/phetsims/geometric-optics/issues/372 - [x] https://github.com/phetsims/geometric-optics/issues/149 - [x] https://github.com/phetsims/geometric-optics/issues/147 - [x] https://github.com/phetsims/geometric-optics/issues/146 (will not be completed until after RC testing) ## **Build and Run Checks** If any of these items fail, pause code review. - [x] Does the sim build without warnings or errors? - [x] Does the html file size seem reasonable, compared to other similar sims? - [x] Does the sim start up? (unbuilt and built versions) - [x] Does the sim experience any assertion failures? (run with query parameter `ea`) - [x] Does the sim pass a scenery fuzz test? (run with query parameters `fuzz&ea`) - [x] Does the sim behave correctly when listener order is shuffled? (run with query parameters `ea&shuffleListeners` and `ea&shuffleListeners&fuzz`) - [x] Does the sim output any deprecation warnings? Run with `?deprecationWarnings`. Do not use deprecated methods in new code. ## **Memory Leaks** - [x] Does a heap comparison using Chrome Developer Tools indicate a memory leak? (Describing this process is beyond the scope of this document.) Test on a version built using `grunt --minify.mangle=false`. Compare to testing results done by the responsible developer. Results can be found in {{GITHUB_ISSUE_LINK}}. - [x] For each common-code component (sun, scenery-phet, vegas, …) that opaquely registers observers or listeners, is there a call to that component’s `dispose` function, or is it obvious why it isn't necessary, or is there documentation about why `dispose` isn't called? An example of why no call to `dispose` is needed is if the component is used in a `ScreenView` that would never be removed from the scene graph. Note that it's also acceptable (and encouraged!) to describe what needs to be disposed in implementation-notes.md. - [x] Are there leaks due to registering observers or listeners? The following guidelines should be followed unless documentation (in-line or in implementation-notes.md) describes why following them is not necessary. * AXON: `Property.link` or `lazyLink` is accompanied by `unlink`. * AXON: `Property.multilink` is accompanied by `unmultilink`. * AXON: Creation of `Multilink` is accompanied by `dispose`. * AXON: Creation of `DerivedProperty` is accompanied by `dispose`. * AXON: `Emitter.addListener` is accompanied by `removeListener`. * AXON: `ObservableArrayDef.element*Emitter.addListener` is accompanied by `ObservableArrayDef.element*Emitter.removeListener` * SCENERY: `Node.addInputListener` is accompanied by `removeInputListener` * TANDEM: Creation of an instrumented `PhetioObject` is accompanied by `dispose`. - [x] Do all types that require a `dispose` function have one? This should expose a public `dispose` function that calls `this.disposeMyType()`, where `disposeMyType` is a private function declared in the constructor. `MyType` should exactly match the filename. ## **Performance** - [x] Play with sim, identify any obvious performance issues. Examples: animation that slows down with large numbers of objects; animation that pauses or "hitches" during garbage collection. - [x] ~~If the sim uses WebGL, does it have a fallback? Does the fallback perform reasonably well? (run with query parameter `webgl=false`)~~ ## **Usability** - [x] Are UI components sufficiently responsive? (especially continuous UI components, such as sliders) - [x] Are pointer areas optimized, especially for touch? (run with query parameter `showPointerAreas`) - [x] Do pointer areas overlap? (run with query parameter `showPointerAreas`) Overlap may be OK in some cases, depending on the z-ordering (if the front-most object is supposed to occlude pointer areas) and whether objects can be moved. ## **Internationalization** - [x] Are there any strings that are not internationalized, and does the sim layout gracefully handle internationalized strings that are shorter than the English strings? (run with query parameter `stringTest=X`. You should see nothing but 'X' strings.) - [x] Does the sim layout gracefully handle internationalized strings that are longer than the English strings? (run with query parameters `stringTest=double` and `stringTest=long`) - [x] Does the sim stay on the sim page (doesn't redirect to an external page) when running with the query parameter `stringTest=xss`? This test passes if sim does not redirect, OK if sim crashes or fails to fully start. Only test on one desktop platform. For PhET-iO sims, additionally test `?stringTest=xss` in Studio to make sure i18n strings didn't leak to phetioDocumentation, see https://github.com/phetsims/phet-io/issues/1377 - [x] Avoid using concatenation to create strings that will be visible in the user interface. Use `StringUtils.fillIn` and a string pattern to ensure that strings are properly localized. - [x] Use named placeholders (e.g. `"{{value}} {{units}}"`) instead of numbered placeholders (e.g. `"{0} {1}"`). - [x] Make sure the string keys are all perfect, because they are difficult to change after 1.0.0 is published. Guidelines for string keys are: (1) Strings keys should generally match their values. E.g.: ```js "helloWorld": { value: "Hello World!" }, "quadraticTerms": { value: "Quadratic Terms" } ``` (2) If a string key would be exceptionally long, use a key name that is an abbreviated form of the string value, or that captures the purpose/essence of the value. E.g.: ```js // key is abbreviated "iWentToTheStore": { value: "I went to the store to get milk, eggs, butter, and sugar." }, // key is based on purpose "describeTheScreen": { value: "The Play Area is a small room. The Control Panel has buttons, a checkbox, and radio buttons to change conditions in the room." } ``` (3) If string key names would collide, use your judgment to disambiguate. E.g.: ```js "simplifyTitle": { value: "Simplify!" }, "simplifyCheckbox": { value: "simplify" } ``` (4) String keys for screen names should have the general form `"screen.{{screenName}}"`. E.g.: ```js "screen.explore": { "value": "Explore" }, ``` (5) String patterns that contain placeholders (e.g. `"My name is {{first}} {{last}}"`) should use keys that are unlikely to conflict with strings that might be needed in the future. For example, for `"{{price}}"` consider using key `"pricePattern"` instead of `"price"`, if you think there might be a future need for a `"price"` string. (6) It is acceptable to prefix families of strings with a prefix, like so: ```json "material.water": { "value": "Water" }, "material.wood": { "value": "Wood" }, "shape.block": { "value": "Block" }, "shape.cone": { "value": "Cone" }, ``` Nested substructure is not yet fully supported. - [x] ~~If the sim was already released, make sure none of the original string keys have changed. If they have changed, make sure any changes have a good reason and have been discussed with @jbphet (it is likely that an issue like https://github.com/phetsims/gravity-force-lab/issues/166 should be created).~~ ## **Repository Structure** - [x] #434 - [x] Are all required files and directories present? For a sim repository named “my-repo”, the general structure should look like this (where assets/, images/, mipmaps/ or sounds/ may be omitted if the sim doesn’t have those types of resource files). ``` my-repo/ assets/ doc/ images/ *see annotation model.md implementation-notes.md images/ license.json js/ (see section below) mipmaps/ license.json sound/ license.json dependencies.json .gitignore my-repo_en.html my-repo-strings_en.json Gruntfile.js LICENSE package.json README.md ``` *Any images used in model.md or implementation-notes.md should be added here. Images specific to aiding with documentation do not need their own license. - [x] ~~Verify that the same image file is not present in both images/ and mipmaps/. If you need a mipmap, use it for all occurrences of the image.~~ - [x] Is the js/ directory properly structured? All JavaScript source should be in the js/ directory. There should be a subdirectory for each screen (this also applies for single-screen sims, where the subdirectory matches the repo name). For a multi-screen sim, code shared by 2 or more screens should be in a js/common/ subdirectory. Model and view code should be in model/ and view/ subdirectories for each screen and common/. For example, for a sim with screens “Introduction” and “Lab”, the general directory structure should look like this: ``` my-repo/ js/ common/ model/ view/ introduction/ model/ view/ lab/ model/ view/ my-repo-main.js myRepo.js myRepoStrings.js ``` - [x] Do filenames use an appropriate prefix? Some filenames may be prefixed with the repository name, e.g. `MolarityConstants.js` in molarity. If the repository name is long, the developer may choose to abbreviate the repository name, e.g. `EEConstants.js` in expression-exchange. If the abbreviation is already used by another respository, then the full name must be used. For example, if the "EE" abbreviation is already used by expression-exchange, then it should not be used in equality-explorer. Whichever convention is used, it should be used consistently within a repository - don't mix abbreviations and full names. - [x] Is there a file in assets/ for every resource file in sound/ and images/? Note that there is *not necessarily* a 1:1 correspondence between asset and resource files; for example, several related images may be in the same .ai file. Check license.json for possible documentation of why some reesources might not have a corresponding asset file. - [x] For simulations, was the README.md generated by `grunt published-README` or `grunt unpublished-README`? Common code repos can have custom README files. - [x] Does package.json refer to any dependencies that are not used by the sim? - [x] Is the LICENSE file correct? (Generally GPL v3 for sims and MIT for common code, see [this thread](https://github.com/phetsims/tasks/issues/875#issuecomment-312168646) for additional information). - [x] Does .gitignore match the one in simula-rasa? - [x] In GitHub, verify that all non-release branches have an associated issue that describes their purpose. - [x] Are there any GitHub branches that are no longer needed and should be deleted? - [x] Sim-specific query parameters (if any) should be identified and documented in one .js file in js/common/ or js/ (if there is no common/). The .js file should be named `{{PREFIX}}QueryParameters.js`, for example ArithmeticQueryParameters.js for the aritmetic repository, or FBQueryParameters.js for Function Builder (where the `FB` prefix is used). - [x] Query parameters that are public-facing should be identified using `public: true` in the schema. - [x] All sims should use a color file named `MyRepoColors.js` or, if using abbreviations, `MRColors.js`, and use `ProfileColorProperty` where appropriate, even if they have a single (default) profile (to support color editing and PhET-iO Studio). The `ColorProfile` pattern was converted to `*Colors.js` files in https://github.com/phetsims/scenery-phet/issues/515. Please see [GasPropertiesColors.js](https://github.com/phetsims/gas-properties/blob/master/js/common/GasPropertiesColors.js) for a good example. ## **Coding Conventions** - [x] Are coding conventions outlined in [PhET's Coding Conventions Document](https://github.com/phetsims/phet-info/blob/master/doc/coding-conventions.md) followed and adhered to? This document deals with PhET coding conventions. You do not need to exhaustively check every item in this section, nor do you necessarily need to check these items one at a time. The goal is to determine whether the code generally meets PhET standards. ## **TypeScript Conventions** - [x] Are TypeScript conventions outlined in the [TypeScript Conventions Document](https://github.com/phetsims/phet-info/blob/master/doc/typescript-conventions.md) followed and adhered to? ## **Math Libraries** - [x] `DOT/Utils.toFixed` or `DOT/Utils.toFixedNumber` should be used instead of `toFixed`. JavaScript's `toFixed` is notoriously buggy. Behavior differs depending on browser, because the spec doesn't specify whether to round or floor. ## IE11 - [x] ~~IE is no longer supported. With that in mind remove IE-specific workarounds~~ - [x] Use `string.includes` and `string.startsWith` where possible. ## **Organization, Readability, and Maintainability** - [x] Does the organization and structure of the code make sense? Do the model and view contain types that you would expect (or guess!) by looking at the sim? Do the names of things correspond to the names that you see in the user interface? - [x] Are appropriate design patterns used? See [phet-software-design-patterns.md](https://github.com/phetsims/phet-info/blob/master/doc/phet-software-design-patterns.md). If new or inappropriate patterns are identified, create an issue. - [x] Is inheritance used where appropriate? Does the type hierarchy make sense? - [x] Is composition favored over inheritance where appropriate? See https://en.wikipedia.org/wiki/Composition_over_inheritance. - [x] Is there any unnecessary coupling? (e.g., by passing large objects to constructors, or exposing unnecessary properties/functions) - [x] Is there too much unnecessary decoupling? (e.g. by passing all of the properties of an object independently instead of passing the object itself)? - [x] Are the source files reasonable in size? Scrutinize large files with too many responsibilities - can responsibilities be broken into smaller delegates? - [x] Are any significant chunks of code duplicated? In addition to manual identification, tools include: WebStorm _Code > Analyze Code > Locate Duplicates_ and https://github.com/danielstjules/jsinspect. - [x] Is there anything that should be generalized and migrated to common code? - [x] Are there any `TODO` or `FIXME` or `REVIEW` comments in the code? They should be addressed or promoted to GitHub issues. - [x] Are there any [magic numbers](https://en.wikipedia.org/wiki/Magic_number_(programming)) that should be factored out as constants and documented? - [x] Are there any constants that are duplicated in multiple files that should be factored out into a `{{REPO}}Constants.js` file? - [x] Does the implementation rely on any specific constant values that are likely to change in the future? Identify constants that might be changed in the future. (Use your judgement about which constants are likely candidates.) Does changing the values of these constants break the sim? For example, see https://github.com/phetsims/plinko-probability/issues/84. - [x] Is [PhetColorScheme](https://github.com/phetsims/scenery-phet/blob/master/js/PhetColorScheme.js) used where appropriate? Verify that the sim is not inventing/creating its own colors for things that have been standardized in `PhetColorScheme`. Identify any colors that might be worth adding to `PhetColorScheme`. - [x] Are all dependent Properties modeled as `DerivedProperty` instead of `Property`? - [x] All dynamics should be called from Sim.step(dt), do not use window.setTimeout or window.setInterval. This will help support Legends of Learning and PhET-iO. ## **Accessibility** This section may be omitted if the sim has not been instrumented with accessibility features. Accessibility includes various features, not all are always include. Ignore sections that do not apply. ### General - [x] Are accessibility features integrated well into the code. They should be added in a maintainable way, even if that requires upfront refactoring. ### Alternative Input - [x] Does the sim pass an accessibility fuzz test? (run with query parameters `fuzzBoard&ea`) - [x] ~~Does this sim use specific keyboard shortcuts that overlap with global shortcuts? This includes the use of modifier keys like in https://github.com/phetsims/ratio-and-proportion/issues/287.~~ **NOTE: There is currently no list of global shortcuts, and therefore no way to complete this checklist item. See https://github.com/phetsims/phet-info/issues/188.** ### ~~Interactive Description~~ - [x] ~~Run the entire built sim HTML file through an [HTML validator](https://validator.w3.org/nu/#textarea), does the HTML pass?~~ - [x] ~~If applicable, are good design patterns used for interactive description, see [interactive-description-technical-guide.md](https://github.com/phetsims/phet-info/blob/master/doc/interactive-description-technical-guide.md)~~ - [x] ~~Does resetting the simulation also reset the entire PDOM?~~ - [x] ~~Is `Node.pdomOrder` used appropriately to maintain visual and PDOM layout balance?~~ - [x] ~~Make sure accessibility strings aren't being adjusted with ascii specific javascript methods like `toUpperCase()`. Remember that one day these strings will be translatable~~ - [x] ~~Make sure for accessibility strings that all end of sentence periods do not have a leading space before it. Some screen readers will read these as "dot." This can occur often when a clause is conditionally added.~~ ## **PhET-iO** This section may be omitted if the sim has not been instrumented for PhET-iO, but is likely good to glance at no matter. - [x] Does instrumentation follow the conventions described in [PhET-iO Instrumentation Guide](https://github.com/phetsims/phet-io/blob/master/doc/phet-io-instrumentation-technical-guide.md)? This could be an extensive bullet. At the very least, be sure to know what amount of instrumentation this sim supports. Describing this further goes beyond the scope of this document. - [x] PhET-iO instantiates different objects and wires up listeners that are not present in the PhET-branded simulation. It needs to be tested separately for memory leaks. To help isolate the nature of the memory leak, this test should be run separately from the PhET brand memory leak test. Test with a colorized Data Stream, and Studio (easily accessed from phetmarks). Compare to testing results done by the responsible developer and previous releases. - [x] ~~Make sure unused `PhetioObject` instances are disposed, which unregisters their tandems.~~ - [x] Make sure JOIST `dt` values are used instead of `Date.now()` or other Date functions. Perhaps try `phet.joist.elapsedTime`. Though this has already been mentioned, it is necessary for reproducible playback via input events and deserves a comment in this PhET-iO section. - [x] Are random numbers using `DOT/dotRandom` as an imported module (not a global), and all doing so after modules are declared (non-statically)? For example, the following methods (and perhaps others) should not be used: `Math.random`, `_.shuffle`, `_.sample`, `_.random`. This also deserves re-iteration due to its effect on record/playback for PhET-iO. - [x] ~~Like JSON, keys for `undefined` values are omitted when serializing objects across frames. Consider this when determining whether `toStateObject` should use `null` or `undefined` values.~~ - [x] PhET prefers to use the term "position" to refer to the physical (x,y) position of objects. This applies to both brands, but is more important for the PhET-iO API. See https://github.com/phetsims/phet-info/issues/126 - [x] ~~Are your IOType state methods violating the API of the core type by accessing private fields?~~ - [x] ~~When defining a boolean Property to indicate whether something is enabled, use `AXON/EnabledProperty`. This should be done in both the model and the view. If you're using a DerivedProperty, skip this item.~~ - [x] Do not use translated strings in `phetioDocumentaton` - it changes the PhET-iO API!
1.0
Code review - # PhET Code-Review Checklist (a.k.a "CRC") * The responsible dev is responsible for removing the irrelevant parts (@pixelzoom: I've done this ~~using strikethrough text~~.) * A checked-off item doesn't mean "no problem here", it means "it was reviewed" * Problems can be noted in side issues that reference this issue, or through `// REVIEW` comments in the code ## GitHub Issues The following standard GitHub issues should exist. _If these issues are missing, or have not been completed, pause code review until the issues have been created and addressed by the responsible dev._ - [x] https://github.com/phetsims/geometric-optics/issues/134 -- Familiarize yourself with the model by reading model.md. Does it adequately describe the model, in terms appropriate for teachers? Has it been reviewed by the sim designer? - [x] https://github.com/phetsims/geometric-optics/issues/135 -- Familiarize yourself with the implementation by reading implementation-notes.md. Does it provide an overview that will be useful to future maintainers? - [x] https://github.com/phetsims/geometric-optics/issues/372 - [x] https://github.com/phetsims/geometric-optics/issues/372 - [x] https://github.com/phetsims/geometric-optics/issues/149 - [x] https://github.com/phetsims/geometric-optics/issues/147 - [x] https://github.com/phetsims/geometric-optics/issues/146 (will not be completed until after RC testing) ## **Build and Run Checks** If any of these items fail, pause code review. - [x] Does the sim build without warnings or errors? - [x] Does the html file size seem reasonable, compared to other similar sims? - [x] Does the sim start up? (unbuilt and built versions) - [x] Does the sim experience any assertion failures? (run with query parameter `ea`) - [x] Does the sim pass a scenery fuzz test? (run with query parameters `fuzz&ea`) - [x] Does the sim behave correctly when listener order is shuffled? (run with query parameters `ea&shuffleListeners` and `ea&shuffleListeners&fuzz`) - [x] Does the sim output any deprecation warnings? Run with `?deprecationWarnings`. Do not use deprecated methods in new code. ## **Memory Leaks** - [x] Does a heap comparison using Chrome Developer Tools indicate a memory leak? (Describing this process is beyond the scope of this document.) Test on a version built using `grunt --minify.mangle=false`. Compare to testing results done by the responsible developer. Results can be found in {{GITHUB_ISSUE_LINK}}. - [x] For each common-code component (sun, scenery-phet, vegas, …) that opaquely registers observers or listeners, is there a call to that component’s `dispose` function, or is it obvious why it isn't necessary, or is there documentation about why `dispose` isn't called? An example of why no call to `dispose` is needed is if the component is used in a `ScreenView` that would never be removed from the scene graph. Note that it's also acceptable (and encouraged!) to describe what needs to be disposed in implementation-notes.md. - [x] Are there leaks due to registering observers or listeners? The following guidelines should be followed unless documentation (in-line or in implementation-notes.md) describes why following them is not necessary. * AXON: `Property.link` or `lazyLink` is accompanied by `unlink`. * AXON: `Property.multilink` is accompanied by `unmultilink`. * AXON: Creation of `Multilink` is accompanied by `dispose`. * AXON: Creation of `DerivedProperty` is accompanied by `dispose`. * AXON: `Emitter.addListener` is accompanied by `removeListener`. * AXON: `ObservableArrayDef.element*Emitter.addListener` is accompanied by `ObservableArrayDef.element*Emitter.removeListener` * SCENERY: `Node.addInputListener` is accompanied by `removeInputListener` * TANDEM: Creation of an instrumented `PhetioObject` is accompanied by `dispose`. - [x] Do all types that require a `dispose` function have one? This should expose a public `dispose` function that calls `this.disposeMyType()`, where `disposeMyType` is a private function declared in the constructor. `MyType` should exactly match the filename. ## **Performance** - [x] Play with sim, identify any obvious performance issues. Examples: animation that slows down with large numbers of objects; animation that pauses or "hitches" during garbage collection. - [x] ~~If the sim uses WebGL, does it have a fallback? Does the fallback perform reasonably well? (run with query parameter `webgl=false`)~~ ## **Usability** - [x] Are UI components sufficiently responsive? (especially continuous UI components, such as sliders) - [x] Are pointer areas optimized, especially for touch? (run with query parameter `showPointerAreas`) - [x] Do pointer areas overlap? (run with query parameter `showPointerAreas`) Overlap may be OK in some cases, depending on the z-ordering (if the front-most object is supposed to occlude pointer areas) and whether objects can be moved. ## **Internationalization** - [x] Are there any strings that are not internationalized, and does the sim layout gracefully handle internationalized strings that are shorter than the English strings? (run with query parameter `stringTest=X`. You should see nothing but 'X' strings.) - [x] Does the sim layout gracefully handle internationalized strings that are longer than the English strings? (run with query parameters `stringTest=double` and `stringTest=long`) - [x] Does the sim stay on the sim page (doesn't redirect to an external page) when running with the query parameter `stringTest=xss`? This test passes if sim does not redirect, OK if sim crashes or fails to fully start. Only test on one desktop platform. For PhET-iO sims, additionally test `?stringTest=xss` in Studio to make sure i18n strings didn't leak to phetioDocumentation, see https://github.com/phetsims/phet-io/issues/1377 - [x] Avoid using concatenation to create strings that will be visible in the user interface. Use `StringUtils.fillIn` and a string pattern to ensure that strings are properly localized. - [x] Use named placeholders (e.g. `"{{value}} {{units}}"`) instead of numbered placeholders (e.g. `"{0} {1}"`). - [x] Make sure the string keys are all perfect, because they are difficult to change after 1.0.0 is published. Guidelines for string keys are: (1) Strings keys should generally match their values. E.g.: ```js "helloWorld": { value: "Hello World!" }, "quadraticTerms": { value: "Quadratic Terms" } ``` (2) If a string key would be exceptionally long, use a key name that is an abbreviated form of the string value, or that captures the purpose/essence of the value. E.g.: ```js // key is abbreviated "iWentToTheStore": { value: "I went to the store to get milk, eggs, butter, and sugar." }, // key is based on purpose "describeTheScreen": { value: "The Play Area is a small room. The Control Panel has buttons, a checkbox, and radio buttons to change conditions in the room." } ``` (3) If string key names would collide, use your judgment to disambiguate. E.g.: ```js "simplifyTitle": { value: "Simplify!" }, "simplifyCheckbox": { value: "simplify" } ``` (4) String keys for screen names should have the general form `"screen.{{screenName}}"`. E.g.: ```js "screen.explore": { "value": "Explore" }, ``` (5) String patterns that contain placeholders (e.g. `"My name is {{first}} {{last}}"`) should use keys that are unlikely to conflict with strings that might be needed in the future. For example, for `"{{price}}"` consider using key `"pricePattern"` instead of `"price"`, if you think there might be a future need for a `"price"` string. (6) It is acceptable to prefix families of strings with a prefix, like so: ```json "material.water": { "value": "Water" }, "material.wood": { "value": "Wood" }, "shape.block": { "value": "Block" }, "shape.cone": { "value": "Cone" }, ``` Nested substructure is not yet fully supported. - [x] ~~If the sim was already released, make sure none of the original string keys have changed. If they have changed, make sure any changes have a good reason and have been discussed with @jbphet (it is likely that an issue like https://github.com/phetsims/gravity-force-lab/issues/166 should be created).~~ ## **Repository Structure** - [x] #434 - [x] Are all required files and directories present? For a sim repository named “my-repo”, the general structure should look like this (where assets/, images/, mipmaps/ or sounds/ may be omitted if the sim doesn’t have those types of resource files). ``` my-repo/ assets/ doc/ images/ *see annotation model.md implementation-notes.md images/ license.json js/ (see section below) mipmaps/ license.json sound/ license.json dependencies.json .gitignore my-repo_en.html my-repo-strings_en.json Gruntfile.js LICENSE package.json README.md ``` *Any images used in model.md or implementation-notes.md should be added here. Images specific to aiding with documentation do not need their own license. - [x] ~~Verify that the same image file is not present in both images/ and mipmaps/. If you need a mipmap, use it for all occurrences of the image.~~ - [x] Is the js/ directory properly structured? All JavaScript source should be in the js/ directory. There should be a subdirectory for each screen (this also applies for single-screen sims, where the subdirectory matches the repo name). For a multi-screen sim, code shared by 2 or more screens should be in a js/common/ subdirectory. Model and view code should be in model/ and view/ subdirectories for each screen and common/. For example, for a sim with screens “Introduction” and “Lab”, the general directory structure should look like this: ``` my-repo/ js/ common/ model/ view/ introduction/ model/ view/ lab/ model/ view/ my-repo-main.js myRepo.js myRepoStrings.js ``` - [x] Do filenames use an appropriate prefix? Some filenames may be prefixed with the repository name, e.g. `MolarityConstants.js` in molarity. If the repository name is long, the developer may choose to abbreviate the repository name, e.g. `EEConstants.js` in expression-exchange. If the abbreviation is already used by another respository, then the full name must be used. For example, if the "EE" abbreviation is already used by expression-exchange, then it should not be used in equality-explorer. Whichever convention is used, it should be used consistently within a repository - don't mix abbreviations and full names. - [x] Is there a file in assets/ for every resource file in sound/ and images/? Note that there is *not necessarily* a 1:1 correspondence between asset and resource files; for example, several related images may be in the same .ai file. Check license.json for possible documentation of why some reesources might not have a corresponding asset file. - [x] For simulations, was the README.md generated by `grunt published-README` or `grunt unpublished-README`? Common code repos can have custom README files. - [x] Does package.json refer to any dependencies that are not used by the sim? - [x] Is the LICENSE file correct? (Generally GPL v3 for sims and MIT for common code, see [this thread](https://github.com/phetsims/tasks/issues/875#issuecomment-312168646) for additional information). - [x] Does .gitignore match the one in simula-rasa? - [x] In GitHub, verify that all non-release branches have an associated issue that describes their purpose. - [x] Are there any GitHub branches that are no longer needed and should be deleted? - [x] Sim-specific query parameters (if any) should be identified and documented in one .js file in js/common/ or js/ (if there is no common/). The .js file should be named `{{PREFIX}}QueryParameters.js`, for example ArithmeticQueryParameters.js for the aritmetic repository, or FBQueryParameters.js for Function Builder (where the `FB` prefix is used). - [x] Query parameters that are public-facing should be identified using `public: true` in the schema. - [x] All sims should use a color file named `MyRepoColors.js` or, if using abbreviations, `MRColors.js`, and use `ProfileColorProperty` where appropriate, even if they have a single (default) profile (to support color editing and PhET-iO Studio). The `ColorProfile` pattern was converted to `*Colors.js` files in https://github.com/phetsims/scenery-phet/issues/515. Please see [GasPropertiesColors.js](https://github.com/phetsims/gas-properties/blob/master/js/common/GasPropertiesColors.js) for a good example. ## **Coding Conventions** - [x] Are coding conventions outlined in [PhET's Coding Conventions Document](https://github.com/phetsims/phet-info/blob/master/doc/coding-conventions.md) followed and adhered to? This document deals with PhET coding conventions. You do not need to exhaustively check every item in this section, nor do you necessarily need to check these items one at a time. The goal is to determine whether the code generally meets PhET standards. ## **TypeScript Conventions** - [x] Are TypeScript conventions outlined in the [TypeScript Conventions Document](https://github.com/phetsims/phet-info/blob/master/doc/typescript-conventions.md) followed and adhered to? ## **Math Libraries** - [x] `DOT/Utils.toFixed` or `DOT/Utils.toFixedNumber` should be used instead of `toFixed`. JavaScript's `toFixed` is notoriously buggy. Behavior differs depending on browser, because the spec doesn't specify whether to round or floor. ## IE11 - [x] ~~IE is no longer supported. With that in mind remove IE-specific workarounds~~ - [x] Use `string.includes` and `string.startsWith` where possible. ## **Organization, Readability, and Maintainability** - [x] Does the organization and structure of the code make sense? Do the model and view contain types that you would expect (or guess!) by looking at the sim? Do the names of things correspond to the names that you see in the user interface? - [x] Are appropriate design patterns used? See [phet-software-design-patterns.md](https://github.com/phetsims/phet-info/blob/master/doc/phet-software-design-patterns.md). If new or inappropriate patterns are identified, create an issue. - [x] Is inheritance used where appropriate? Does the type hierarchy make sense? - [x] Is composition favored over inheritance where appropriate? See https://en.wikipedia.org/wiki/Composition_over_inheritance. - [x] Is there any unnecessary coupling? (e.g., by passing large objects to constructors, or exposing unnecessary properties/functions) - [x] Is there too much unnecessary decoupling? (e.g. by passing all of the properties of an object independently instead of passing the object itself)? - [x] Are the source files reasonable in size? Scrutinize large files with too many responsibilities - can responsibilities be broken into smaller delegates? - [x] Are any significant chunks of code duplicated? In addition to manual identification, tools include: WebStorm _Code > Analyze Code > Locate Duplicates_ and https://github.com/danielstjules/jsinspect. - [x] Is there anything that should be generalized and migrated to common code? - [x] Are there any `TODO` or `FIXME` or `REVIEW` comments in the code? They should be addressed or promoted to GitHub issues. - [x] Are there any [magic numbers](https://en.wikipedia.org/wiki/Magic_number_(programming)) that should be factored out as constants and documented? - [x] Are there any constants that are duplicated in multiple files that should be factored out into a `{{REPO}}Constants.js` file? - [x] Does the implementation rely on any specific constant values that are likely to change in the future? Identify constants that might be changed in the future. (Use your judgement about which constants are likely candidates.) Does changing the values of these constants break the sim? For example, see https://github.com/phetsims/plinko-probability/issues/84. - [x] Is [PhetColorScheme](https://github.com/phetsims/scenery-phet/blob/master/js/PhetColorScheme.js) used where appropriate? Verify that the sim is not inventing/creating its own colors for things that have been standardized in `PhetColorScheme`. Identify any colors that might be worth adding to `PhetColorScheme`. - [x] Are all dependent Properties modeled as `DerivedProperty` instead of `Property`? - [x] All dynamics should be called from Sim.step(dt), do not use window.setTimeout or window.setInterval. This will help support Legends of Learning and PhET-iO. ## **Accessibility** This section may be omitted if the sim has not been instrumented with accessibility features. Accessibility includes various features, not all are always include. Ignore sections that do not apply. ### General - [x] Are accessibility features integrated well into the code. They should be added in a maintainable way, even if that requires upfront refactoring. ### Alternative Input - [x] Does the sim pass an accessibility fuzz test? (run with query parameters `fuzzBoard&ea`) - [x] ~~Does this sim use specific keyboard shortcuts that overlap with global shortcuts? This includes the use of modifier keys like in https://github.com/phetsims/ratio-and-proportion/issues/287.~~ **NOTE: There is currently no list of global shortcuts, and therefore no way to complete this checklist item. See https://github.com/phetsims/phet-info/issues/188.** ### ~~Interactive Description~~ - [x] ~~Run the entire built sim HTML file through an [HTML validator](https://validator.w3.org/nu/#textarea), does the HTML pass?~~ - [x] ~~If applicable, are good design patterns used for interactive description, see [interactive-description-technical-guide.md](https://github.com/phetsims/phet-info/blob/master/doc/interactive-description-technical-guide.md)~~ - [x] ~~Does resetting the simulation also reset the entire PDOM?~~ - [x] ~~Is `Node.pdomOrder` used appropriately to maintain visual and PDOM layout balance?~~ - [x] ~~Make sure accessibility strings aren't being adjusted with ascii specific javascript methods like `toUpperCase()`. Remember that one day these strings will be translatable~~ - [x] ~~Make sure for accessibility strings that all end of sentence periods do not have a leading space before it. Some screen readers will read these as "dot." This can occur often when a clause is conditionally added.~~ ## **PhET-iO** This section may be omitted if the sim has not been instrumented for PhET-iO, but is likely good to glance at no matter. - [x] Does instrumentation follow the conventions described in [PhET-iO Instrumentation Guide](https://github.com/phetsims/phet-io/blob/master/doc/phet-io-instrumentation-technical-guide.md)? This could be an extensive bullet. At the very least, be sure to know what amount of instrumentation this sim supports. Describing this further goes beyond the scope of this document. - [x] PhET-iO instantiates different objects and wires up listeners that are not present in the PhET-branded simulation. It needs to be tested separately for memory leaks. To help isolate the nature of the memory leak, this test should be run separately from the PhET brand memory leak test. Test with a colorized Data Stream, and Studio (easily accessed from phetmarks). Compare to testing results done by the responsible developer and previous releases. - [x] ~~Make sure unused `PhetioObject` instances are disposed, which unregisters their tandems.~~ - [x] Make sure JOIST `dt` values are used instead of `Date.now()` or other Date functions. Perhaps try `phet.joist.elapsedTime`. Though this has already been mentioned, it is necessary for reproducible playback via input events and deserves a comment in this PhET-iO section. - [x] Are random numbers using `DOT/dotRandom` as an imported module (not a global), and all doing so after modules are declared (non-statically)? For example, the following methods (and perhaps others) should not be used: `Math.random`, `_.shuffle`, `_.sample`, `_.random`. This also deserves re-iteration due to its effect on record/playback for PhET-iO. - [x] ~~Like JSON, keys for `undefined` values are omitted when serializing objects across frames. Consider this when determining whether `toStateObject` should use `null` or `undefined` values.~~ - [x] PhET prefers to use the term "position" to refer to the physical (x,y) position of objects. This applies to both brands, but is more important for the PhET-iO API. See https://github.com/phetsims/phet-info/issues/126 - [x] ~~Are your IOType state methods violating the API of the core type by accessing private fields?~~ - [x] ~~When defining a boolean Property to indicate whether something is enabled, use `AXON/EnabledProperty`. This should be done in both the model and the view. If you're using a DerivedProperty, skip this item.~~ - [x] Do not use translated strings in `phetioDocumentaton` - it changes the PhET-iO API!
non_process
code review phet code review checklist a k a crc the responsible dev is responsible for removing the irrelevant parts pixelzoom i ve done this using strikethrough text a checked off item doesn t mean no problem here it means it was reviewed problems can be noted in side issues that reference this issue or through review comments in the code github issues the following standard github issues should exist if these issues are missing or have not been completed pause code review until the issues have been created and addressed by the responsible dev familiarize yourself with the model by reading model md does it adequately describe the model in terms appropriate for teachers has it been reviewed by the sim designer familiarize yourself with the implementation by reading implementation notes md does it provide an overview that will be useful to future maintainers will not be completed until after rc testing build and run checks if any of these items fail pause code review does the sim build without warnings or errors does the html file size seem reasonable compared to other similar sims does the sim start up unbuilt and built versions does the sim experience any assertion failures run with query parameter ea does the sim pass a scenery fuzz test run with query parameters fuzz ea does the sim behave correctly when listener order is shuffled run with query parameters ea shufflelisteners and ea shufflelisteners fuzz does the sim output any deprecation warnings run with deprecationwarnings do not use deprecated methods in new code memory leaks does a heap comparison using chrome developer tools indicate a memory leak describing this process is beyond the scope of this document test on a version built using grunt minify mangle false compare to testing results done by the responsible developer results can be found in github issue link for each common code component sun scenery phet vegas … that opaquely registers observers or listeners is there a call to that component’s dispose function or is it obvious why it isn t necessary or is there documentation about why dispose isn t called an example of why no call to dispose is needed is if the component is used in a screenview that would never be removed from the scene graph note that it s also acceptable and encouraged to describe what needs to be disposed in implementation notes md are there leaks due to registering observers or listeners the following guidelines should be followed unless documentation in line or in implementation notes md describes why following them is not necessary axon property link or lazylink is accompanied by unlink axon property multilink is accompanied by unmultilink axon creation of multilink is accompanied by dispose axon creation of derivedproperty is accompanied by dispose axon emitter addlistener is accompanied by removelistener axon observablearraydef element emitter addlistener is accompanied by observablearraydef element emitter removelistener scenery node addinputlistener is accompanied by removeinputlistener tandem creation of an instrumented phetioobject is accompanied by dispose do all types that require a dispose function have one this should expose a public dispose function that calls this disposemytype where disposemytype is a private function declared in the constructor mytype should exactly match the filename performance play with sim identify any obvious performance issues examples animation that slows down with large numbers of objects animation that pauses or hitches during garbage collection if the sim uses webgl does it have a fallback does the fallback perform reasonably well run with query parameter webgl false usability are ui components sufficiently responsive especially continuous ui components such as sliders are pointer areas optimized especially for touch run with query parameter showpointerareas do pointer areas overlap run with query parameter showpointerareas overlap may be ok in some cases depending on the z ordering if the front most object is supposed to occlude pointer areas and whether objects can be moved internationalization are there any strings that are not internationalized and does the sim layout gracefully handle internationalized strings that are shorter than the english strings run with query parameter stringtest x you should see nothing but x strings does the sim layout gracefully handle internationalized strings that are longer than the english strings run with query parameters stringtest double and stringtest long does the sim stay on the sim page doesn t redirect to an external page when running with the query parameter stringtest xss this test passes if sim does not redirect ok if sim crashes or fails to fully start only test on one desktop platform for phet io sims additionally test stringtest xss in studio to make sure strings didn t leak to phetiodocumentation see avoid using concatenation to create strings that will be visible in the user interface use stringutils fillin and a string pattern to ensure that strings are properly localized use named placeholders e g value units instead of numbered placeholders e g make sure the string keys are all perfect because they are difficult to change after is published guidelines for string keys are strings keys should generally match their values e g js helloworld value hello world quadraticterms value quadratic terms if a string key would be exceptionally long use a key name that is an abbreviated form of the string value or that captures the purpose essence of the value e g js key is abbreviated iwenttothestore value i went to the store to get milk eggs butter and sugar key is based on purpose describethescreen value the play area is a small room the control panel has buttons a checkbox and radio buttons to change conditions in the room if string key names would collide use your judgment to disambiguate e g js simplifytitle value simplify simplifycheckbox value simplify string keys for screen names should have the general form screen screenname e g js screen explore value explore string patterns that contain placeholders e g my name is first last should use keys that are unlikely to conflict with strings that might be needed in the future for example for price consider using key pricepattern instead of price if you think there might be a future need for a price string it is acceptable to prefix families of strings with a prefix like so json material water value water material wood value wood shape block value block shape cone value cone nested substructure is not yet fully supported if the sim was already released make sure none of the original string keys have changed if they have changed make sure any changes have a good reason and have been discussed with jbphet it is likely that an issue like should be created repository structure are all required files and directories present for a sim repository named “my repo” the general structure should look like this where assets images mipmaps or sounds may be omitted if the sim doesn’t have those types of resource files my repo assets doc images see annotation model md implementation notes md images license json js see section below mipmaps license json sound license json dependencies json gitignore my repo en html my repo strings en json gruntfile js license package json readme md any images used in model md or implementation notes md should be added here images specific to aiding with documentation do not need their own license verify that the same image file is not present in both images and mipmaps if you need a mipmap use it for all occurrences of the image is the js directory properly structured all javascript source should be in the js directory there should be a subdirectory for each screen this also applies for single screen sims where the subdirectory matches the repo name for a multi screen sim code shared by or more screens should be in a js common subdirectory model and view code should be in model and view subdirectories for each screen and common for example for a sim with screens “introduction” and “lab” the general directory structure should look like this my repo js common model view introduction model view lab model view my repo main js myrepo js myrepostrings js do filenames use an appropriate prefix some filenames may be prefixed with the repository name e g molarityconstants js in molarity if the repository name is long the developer may choose to abbreviate the repository name e g eeconstants js in expression exchange if the abbreviation is already used by another respository then the full name must be used for example if the ee abbreviation is already used by expression exchange then it should not be used in equality explorer whichever convention is used it should be used consistently within a repository don t mix abbreviations and full names is there a file in assets for every resource file in sound and images note that there is not necessarily a correspondence between asset and resource files for example several related images may be in the same ai file check license json for possible documentation of why some reesources might not have a corresponding asset file for simulations was the readme md generated by grunt published readme or grunt unpublished readme common code repos can have custom readme files does package json refer to any dependencies that are not used by the sim is the license file correct generally gpl for sims and mit for common code see for additional information does gitignore match the one in simula rasa in github verify that all non release branches have an associated issue that describes their purpose are there any github branches that are no longer needed and should be deleted sim specific query parameters if any should be identified and documented in one js file in js common or js if there is no common the js file should be named prefix queryparameters js for example arithmeticqueryparameters js for the aritmetic repository or fbqueryparameters js for function builder where the fb prefix is used query parameters that are public facing should be identified using public true in the schema all sims should use a color file named myrepocolors js or if using abbreviations mrcolors js and use profilecolorproperty where appropriate even if they have a single default profile to support color editing and phet io studio the colorprofile pattern was converted to colors js files in please see for a good example coding conventions are coding conventions outlined in followed and adhered to this document deals with phet coding conventions you do not need to exhaustively check every item in this section nor do you necessarily need to check these items one at a time the goal is to determine whether the code generally meets phet standards typescript conventions are typescript conventions outlined in the followed and adhered to math libraries dot utils tofixed or dot utils tofixednumber should be used instead of tofixed javascript s tofixed is notoriously buggy behavior differs depending on browser because the spec doesn t specify whether to round or floor ie is no longer supported with that in mind remove ie specific workarounds use string includes and string startswith where possible organization readability and maintainability does the organization and structure of the code make sense do the model and view contain types that you would expect or guess by looking at the sim do the names of things correspond to the names that you see in the user interface are appropriate design patterns used see if new or inappropriate patterns are identified create an issue is inheritance used where appropriate does the type hierarchy make sense is composition favored over inheritance where appropriate see is there any unnecessary coupling e g by passing large objects to constructors or exposing unnecessary properties functions is there too much unnecessary decoupling e g by passing all of the properties of an object independently instead of passing the object itself are the source files reasonable in size scrutinize large files with too many responsibilities can responsibilities be broken into smaller delegates are any significant chunks of code duplicated in addition to manual identification tools include webstorm code analyze code locate duplicates and is there anything that should be generalized and migrated to common code are there any todo or fixme or review comments in the code they should be addressed or promoted to github issues are there any that should be factored out as constants and documented are there any constants that are duplicated in multiple files that should be factored out into a repo constants js file does the implementation rely on any specific constant values that are likely to change in the future identify constants that might be changed in the future use your judgement about which constants are likely candidates does changing the values of these constants break the sim for example see is used where appropriate verify that the sim is not inventing creating its own colors for things that have been standardized in phetcolorscheme identify any colors that might be worth adding to phetcolorscheme are all dependent properties modeled as derivedproperty instead of property all dynamics should be called from sim step dt do not use window settimeout or window setinterval this will help support legends of learning and phet io accessibility this section may be omitted if the sim has not been instrumented with accessibility features accessibility includes various features not all are always include ignore sections that do not apply general are accessibility features integrated well into the code they should be added in a maintainable way even if that requires upfront refactoring alternative input does the sim pass an accessibility fuzz test run with query parameters fuzzboard ea does this sim use specific keyboard shortcuts that overlap with global shortcuts this includes the use of modifier keys like in note there is currently no list of global shortcuts and therefore no way to complete this checklist item see interactive description run the entire built sim html file through an does the html pass if applicable are good design patterns used for interactive description see does resetting the simulation also reset the entire pdom is node pdomorder used appropriately to maintain visual and pdom layout balance make sure accessibility strings aren t being adjusted with ascii specific javascript methods like touppercase remember that one day these strings will be translatable make sure for accessibility strings that all end of sentence periods do not have a leading space before it some screen readers will read these as dot this can occur often when a clause is conditionally added phet io this section may be omitted if the sim has not been instrumented for phet io but is likely good to glance at no matter does instrumentation follow the conventions described in this could be an extensive bullet at the very least be sure to know what amount of instrumentation this sim supports describing this further goes beyond the scope of this document phet io instantiates different objects and wires up listeners that are not present in the phet branded simulation it needs to be tested separately for memory leaks to help isolate the nature of the memory leak this test should be run separately from the phet brand memory leak test test with a colorized data stream and studio easily accessed from phetmarks compare to testing results done by the responsible developer and previous releases make sure unused phetioobject instances are disposed which unregisters their tandems make sure joist dt values are used instead of date now or other date functions perhaps try phet joist elapsedtime though this has already been mentioned it is necessary for reproducible playback via input events and deserves a comment in this phet io section are random numbers using dot dotrandom as an imported module not a global and all doing so after modules are declared non statically for example the following methods and perhaps others should not be used math random shuffle sample random this also deserves re iteration due to its effect on record playback for phet io like json keys for undefined values are omitted when serializing objects across frames consider this when determining whether tostateobject should use null or undefined values phet prefers to use the term position to refer to the physical x y position of objects this applies to both brands but is more important for the phet io api see are your iotype state methods violating the api of the core type by accessing private fields when defining a boolean property to indicate whether something is enabled use axon enabledproperty this should be done in both the model and the view if you re using a derivedproperty skip this item do not use translated strings in phetiodocumentaton it changes the phet io api
0
256,881
22,108,858,981
IssuesEvent
2022-06-01 19:19:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: sequelize failed
C-test-failure O-robot O-roachtest branch-master T-sql-experience
roachtest.sequelize [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5336174?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5336174?buildTab=artifacts#/sequelize) on master @ [1cea73c8a18623949b81705eb5f75179e6cd8d86](https://github.com/cockroachdb/cockroach/commits/1cea73c8a18623949b81705eb5f75179e6cd8d86): ``` | initialize submodules in the clone | -j, --jobs <n> number of submodules cloned in parallel | --template <template-directory> | directory from which templates will be used | --reference <repo> reference repository | --reference-if-able <repo> | reference repository | --dissociate use --reference only while cloning | -o, --origin <name> use <name> instead of 'origin' to track upstream | -b, --branch <branch> | checkout <branch> instead of the remote's HEAD | -u, --upload-pack <path> | path to git-upload-pack on the remote | --depth <depth> create a shallow clone of that depth | --shallow-since <time> | create a shallow clone since a specific time | --shallow-exclude <revision> | deepen history of shallow clone, excluding rev | --single-branch clone only one branch, HEAD or --branch | --no-tags don't clone any tags, and make later fetches not to follow them | --shallow-submodules any cloned submodules will be shallow | --separate-git-dir <gitdir> | separate git dir from working tree | -c, --config <key=value> | set config inside the new repository | --server-option <server-specific> | option to transmit | -4, --ipv4 use IPv4 addresses only | -6, --ipv6 use IPv6 addresses only | --filter <args> object filtering | --remote-submodules any cloned submodules will use their remote-tracking branch | --sparse initialize sparse-checkout file to include only files at root | | | stdout: Wraps: (8) COMMAND_PROBLEM Wraps: (9) Node 1. Command with error: | `````` | bash -e -c ' | if ! test -d /mnt/data1/sequelize; then | git clone -b v6.0.5 --depth 1 https://github.com/cockroachdb/sequelize-cockroachdb.git /mnt/data1/sequelize --add safe.directory /mnt/data1/sequelize | else | cd /mnt/data1/sequelize | git fetch origin | git checkout origin/v6.0.5 | fi | ' | `````` Wraps: (10) exit status 129 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.withPrefix (7) *cluster.WithCommandDetails (8) errors.Cmd (9) *hintdetail.withDetail (10) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sequelize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16261
2.0
roachtest: sequelize failed - roachtest.sequelize [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5336174?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5336174?buildTab=artifacts#/sequelize) on master @ [1cea73c8a18623949b81705eb5f75179e6cd8d86](https://github.com/cockroachdb/cockroach/commits/1cea73c8a18623949b81705eb5f75179e6cd8d86): ``` | initialize submodules in the clone | -j, --jobs <n> number of submodules cloned in parallel | --template <template-directory> | directory from which templates will be used | --reference <repo> reference repository | --reference-if-able <repo> | reference repository | --dissociate use --reference only while cloning | -o, --origin <name> use <name> instead of 'origin' to track upstream | -b, --branch <branch> | checkout <branch> instead of the remote's HEAD | -u, --upload-pack <path> | path to git-upload-pack on the remote | --depth <depth> create a shallow clone of that depth | --shallow-since <time> | create a shallow clone since a specific time | --shallow-exclude <revision> | deepen history of shallow clone, excluding rev | --single-branch clone only one branch, HEAD or --branch | --no-tags don't clone any tags, and make later fetches not to follow them | --shallow-submodules any cloned submodules will be shallow | --separate-git-dir <gitdir> | separate git dir from working tree | -c, --config <key=value> | set config inside the new repository | --server-option <server-specific> | option to transmit | -4, --ipv4 use IPv4 addresses only | -6, --ipv6 use IPv6 addresses only | --filter <args> object filtering | --remote-submodules any cloned submodules will use their remote-tracking branch | --sparse initialize sparse-checkout file to include only files at root | | | stdout: Wraps: (8) COMMAND_PROBLEM Wraps: (9) Node 1. Command with error: | `````` | bash -e -c ' | if ! test -d /mnt/data1/sequelize; then | git clone -b v6.0.5 --depth 1 https://github.com/cockroachdb/sequelize-cockroachdb.git /mnt/data1/sequelize --add safe.directory /mnt/data1/sequelize | else | cd /mnt/data1/sequelize | git fetch origin | git checkout origin/v6.0.5 | fi | ' | `````` Wraps: (10) exit status 129 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.withPrefix (7) *cluster.WithCommandDetails (8) errors.Cmd (9) *hintdetail.withDetail (10) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sequelize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16261
non_process
roachtest sequelize failed roachtest sequelize with on master initialize submodules in the clone j jobs number of submodules cloned in parallel template directory from which templates will be used reference reference repository reference if able reference repository dissociate use reference only while cloning o origin use instead of origin to track upstream b branch checkout instead of the remote s head u upload pack path to git upload pack on the remote depth create a shallow clone of that depth shallow since create a shallow clone since a specific time shallow exclude deepen history of shallow clone excluding rev single branch clone only one branch head or branch no tags don t clone any tags and make later fetches not to follow them shallow submodules any cloned submodules will be shallow separate git dir separate git dir from working tree c config set config inside the new repository server option option to transmit use addresses only use addresses only filter object filtering remote submodules any cloned submodules will use their remote tracking branch sparse initialize sparse checkout file to include only files at root stdout wraps command problem wraps node command with error bash e c if test d mnt sequelize then git clone b depth mnt sequelize add safe directory mnt sequelize else cd mnt sequelize git fetch origin git checkout origin fi wraps exit status error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror help see see cc cockroachdb sql experience jira issue crdb
0
19,129
25,183,580,614
IssuesEvent
2022-11-11 15:47:36
googleapis/python-bigquery-sqlalchemy
https://api.github.com/repos/googleapis/python-bigquery-sqlalchemy
closed
specify version bounds for dependencies in "extras"
type: process priority: p3 api: bigquery
Currently, the extras don't have any version bounds for GeoAlchemy2, shapely, or alembic. This makes me a bit nervous, as it's possible they release a breaking change that means our users have to manually pin those version. (Because of how extras work, they likely need to pin them anyway, but at least if they install the extras from our package, pip could give hints as to what is expected to work)
1.0
specify version bounds for dependencies in "extras" - Currently, the extras don't have any version bounds for GeoAlchemy2, shapely, or alembic. This makes me a bit nervous, as it's possible they release a breaking change that means our users have to manually pin those version. (Because of how extras work, they likely need to pin them anyway, but at least if they install the extras from our package, pip could give hints as to what is expected to work)
process
specify version bounds for dependencies in extras currently the extras don t have any version bounds for shapely or alembic this makes me a bit nervous as it s possible they release a breaking change that means our users have to manually pin those version because of how extras work they likely need to pin them anyway but at least if they install the extras from our package pip could give hints as to what is expected to work
1
758
8,050,172,229
IssuesEvent
2018-08-01 12:40:53
cheerup-ehime/cheerup-ehime.github.io
https://api.github.com/repos/cheerup-ehime/cheerup-ehime.github.io
opened
被災地支援者としてふるさと納税の義援金の自治体毎の寄付金額合計をビジュアルに見たい
Automation Useful For Supporters Visualization
## 誰のために? 被災地を義援金で支援したい人のために ## 何を実現したい? ふるさと納税の義援金を自治体毎にいくらになっているかを取得して記事化・グラフ化したい。 そのために、ふるさと納税のサイトをスクレイピングして定時取得し合計し、グラフ化する。 ## それはなぜ? 金額だけだとピンとこないのと、代理寄付もあり、ある自治体の合計金額が見えない。 被害規模に対して、どれくらい集まっているのか、そうでないのかが見えない。 ## どうなったら完了? - ふるさと納税サイトからデータが収集できCSVに出力できている - CSVを読み込み、自治体ごとの小計ができている - データの日次毎の金額の推移を記録してCSVに出力できている - 自治体ごとに時系列毎のグラフ出力(Y軸金額、X軸日時)ができている - 自治体ごとに義援金の合計がグラフ化できている
1.0
被災地支援者としてふるさと納税の義援金の自治体毎の寄付金額合計をビジュアルに見たい - ## 誰のために? 被災地を義援金で支援したい人のために ## 何を実現したい? ふるさと納税の義援金を自治体毎にいくらになっているかを取得して記事化・グラフ化したい。 そのために、ふるさと納税のサイトをスクレイピングして定時取得し合計し、グラフ化する。 ## それはなぜ? 金額だけだとピンとこないのと、代理寄付もあり、ある自治体の合計金額が見えない。 被害規模に対して、どれくらい集まっているのか、そうでないのかが見えない。 ## どうなったら完了? - ふるさと納税サイトからデータが収集できCSVに出力できている - CSVを読み込み、自治体ごとの小計ができている - データの日次毎の金額の推移を記録してCSVに出力できている - 自治体ごとに時系列毎のグラフ出力(Y軸金額、X軸日時)ができている - 自治体ごとに義援金の合計がグラフ化できている
non_process
被災地支援者としてふるさと納税の義援金の自治体毎の寄付金額合計をビジュアルに見たい 誰のために? 被災地を義援金で支援したい人のために 何を実現したい? ふるさと納税の義援金を自治体毎にいくらになっているかを取得して記事化・グラフ化したい。 そのために、ふるさと納税のサイトをスクレイピングして定時取得し合計し、グラフ化する。 それはなぜ? 金額だけだとピンとこないのと、代理寄付もあり、ある自治体の合計金額が見えない。 被害規模に対して、どれくらい集まっているのか、そうでないのかが見えない。 どうなったら完了? ふるさと納税サイトからデータが収集できcsvに出力できている csvを読み込み、自治体ごとの小計ができている データの日次毎の金額の推移を記録してcsvに出力できている 自治体ごとに時系列毎のグラフ出力 y軸金額、x軸日時)ができている 自治体ごとに義援金の合計がグラフ化できている
0
4,030
6,963,968,402
IssuesEvent
2017-12-08 19:36:07
nodejs/node
https://api.github.com/repos/nodejs/node
closed
test: investigate - async-hooks/test-graph.signal
async_hooks CI / flaky test process
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: `master` * **Platform**: linux * **Subsystem**: async_hooks <!-- Enter your issue details below this comment. --> https://ci.nodejs.org/job/node-test-commit-linux/11523/nodes=centos5-64 ``` not ok 26 async-hooks/test-graph.signal --- duration_ms: 0.155 severity: fail stack: |- Mismatched onsigusr2 function calls. Expected exactly 2, actual 0. at Object.exports.mustCall (/home/iojs/build/workspace/node-test-commit-linux/nodes/centos5-64/test/common/index.js:475:10) at Object.<anonymous> (/home/iojs/build/workspace/node-test-commit-linux/nodes/centos5-64/test/async-hooks/test-graph.signal.js:16:30) at Module._compile (module.js:573:30) at Object.Module._extensions..js (module.js:584:10) at Module.load (module.js:507:32) at tryModuleLoad (module.js:470:12) at Function.Module._load (module.js:462:3) at Function.Module.runMain (module.js:609:10) at startup (bootstrap_node.js:158:16) ... ``` ### [code](https://github.com/nodejs/node/blob/master/test/async-hooks/test-graph.signal.js#L16) ```js ... hooks.enable(); process.on('SIGUSR2', common.mustCall(onsigusr2, 2)); let count = 0; exec(`kill -USR2 ${process.pid}`); ... ``` /cc @nodejs/async_hooks @Fishrock123 @thlorenz
1.0
test: investigate - async-hooks/test-graph.signal - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: `master` * **Platform**: linux * **Subsystem**: async_hooks <!-- Enter your issue details below this comment. --> https://ci.nodejs.org/job/node-test-commit-linux/11523/nodes=centos5-64 ``` not ok 26 async-hooks/test-graph.signal --- duration_ms: 0.155 severity: fail stack: |- Mismatched onsigusr2 function calls. Expected exactly 2, actual 0. at Object.exports.mustCall (/home/iojs/build/workspace/node-test-commit-linux/nodes/centos5-64/test/common/index.js:475:10) at Object.<anonymous> (/home/iojs/build/workspace/node-test-commit-linux/nodes/centos5-64/test/async-hooks/test-graph.signal.js:16:30) at Module._compile (module.js:573:30) at Object.Module._extensions..js (module.js:584:10) at Module.load (module.js:507:32) at tryModuleLoad (module.js:470:12) at Function.Module._load (module.js:462:3) at Function.Module.runMain (module.js:609:10) at startup (bootstrap_node.js:158:16) ... ``` ### [code](https://github.com/nodejs/node/blob/master/test/async-hooks/test-graph.signal.js#L16) ```js ... hooks.enable(); process.on('SIGUSR2', common.mustCall(onsigusr2, 2)); let count = 0; exec(`kill -USR2 ${process.pid}`); ... ``` /cc @nodejs/async_hooks @Fishrock123 @thlorenz
process
test investigate async hooks test graph signal thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version master platform linux subsystem async hooks not ok async hooks test graph signal duration ms severity fail stack mismatched function calls expected exactly actual at object exports mustcall home iojs build workspace node test commit linux nodes test common index js at object home iojs build workspace node test commit linux nodes test async hooks test graph signal js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js at function module load module js at function module runmain module js at startup bootstrap node js js hooks enable process on common mustcall let count exec kill process pid cc nodejs async hooks thlorenz
1
619,967
19,541,148,837
IssuesEvent
2021-12-31 23:50:46
obsproject/obs-websocket
https://api.github.com/repos/obsproject/obs-websocket
closed
Bug: Events don't fire reliably
Medium Priority Issue: Bug - Unconfirmed Needs Research
##### Issue type - Bug report ##### Description Some of the events don't fire reliably. - PreviewSceneChanged sometimes only fires the first time it is triggered. Every subsequent selection does not fire the event. Or it doesn't fire at all. - SceneCollectionChanged doesn't fire at all when changing the scene collection ##### Steps to reproduce and other useful info For me, this reliably happens as soon as I call GetSceneList. After this I don't get any events on the connection until I reestablish a new WebSocket connection. ##### Technical information - **Operating System** : Win 10 Build 19042 - **OBS Studio version**: 27.0.1 - **obs-websocket version** 4.91. ##### Development Environment Not sure if the error is because of the API I use or in the plugin itself. All events are consumed using the .NET implementation [obs-websocket-dotnet](https://github.com/BarRaider/obs-websocket-dotnet)
1.0
Bug: Events don't fire reliably - ##### Issue type - Bug report ##### Description Some of the events don't fire reliably. - PreviewSceneChanged sometimes only fires the first time it is triggered. Every subsequent selection does not fire the event. Or it doesn't fire at all. - SceneCollectionChanged doesn't fire at all when changing the scene collection ##### Steps to reproduce and other useful info For me, this reliably happens as soon as I call GetSceneList. After this I don't get any events on the connection until I reestablish a new WebSocket connection. ##### Technical information - **Operating System** : Win 10 Build 19042 - **OBS Studio version**: 27.0.1 - **obs-websocket version** 4.91. ##### Development Environment Not sure if the error is because of the API I use or in the plugin itself. All events are consumed using the .NET implementation [obs-websocket-dotnet](https://github.com/BarRaider/obs-websocket-dotnet)
non_process
bug events don t fire reliably issue type bug report description some of the events don t fire reliably previewscenechanged sometimes only fires the first time it is triggered every subsequent selection does not fire the event or it doesn t fire at all scenecollectionchanged doesn t fire at all when changing the scene collection steps to reproduce and other useful info for me this reliably happens as soon as i call getscenelist after this i don t get any events on the connection until i reestablish a new websocket connection technical information operating system win build obs studio version obs websocket version development environment not sure if the error is because of the api i use or in the plugin itself all events are consumed using the net implementation
0
182,287
21,664,496,832
IssuesEvent
2022-05-07 01:33:04
emilwareus/NodeGoat
https://api.github.com/repos/emilwareus/NodeGoat
closed
CVE-2020-7774 (High) detected in y18n-3.2.1.tgz - autoclosed
security vulnerability
## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-3.2.1.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz">https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - yargs-4.8.1.tgz - :x: **y18n-3.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/emilwareus/NodeGoat/commit/8eddc2830777251e3d86693a089db39bd3090b5d">8eddc2830777251e3d86693a089db39bd3090b5d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7774 (High) detected in y18n-3.2.1.tgz - autoclosed - ## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-3.2.1.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz">https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - yargs-4.8.1.tgz - :x: **y18n-3.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/emilwareus/NodeGoat/commit/8eddc2830777251e3d86693a089db39bd3090b5d">8eddc2830777251e3d86693a089db39bd3090b5d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tgz autoclosed cve high severity vulnerability vulnerable library tgz the bare bones internationalization library used by yargs library home page a href path to dependency file package json path to vulnerable library node modules nyc node modules package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz yargs tgz x tgz vulnerable library found in head commit a href vulnerability details this affects the package before and poc by const require setlocale proto updatelocale polluted true console log polluted true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,655
26,011,478,208
IssuesEvent
2022-12-21 02:28:00
NCAR/comp-pipeline
https://api.github.com/repos/NCAR/comp-pipeline
opened
First pass at reprocessing pre-December 2012
level 1 process
Need to get the raw data from the archive, reprocess, and do some analysis on the results to see if they are consistent with the rest of the mission. The sub-tasks are: - [ ] download pre-December 2012 from Campaign Storage - [ ] reprocess level 1 and 2 - [ ] update flat plot - [ ] update velocity data file, i.e., run `comp_quick_invert_restwvl`
1.0
First pass at reprocessing pre-December 2012 - Need to get the raw data from the archive, reprocess, and do some analysis on the results to see if they are consistent with the rest of the mission. The sub-tasks are: - [ ] download pre-December 2012 from Campaign Storage - [ ] reprocess level 1 and 2 - [ ] update flat plot - [ ] update velocity data file, i.e., run `comp_quick_invert_restwvl`
process
first pass at reprocessing pre december need to get the raw data from the archive reprocess and do some analysis on the results to see if they are consistent with the rest of the mission the sub tasks are download pre december from campaign storage reprocess level and update flat plot update velocity data file i e run comp quick invert restwvl
1
8,121
11,303,181,405
IssuesEvent
2020-01-17 19:28:27
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
OCM Scrum Way of Working - iteration 2
OCM_Int Processes
- [x] Developed description of each pipeline & sent for review ... - [x] Katherine review - [ ] KM review
1.0
OCM Scrum Way of Working - iteration 2 - - [x] Developed description of each pipeline & sent for review ... - [x] Katherine review - [ ] KM review
process
ocm scrum way of working iteration developed description of each pipeline sent for review katherine review km review
1
34,591
9,417,825,578
IssuesEvent
2019-04-10 17:41:29
alter-rebbe/site
https://api.github.com/repos/alter-rebbe/site
opened
Add support for translations
build content
- [ ] detect secondary documents (xxx-ru.xml); - [ ] fold them into Document; - [ ] generate references to secondary documents in the index entries; - [ ] generate references to translations in the document viewer; - [ ] test on 059.
1.0
Add support for translations - - [ ] detect secondary documents (xxx-ru.xml); - [ ] fold them into Document; - [ ] generate references to secondary documents in the index entries; - [ ] generate references to translations in the document viewer; - [ ] test on 059.
non_process
add support for translations detect secondary documents xxx ru xml fold them into document generate references to secondary documents in the index entries generate references to translations in the document viewer test on
0
361,207
10,705,713,791
IssuesEvent
2019-10-24 14:11:20
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Marketplace bp sites pages are displayed incorrectly when different sites exists, depending on which site was created first
bug priority: high
## Describe the bug Marketplace bp sites pages are displayed incorrectly when different sites exists that uses say css files of the same name, depending on which site was created first ## To Reproduce Steps to reproduce the behavior: 1. Create a site using the `Art Showcase` bp from the marketplace. Below is how the site looks like after creation: ![Screen Shot 2019-10-16 at 11 56 27 AM](https://user-images.githubusercontent.com/25483966/66937828-2146b080-f00e-11e9-921d-72048fed9d56.png) 2. Create a site using the hotel bp from the marketplace. 3. Click on the `Main Menu` and preview the `Art Showcase` site created from step 1. Notice how the site looks now: ![Screen Shot 2019-10-16 at 11 53 08 AM](https://user-images.githubusercontent.com/25483966/66937960-54893f80-f00e-11e9-833f-f0f85fddfaa5.png) ## Expected behavior Previewing art showcase or hotel sites, etc should not be affected on which site was created first ## Screenshots Here's a short clip: https://www.loom.com/share/f4d909d7549640b196e62aa7cb143c39 ## Logs {{If applicable, attach the logs/stack trace (use https://gist.github.com).}} ## Specs ### Version Studio Version Number: 3.1.4-SNAPSHOT-31df36 Build Number: 31df362f307966926633b458806695069f844f77 Build Date/Time: 10-16-2019 10:01:14 -0400 ### OS OS X ### Browser Chrome browser ## Additional context This is similar to this ticket: https://github.com/craftercms/craftercms/issues/3010
1.0
[studio] Marketplace bp sites pages are displayed incorrectly when different sites exists, depending on which site was created first - ## Describe the bug Marketplace bp sites pages are displayed incorrectly when different sites exists that uses say css files of the same name, depending on which site was created first ## To Reproduce Steps to reproduce the behavior: 1. Create a site using the `Art Showcase` bp from the marketplace. Below is how the site looks like after creation: ![Screen Shot 2019-10-16 at 11 56 27 AM](https://user-images.githubusercontent.com/25483966/66937828-2146b080-f00e-11e9-921d-72048fed9d56.png) 2. Create a site using the hotel bp from the marketplace. 3. Click on the `Main Menu` and preview the `Art Showcase` site created from step 1. Notice how the site looks now: ![Screen Shot 2019-10-16 at 11 53 08 AM](https://user-images.githubusercontent.com/25483966/66937960-54893f80-f00e-11e9-833f-f0f85fddfaa5.png) ## Expected behavior Previewing art showcase or hotel sites, etc should not be affected on which site was created first ## Screenshots Here's a short clip: https://www.loom.com/share/f4d909d7549640b196e62aa7cb143c39 ## Logs {{If applicable, attach the logs/stack trace (use https://gist.github.com).}} ## Specs ### Version Studio Version Number: 3.1.4-SNAPSHOT-31df36 Build Number: 31df362f307966926633b458806695069f844f77 Build Date/Time: 10-16-2019 10:01:14 -0400 ### OS OS X ### Browser Chrome browser ## Additional context This is similar to this ticket: https://github.com/craftercms/craftercms/issues/3010
non_process
marketplace bp sites pages are displayed incorrectly when different sites exists depending on which site was created first describe the bug marketplace bp sites pages are displayed incorrectly when different sites exists that uses say css files of the same name depending on which site was created first to reproduce steps to reproduce the behavior create a site using the art showcase bp from the marketplace below is how the site looks like after creation create a site using the hotel bp from the marketplace click on the main menu and preview the art showcase site created from step notice how the site looks now expected behavior previewing art showcase or hotel sites etc should not be affected on which site was created first screenshots here s a short clip logs if applicable attach the logs stack trace use specs version studio version number snapshot build number build date time os os x browser chrome browser additional context this is similar to this ticket
0
20,604
27,268,054,094
IssuesEvent
2023-02-22 19:45:09
StormSurgeLive/asgs
https://api.github.com/repos/StormSurgeLive/asgs
opened
clean up logging so that syslog is not cluttered with low impact messages
enhancement important non-critical opendap reliability monitoring postprocessing
The writing of messages to the ASGS log file has become excessive to the point that the signal-to-noise ratio has become very low. In particular, the log messages produced by copying files to an external server (usually via `output/opendap_post2.sh` and especially for frequently updated status messages) are burdensome. The low-impact messages are still important, but should be diverted to a different log file.
1.0
clean up logging so that syslog is not cluttered with low impact messages - The writing of messages to the ASGS log file has become excessive to the point that the signal-to-noise ratio has become very low. In particular, the log messages produced by copying files to an external server (usually via `output/opendap_post2.sh` and especially for frequently updated status messages) are burdensome. The low-impact messages are still important, but should be diverted to a different log file.
process
clean up logging so that syslog is not cluttered with low impact messages the writing of messages to the asgs log file has become excessive to the point that the signal to noise ratio has become very low in particular the log messages produced by copying files to an external server usually via output opendap sh and especially for frequently updated status messages are burdensome the low impact messages are still important but should be diverted to a different log file
1
13,283
15,762,021,208
IssuesEvent
2021-03-31 10:36:43
KratosMultiphysics/Kratos
https://api.github.com/repos/KratosMultiphysics/Kratos
closed
General Discussion: Output-Process
Applications Discussion Kratos Core Post Process
I am opening this general discussion abt the output-processes (e.g. for GiD). Right now we have multiple ways for writing output, e.g. for gid, paraview. Those are implemented as `Process` but also have an extended interface: `IsOutputStep` and `PrintOutput` This (I assume) is because the base-process has the methods `ExecuteBeforeOutputStep` and `ExecuteAfterOutputStep` => the non-standard functions are then called inbetween those functions. On the other hand there are also other processes that print their output in `ExecuteFinalizeSolutionStep` The `AnalysisStage` deals with both, see [here](https://github.com/KratosMultiphysics/Kratos/blob/master/kratos/python_scripts/analysis_stage.py#L137-L157) I think it would be nice to unify this situation, which would also make it easier to write a wrapper for switching the output without effort. Plus the current situation is confusing for new users I think. - Either we print the output in `ExecuteFinalizeSolutionStep`, but then the functions `ExecuteBeforeOutputStep` and `ExecuteAfterOutputStep` are not really needed/used. Are they used somewhere? - Or we print the Output in `PrintOutput`, then it would probably make sense to have a new baseclass (`OutputProcess`) which derives from `Process`. In there we could also implement `IsOutputStep` in a general way I don't have a strong opinion on either of those, so what do others think? @KratosMultiphysics/technical-committee @KratosMultiphysics/interface-committee @KratosMultiphysics/fluid-dynamics @KratosMultiphysics/structural-mechanics
1.0
General Discussion: Output-Process - I am opening this general discussion abt the output-processes (e.g. for GiD). Right now we have multiple ways for writing output, e.g. for gid, paraview. Those are implemented as `Process` but also have an extended interface: `IsOutputStep` and `PrintOutput` This (I assume) is because the base-process has the methods `ExecuteBeforeOutputStep` and `ExecuteAfterOutputStep` => the non-standard functions are then called inbetween those functions. On the other hand there are also other processes that print their output in `ExecuteFinalizeSolutionStep` The `AnalysisStage` deals with both, see [here](https://github.com/KratosMultiphysics/Kratos/blob/master/kratos/python_scripts/analysis_stage.py#L137-L157) I think it would be nice to unify this situation, which would also make it easier to write a wrapper for switching the output without effort. Plus the current situation is confusing for new users I think. - Either we print the output in `ExecuteFinalizeSolutionStep`, but then the functions `ExecuteBeforeOutputStep` and `ExecuteAfterOutputStep` are not really needed/used. Are they used somewhere? - Or we print the Output in `PrintOutput`, then it would probably make sense to have a new baseclass (`OutputProcess`) which derives from `Process`. In there we could also implement `IsOutputStep` in a general way I don't have a strong opinion on either of those, so what do others think? @KratosMultiphysics/technical-committee @KratosMultiphysics/interface-committee @KratosMultiphysics/fluid-dynamics @KratosMultiphysics/structural-mechanics
process
general discussion output process i am opening this general discussion abt the output processes e g for gid right now we have multiple ways for writing output e g for gid paraview those are implemented as process but also have an extended interface isoutputstep and printoutput this i assume is because the base process has the methods executebeforeoutputstep and executeafteroutputstep the non standard functions are then called inbetween those functions on the other hand there are also other processes that print their output in executefinalizesolutionstep the analysisstage deals with both see i think it would be nice to unify this situation which would also make it easier to write a wrapper for switching the output without effort plus the current situation is confusing for new users i think either we print the output in executefinalizesolutionstep but then the functions executebeforeoutputstep and executeafteroutputstep are not really needed used are they used somewhere or we print the output in printoutput then it would probably make sense to have a new baseclass outputprocess which derives from process in there we could also implement isoutputstep in a general way i don t have a strong opinion on either of those so what do others think kratosmultiphysics technical committee kratosmultiphysics interface committee kratosmultiphysics fluid dynamics kratosmultiphysics structural mechanics
1
7,533
10,608,385,832
IssuesEvent
2019-10-11 07:24:18
axa-group/Parsr
https://api.github.com/repos/axa-group/Parsr
closed
Duplicates not removed because of slight translation in coordinates
bug processing
If there are duplicates of certain textual elements with not the exact coordinates but slightly translated from one another, the duplicate element removal module does not treat them properly. <img width="678" alt="Screen Shot 2019-08-19 at 3 26 32 pm" src="https://user-images.githubusercontent.com/1119728/63241195-197cc100-c296-11e9-8bc4-f19593d08057.png"> UPDATE: After e5cba27 , <img width="481" alt="Screen Shot 2019-10-03 at 4 01 06 pm" src="https://user-images.githubusercontent.com/1119728/66102827-08460480-e5f7-11e9-8b76-de473ff62faf.png">
1.0
Duplicates not removed because of slight translation in coordinates - If there are duplicates of certain textual elements with not the exact coordinates but slightly translated from one another, the duplicate element removal module does not treat them properly. <img width="678" alt="Screen Shot 2019-08-19 at 3 26 32 pm" src="https://user-images.githubusercontent.com/1119728/63241195-197cc100-c296-11e9-8bc4-f19593d08057.png"> UPDATE: After e5cba27 , <img width="481" alt="Screen Shot 2019-10-03 at 4 01 06 pm" src="https://user-images.githubusercontent.com/1119728/66102827-08460480-e5f7-11e9-8b76-de473ff62faf.png">
process
duplicates not removed because of slight translation in coordinates if there are duplicates of certain textual elements with not the exact coordinates but slightly translated from one another the duplicate element removal module does not treat them properly img width alt screen shot at pm src update after img width alt screen shot at pm src
1