Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
12,000
14,738,129,531
IssuesEvent
2021-01-07 03:50:37
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
068- portland - Payment issue
anc-ops anc-process anp-0.5 ant-bug ant-support
In GitLab by @kdjstudios on May 8, 2018, 12:13 **Submitted by:** "Lettice Ross" <lettice.ross@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-23481/conversation **Server:** Internal **Client/Site:** 068 **Account:** Lucas Emergy B00671 **Issue:** I’m sending this email, because I have received an error message when trying to process credit card payment. The account I’m trying to process payment for is Lucas Emergy B00671. The error message read: We’re sorry, but something went wrong. If you can please help, and let me know once finish.
1.0
068- portland - Payment issue - In GitLab by @kdjstudios on May 8, 2018, 12:13 **Submitted by:** "Lettice Ross" <lettice.ross@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-23481/conversation **Server:** Internal **Client/Site:** 068 **Account:** Lucas Emergy B00671 **Issue:** I’m sending this email, because I have received an error message when trying to process credit card payment. The account I’m trying to process payment for is Lucas Emergy B00671. The error message read: We’re sorry, but something went wrong. If you can please help, and let me know once finish.
process
portland payment issue in gitlab by kdjstudios on may submitted by lettice ross helpdesk server internal client site account lucas emergy issue i’m sending this email because i have received an error message when trying to process credit card payment the account i’m trying to process payment for is lucas emergy the error message read we’re sorry but something went wrong if you can please help and let me know once finish
1
54,869
7,928,548,666
IssuesEvent
2018-07-06 12:09:02
txtdirect/txtdirect
https://api.github.com/repos/txtdirect/txtdirect
opened
Document that arbitrary key/value pairs are allowed
documentation
<!-- This form is for bug reports and feature requests ONLY! If you're looking for help check out [our support guidelines](/SUPPORT.md). --> **Is this a BUG REPORT or FEATURE REQUEST?**: feature Arbitrary key/value pairs should just be discarded (this allows to upgrade TXT records in case a newer version of TXTDirect uses new keys, without breaking older versions) Arbitrary data, however, shouldn't be allowed. Summary: ``` v=txtv0;to=https://txtdirect.org;code=302 -> works v=txtv0;https://txtdirect.org;code=302 -> works v=txtv0;to=https://txtdirect.org;code=302;tracking=true -> works v=txtv0;https://txtdirect.org;302 -> fails v=txtv0;to=https://txtdirect.org;302 -> fails ```
1.0
Document that arbitrary key/value pairs are allowed - <!-- This form is for bug reports and feature requests ONLY! If you're looking for help check out [our support guidelines](/SUPPORT.md). --> **Is this a BUG REPORT or FEATURE REQUEST?**: feature Arbitrary key/value pairs should just be discarded (this allows to upgrade TXT records in case a newer version of TXTDirect uses new keys, without breaking older versions) Arbitrary data, however, shouldn't be allowed. Summary: ``` v=txtv0;to=https://txtdirect.org;code=302 -> works v=txtv0;https://txtdirect.org;code=302 -> works v=txtv0;to=https://txtdirect.org;code=302;tracking=true -> works v=txtv0;https://txtdirect.org;302 -> fails v=txtv0;to=https://txtdirect.org;302 -> fails ```
non_process
document that arbitrary key value pairs are allowed this form is for bug reports and feature requests only if you re looking for help check out support md is this a bug report or feature request feature arbitrary key value pairs should just be discarded this allows to upgrade txt records in case a newer version of txtdirect uses new keys without breaking older versions arbitrary data however shouldn t be allowed summary v to works v works v to works v fails v to fails
0
677,122
23,151,777,742
IssuesEvent
2022-07-29 09:01:55
wasmerio/wasmer
https://api.github.com/repos/wasmerio/wasmer
closed
Re-add create exe to Wasmer
priority-high create-exe
This is a follow up from #2916 (it's blocked by it). We need to re-add `create-exe` to Wasmer. Steps: 1. Re-use the previous existing code (from C) to create the executable 2. Create a new `create-exe` that uses Zig under the hood. 3. Make zig on-create exe used with wax 5. Include the virtual filesystem into the generated executable (`python` should work) - not required for Wasmer 3.0 ## Details for (2) We need to Zig version of the C generated file, if zig is found in the system, we can use it by default, if not, we need to call a subprocess with `wax zig`. Unknowns: 1. linker might not work the same way ## Details for (3) We will include the virtual filesystem structure into Wasmer, so Python can be actually converted to a native executable
1.0
Re-add create exe to Wasmer - This is a follow up from #2916 (it's blocked by it). We need to re-add `create-exe` to Wasmer. Steps: 1. Re-use the previous existing code (from C) to create the executable 2. Create a new `create-exe` that uses Zig under the hood. 3. Make zig on-create exe used with wax 5. Include the virtual filesystem into the generated executable (`python` should work) - not required for Wasmer 3.0 ## Details for (2) We need to Zig version of the C generated file, if zig is found in the system, we can use it by default, if not, we need to call a subprocess with `wax zig`. Unknowns: 1. linker might not work the same way ## Details for (3) We will include the virtual filesystem structure into Wasmer, so Python can be actually converted to a native executable
non_process
re add create exe to wasmer this is a follow up from it s blocked by it we need to re add create exe to wasmer steps re use the previous existing code from c to create the executable create a new create exe that uses zig under the hood make zig on create exe used with wax include the virtual filesystem into the generated executable python should work not required for wasmer details for we need to zig version of the c generated file if zig is found in the system we can use it by default if not we need to call a subprocess with wax zig unknowns linker might not work the same way details for we will include the virtual filesystem structure into wasmer so python can be actually converted to a native executable
0
22,276
30,826,946,846
IssuesEvent
2023-08-01 20:54:15
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
storage: determine what to do on chunksize=0 [GRPC]
api: storage priority: p2 type: process
Address TODOs around L1472 in grpc_client.go Should we actually use the minimum of 256 KB here [chunksize=0] when the user indicates they want minimal memory usage? We cannot do a zero-copy, bufferless upload like HTTP/JSON can. Must make a note to update documentation accordingly.
1.0
storage: determine what to do on chunksize=0 [GRPC] - Address TODOs around L1472 in grpc_client.go Should we actually use the minimum of 256 KB here [chunksize=0] when the user indicates they want minimal memory usage? We cannot do a zero-copy, bufferless upload like HTTP/JSON can. Must make a note to update documentation accordingly.
process
storage determine what to do on chunksize address todos around in grpc client go should we actually use the minimum of kb here when the user indicates they want minimal memory usage we cannot do a zero copy bufferless upload like http json can must make a note to update documentation accordingly
1
3,302
6,399,092,377
IssuesEvent
2017-08-04 22:42:53
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.ServiceProcess.ServiceController & ServiceBase
area-System.ServiceProcess enhancement port-to-core
We've had requests to port https://github.com/topshelf/topshelf to .NET Core. To do this, we would love to be able to query if we had access to the SCM and act as needed so we don't need to ship a version targeting both core & full profile.
1.0
System.ServiceProcess.ServiceController & ServiceBase - We've had requests to port https://github.com/topshelf/topshelf to .NET Core. To do this, we would love to be able to query if we had access to the SCM and act as needed so we don't need to ship a version targeting both core & full profile.
process
system serviceprocess servicecontroller servicebase we ve had requests to port to net core to do this we would love to be able to query if we had access to the scm and act as needed so we don t need to ship a version targeting both core full profile
1
751,657
26,252,677,466
IssuesEvent
2023-01-05 20:52:28
awesomemotive/affiliatewp-affiliate-area-shortcodes
https://api.github.com/repos/awesomemotive/affiliatewp-affiliate-area-shortcodes
closed
If AffiliateWP is not active, there is no longer a requirement admin notice
type-bug priority-high workflow-has-pr
While this addon exhibits the issue, it's likely also affecting any other addon using the same activation class. I'm opening it here because I need to do some work on this addon for the WordPress.org repo and this is going to hold us back. Let's fix it in this addon first and then push it out to other addons. ## The issue: In the following screenshot I have just activated the Affiliate Area Shortcodes addon. Note how AffiliateWP is not activated and there is no notice that tells the user that AffiliateWP must be activated. ![image](https://user-images.githubusercontent.com/52581/210312253-a764de8f-ad98-4119-a570-62311b1b6161.png) We used to show a message like this: ![image](https://user-images.githubusercontent.com/52581/210311824-f13bd18a-bff9-43e3-b244-9d20659a88cb.png) In the case of this addon, the notice worked for all versions up until `1.2` when this was added: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/issues/37 Aside from the notice not displaying there are other issues: 1. There are two `AffiliateWP_Activation` classes included. See https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/lib/affwp/class-affiliatewp-activation.php and https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/class-activation.php 2. The activation code is super convoluted. No idea why it needs to have that much code. 3. There are incorrect textdomains, likely a copy/paste error when implementing on another addon. See `affiliatewp-afgf` in the codebase This is probably only a small part of this but this conditional is preventing the banner from ever loading: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/affiliatewp-affiliate-area-shortcodes.php#L108-L112 The first conditional with `version_compare( $affwp_version, '2.7', '>=' )` is `true` for me since I have a version higher than `2.7`. Then it tries to load the `bootstrap` method but can't because it runs on the `affwp_plugins_loaded` hook. That hook is only available within AffiliateWP, which of course is not active. To test and at least see the banner in the above screenshot, change `affwp_plugins_loaded` to `plugins_loaded`. So what we need to do here is: 1. Get the notice working again. It should show when AffiliateWP is not active 2. Preserve any changes added which cater to the minimum requirements (you'll see these browsing through the code) 3. Simplify the activation code, remove the duplicate class, fix the text domains After we have thoroughly tested this we need to roll this out to the other addons. It's a super broken experience when a user installs our free addons and there's zero indication of what to do next.
1.0
If AffiliateWP is not active, there is no longer a requirement admin notice - While this addon exhibits the issue, it's likely also affecting any other addon using the same activation class. I'm opening it here because I need to do some work on this addon for the WordPress.org repo and this is going to hold us back. Let's fix it in this addon first and then push it out to other addons. ## The issue: In the following screenshot I have just activated the Affiliate Area Shortcodes addon. Note how AffiliateWP is not activated and there is no notice that tells the user that AffiliateWP must be activated. ![image](https://user-images.githubusercontent.com/52581/210312253-a764de8f-ad98-4119-a570-62311b1b6161.png) We used to show a message like this: ![image](https://user-images.githubusercontent.com/52581/210311824-f13bd18a-bff9-43e3-b244-9d20659a88cb.png) In the case of this addon, the notice worked for all versions up until `1.2` when this was added: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/issues/37 Aside from the notice not displaying there are other issues: 1. There are two `AffiliateWP_Activation` classes included. See https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/lib/affwp/class-affiliatewp-activation.php and https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/class-activation.php 2. The activation code is super convoluted. No idea why it needs to have that much code. 3. There are incorrect textdomains, likely a copy/paste error when implementing on another addon. See `affiliatewp-afgf` in the codebase This is probably only a small part of this but this conditional is preventing the banner from ever loading: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/affiliatewp-affiliate-area-shortcodes.php#L108-L112 The first conditional with `version_compare( $affwp_version, '2.7', '>=' )` is `true` for me since I have a version higher than `2.7`. Then it tries to load the `bootstrap` method but can't because it runs on the `affwp_plugins_loaded` hook. That hook is only available within AffiliateWP, which of course is not active. To test and at least see the banner in the above screenshot, change `affwp_plugins_loaded` to `plugins_loaded`. So what we need to do here is: 1. Get the notice working again. It should show when AffiliateWP is not active 2. Preserve any changes added which cater to the minimum requirements (you'll see these browsing through the code) 3. Simplify the activation code, remove the duplicate class, fix the text domains After we have thoroughly tested this we need to roll this out to the other addons. It's a super broken experience when a user installs our free addons and there's zero indication of what to do next.
non_process
if affiliatewp is not active there is no longer a requirement admin notice while this addon exhibits the issue it s likely also affecting any other addon using the same activation class i m opening it here because i need to do some work on this addon for the wordpress org repo and this is going to hold us back let s fix it in this addon first and then push it out to other addons the issue in the following screenshot i have just activated the affiliate area shortcodes addon note how affiliatewp is not activated and there is no notice that tells the user that affiliatewp must be activated we used to show a message like this in the case of this addon the notice worked for all versions up until when this was added aside from the notice not displaying there are other issues there are two affiliatewp activation classes included see and the activation code is super convoluted no idea why it needs to have that much code there are incorrect textdomains likely a copy paste error when implementing on another addon see affiliatewp afgf in the codebase this is probably only a small part of this but this conditional is preventing the banner from ever loading the first conditional with version compare affwp version is true for me since i have a version higher than then it tries to load the bootstrap method but can t because it runs on the affwp plugins loaded hook that hook is only available within affiliatewp which of course is not active to test and at least see the banner in the above screenshot change affwp plugins loaded to plugins loaded so what we need to do here is get the notice working again it should show when affiliatewp is not active preserve any changes added which cater to the minimum requirements you ll see these browsing through the code simplify the activation code remove the duplicate class fix the text domains after we have thoroughly tested this we need to roll this out to the other addons it s a super broken experience when a user installs our free addons and there s zero indication of what to do next
0
1,209
3,711,323,307
IssuesEvent
2016-03-02 09:51:32
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
NPE in [keyref] processing (DITA-OT 2.2.2)
bug P1 preprocess/keyref
Following is the result log file. ``` [echo] com.antennahouse.i18n_index.2.3 is integrated. init: [echo] com.antennahouse.i18n_index.2.3 is integrated. check-arg: [mkdir] Created dir: D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 log-arg: [echo] ***************************************************************** [echo] * basedir = D:\DITA-OT\dita-ot-2.2.2 [echo] * dita.dir = D:\DITA-OT\dita-ot-2.2.2 [echo] * transtype = pdf5.ml [echo] * tempdir = D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 [echo] * outputdir = D:\DITA-OT\dita-ot-2.2.2\out [echo] * clean.temp = true [echo] * DITA-OT version = 2.2.2 [echo] * XML parser = Xerces [echo] * XSLT processor = Saxon [echo] * collator = ICU [echo] ***************************************************************** [echo] #Ant properties [echo] #Wed Mar 02 16:53:01 GMT+09:00 2016 [echo] args.grammar.cache=yes [echo] args.input=samples/sample_en/sample_en.ditamap [echo] args.logdir=D\:\\DITA-OT\\dita-ot-2.2.2\\out [echo] args.xml.systemid.set=yes [echo] dita.dir=D\:\\DITA-OT\\dita-ot-2.2.2 [echo] dita.plugin.com.antennahouse.dita.dita13.doctypes.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.dita.dita13.doctypes [echo] dita.plugin.com.antennahouse.i18n_index.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.i18n_index.2.3 [echo] dita.plugin.com.antennahouse.pdf5.ml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.pdf5.ml [echo] dita.plugin.com.antennahouse.samples.form.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.samples.form [echo] dita.plugin.com.sophos.tocjs.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.sophos.tocjs [echo] dita.plugin.org.dita.base.dir=D\:\\DITA-OT\\dita-ot-2.2.2 [echo] dita.plugin.org.dita.docbook.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.docbook [echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsecontent [echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsehelp [echo] dita.plugin.org.dita.html5.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.html5 [echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.htmlhelp [echo] dita.plugin.org.dita.javahelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.javahelp [echo] dita.plugin.org.dita.odt.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.odt [echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.axf [echo] dita.plugin.org.dita.pdf2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2 [echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.fop [echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.xep [echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.dita11 [echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.eclipsemap [echo] dita.plugin.org.dita.troff.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.troff [echo] dita.plugin.org.dita.wordrtf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.wordrtf [echo] dita.plugin.org.dita.xhtml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.xhtml [echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_2 [echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_3 [echo] dita.temp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\temp\\temp20160302165301227 [echo] preprocess.copy-html.skip=true [echo] preprocess.copy-image.skip=true [echo] ***************************************************************** build-init: preprocess.init: [echo] ***************************************************************** [echo] * input = samples/sample_en/sample_en.ditamap [echo] ***************************************************************** gen-list: [gen-list] Using Xerces grammar pool for DTD and schema caching. [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml [gen-list] Serializing job specification debug-filter: [filter] Using Xerces grammar pool for DTD and schema caching. [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\outditafiles.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamap.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\image.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\flagimage.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\html.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subjectscheme.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\copytosource.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subtargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\user.input.file.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl mapref-check: mapref: [mapref] Transforming into D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 [mapref] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\preprocess\mapref.xsl [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\m_keydef.ditamap [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\glossary_en.ditamap [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\sample_en.ditamap branch-filter: [branch-filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap copy-image: copy-html: copy-flag-check: copy-flag: copy-files: keyref: [keyref] Reading file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap [keyref] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/topics-en/p_longquoteref.xml BUILD FAILED D:\DITA-OT\dita-ot-2.2.2\build.xml:41: The following error occurred while executing this line: D:\DITA-OT\dita-ot-2.2.2\plugins\org.dita.base\build_preprocess.xml:270: java.lang.NullPointerException at org.dita.dost.writer.KeyrefPaser.processElement(KeyrefPaser.java:374) at org.dita.dost.writer.KeyrefPaser.startElement(KeyrefPaser.java:364) at org.xml.sax.helpers.XMLFilterImpl.startElement(Unknown Source) at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:62) at org.dita.dost.writer.ConkeyrefFilter.startElement(ConkeyrefFilter.java:89) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(Unknown Source) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:404) at net.sf.saxon.event.Sender.send(Sender.java:193) at net.sf.saxon.IdentityTransformer.transform(IdentityTransformer.java:30) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:259) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:219) at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:234) at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:93) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.Project.executeTarget(Project.java:1364) at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.Main.runBuild(Main.java:851) at org.apache.tools.ant.Main.startAnt(Main.java:235) at org.apache.tools.ant.launch.Launcher.run(Launcher.java:280) at org.apache.tools.ant.launch.Launcher.main(Launcher.java:109) Total time: 4 seconds ``` It seems that p_longquoteref.xml has no validation errors. I attached used DITA instance. [sample_en.zip](https://github.com/dita-ot/dita-ot/files/154305/sample_en.zip)
1.0
NPE in [keyref] processing (DITA-OT 2.2.2) - Following is the result log file. ``` [echo] com.antennahouse.i18n_index.2.3 is integrated. init: [echo] com.antennahouse.i18n_index.2.3 is integrated. check-arg: [mkdir] Created dir: D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 log-arg: [echo] ***************************************************************** [echo] * basedir = D:\DITA-OT\dita-ot-2.2.2 [echo] * dita.dir = D:\DITA-OT\dita-ot-2.2.2 [echo] * transtype = pdf5.ml [echo] * tempdir = D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 [echo] * outputdir = D:\DITA-OT\dita-ot-2.2.2\out [echo] * clean.temp = true [echo] * DITA-OT version = 2.2.2 [echo] * XML parser = Xerces [echo] * XSLT processor = Saxon [echo] * collator = ICU [echo] ***************************************************************** [echo] #Ant properties [echo] #Wed Mar 02 16:53:01 GMT+09:00 2016 [echo] args.grammar.cache=yes [echo] args.input=samples/sample_en/sample_en.ditamap [echo] args.logdir=D\:\\DITA-OT\\dita-ot-2.2.2\\out [echo] args.xml.systemid.set=yes [echo] dita.dir=D\:\\DITA-OT\\dita-ot-2.2.2 [echo] dita.plugin.com.antennahouse.dita.dita13.doctypes.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.dita.dita13.doctypes [echo] dita.plugin.com.antennahouse.i18n_index.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.i18n_index.2.3 [echo] dita.plugin.com.antennahouse.pdf5.ml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.pdf5.ml [echo] dita.plugin.com.antennahouse.samples.form.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.samples.form [echo] dita.plugin.com.sophos.tocjs.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.sophos.tocjs [echo] dita.plugin.org.dita.base.dir=D\:\\DITA-OT\\dita-ot-2.2.2 [echo] dita.plugin.org.dita.docbook.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.docbook [echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsecontent [echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsehelp [echo] dita.plugin.org.dita.html5.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.html5 [echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.htmlhelp [echo] dita.plugin.org.dita.javahelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.javahelp [echo] dita.plugin.org.dita.odt.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.odt [echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.axf [echo] dita.plugin.org.dita.pdf2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2 [echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.fop [echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.xep [echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.dita11 [echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.eclipsemap [echo] dita.plugin.org.dita.troff.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.troff [echo] dita.plugin.org.dita.wordrtf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.wordrtf [echo] dita.plugin.org.dita.xhtml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.xhtml [echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_2 [echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_3 [echo] dita.temp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\temp\\temp20160302165301227 [echo] preprocess.copy-html.skip=true [echo] preprocess.copy-image.skip=true [echo] ***************************************************************** build-init: preprocess.init: [echo] ***************************************************************** [echo] * input = samples/sample_en/sample_en.ditamap [echo] ***************************************************************** gen-list: [gen-list] Using Xerces grammar pool for DTD and schema caching. [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml [gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml [gen-list] Serializing job specification debug-filter: [filter] Using Xerces grammar pool for DTD and schema caching. [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml [filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\outditafiles.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamap.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\image.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\flagimage.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\html.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subjectscheme.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\copytosource.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subtargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\user.input.file.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl [job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list [job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl mapref-check: mapref: [mapref] Transforming into D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227 [mapref] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\preprocess\mapref.xsl [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\m_keydef.ditamap [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\glossary_en.ditamap [mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\sample_en.ditamap branch-filter: [branch-filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap copy-image: copy-html: copy-flag-check: copy-flag: copy-files: keyref: [keyref] Reading file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap [keyref] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/topics-en/p_longquoteref.xml BUILD FAILED D:\DITA-OT\dita-ot-2.2.2\build.xml:41: The following error occurred while executing this line: D:\DITA-OT\dita-ot-2.2.2\plugins\org.dita.base\build_preprocess.xml:270: java.lang.NullPointerException at org.dita.dost.writer.KeyrefPaser.processElement(KeyrefPaser.java:374) at org.dita.dost.writer.KeyrefPaser.startElement(KeyrefPaser.java:364) at org.xml.sax.helpers.XMLFilterImpl.startElement(Unknown Source) at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:62) at org.dita.dost.writer.ConkeyrefFilter.startElement(ConkeyrefFilter.java:89) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(Unknown Source) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:404) at net.sf.saxon.event.Sender.send(Sender.java:193) at net.sf.saxon.IdentityTransformer.transform(IdentityTransformer.java:30) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:259) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:219) at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:234) at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:93) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.Project.executeTarget(Project.java:1364) at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.Main.runBuild(Main.java:851) at org.apache.tools.ant.Main.startAnt(Main.java:235) at org.apache.tools.ant.launch.Launcher.run(Launcher.java:280) at org.apache.tools.ant.launch.Launcher.main(Launcher.java:109) Total time: 4 seconds ``` It seems that p_longquoteref.xml has no validation errors. I attached used DITA instance. [sample_en.zip](https://github.com/dita-ot/dita-ot/files/154305/sample_en.zip)
process
npe in processing dita ot following is the result log file com antennahouse index is integrated init com antennahouse index is integrated check arg created dir d dita ot dita ot temp log arg basedir d dita ot dita ot dita dir d dita ot dita ot transtype ml tempdir d dita ot dita ot temp outputdir d dita ot dita ot out clean temp true dita ot version xml parser xerces xslt processor saxon collator icu ant properties wed mar gmt args grammar cache yes args input samples sample en sample en ditamap args logdir d dita ot dita ot out args xml systemid set yes dita dir d dita ot dita ot dita plugin com antennahouse dita doctypes dir d dita ot dita ot plugins com antennahouse dita doctypes dita plugin com antennahouse index dir d dita ot dita ot plugins com antennahouse index dita plugin com antennahouse ml dir d dita ot dita ot plugins com antennahouse ml dita plugin com antennahouse samples form dir d dita ot dita ot plugins com antennahouse samples form dita plugin com sophos tocjs dir d dita ot dita ot plugins com sophos tocjs dita plugin org dita base dir d dita ot dita ot dita plugin org dita docbook dir d dita ot dita ot plugins org dita docbook dita plugin org dita eclipsecontent dir d dita ot dita ot plugins org dita eclipsecontent dita plugin org dita eclipsehelp dir d dita ot dita ot plugins org dita eclipsehelp dita plugin org dita dir d dita ot dita ot plugins org dita dita plugin org dita htmlhelp dir d dita ot dita ot plugins org dita htmlhelp dita plugin org dita javahelp dir d dita ot dita ot plugins org dita javahelp dita plugin org dita odt dir d dita ot dita ot plugins org dita odt dita plugin org dita axf dir d dita ot dita ot plugins org dita axf dita plugin org dita dir d dita ot dita ot plugins org dita dita plugin org dita fop dir d dita ot dita ot plugins org dita fop dita plugin org dita xep dir d dita ot dita ot plugins org dita xep dita plugin org dita specialization dir d dita ot dita ot plugins org dita specialization dita plugin org dita specialization eclipsemap dir d dita ot dita ot plugins org dita specialization eclipsemap dita plugin org dita troff dir d dita ot dita ot plugins org dita troff dita plugin org dita wordrtf dir d dita ot dita ot plugins org dita wordrtf dita plugin org dita xhtml dir d dita ot dita ot plugins org dita xhtml dita plugin org oasis open dita dir d dita ot dita ot plugins org oasis open dita dita plugin org oasis open dita dir d dita ot dita ot plugins org oasis open dita dita temp dir d dita ot dita ot temp preprocess copy html skip true preprocess copy image skip true build init preprocess init input samples sample en sample en ditamap gen list using xerces grammar pool for dtd and schema caching processing file d dita ot dita ot samples sample en sample en ditamap processing file d dita ot dita ot samples sample en topics en dita sample xml processing file d dita ot dita ot samples sample en topics en c preface xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en c test introduction xml processing file d dita ot dita ot samples sample en topics en p abstract xml processing file d dita ot dita ot samples sample en topics en p xref xml processing file d dita ot dita ot samples sample en topics en p note xml processing file d dita ot dita ot samples sample en topics en p bodyelements xml processing file d dita ot dita ot samples sample en topics en p miscellaneouselements xml processing file d dita ot dita ot samples sample en topics en p specializationelements xml processing file d dita ot dita ot samples sample en topics en p typographic xml processing file d dita ot dita ot samples sample en topics en p programmingelements xml processing file d dita ot dita ot samples sample en topics en p softwareelements xml processing file d dita ot dita ot samples sample en topics en p utilityelements xml processing file d dita ot dita ot samples sample en topics en p fig xml processing file d dita ot dita ot samples sample en topics en p table xml processing file d dita ot dita ot samples sample en topics en r properties xml processing file d dita ot dita ot samples sample en topics en t logging in to client xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p longdescref xml processing file d dita ot dita ot samples sample en topics en p longquoteref xml processing file d dita ot dita ot samples sample en topics en m keydef ditamap processing file d dita ot dita ot samples sample en topics en p sectiondiv xml processing file d dita ot dita ot samples sample en topics en p reference to no print xml processing file d dita ot dita ot samples sample en topics en p complecated index example xml processing file d dita ot dita ot samples sample en topics en r xslt xml processing file d dita ot dita ot samples sample en topics en p dir attribute xml processing file d dita ot dita ot samples sample en topics en p backmatter xml processing file d dita ot dita ot samples sample en topics en glossary en ditamap processing file d dita ot dita ot samples sample en topics en p no print xml processing file d dita ot dita ot samples sample en topics en r xml processing file d dita ot dita ot samples sample en topics en gloss xslt xml processing file d dita ot dita ot samples sample en topics en gloss xslfo xml processing file d dita ot dita ot samples sample en topics en gloss xmlandhtml xml processing file d dita ot dita ot samples sample en topics en gloss xmlschema xml processing file d dita ot dita ot samples sample en topics en gloss dtd xml processing file d dita ot dita ot samples sample en topics en gloss dom xml processing file d dita ot dita ot samples sample en topics en gloss sax xml serializing job specification debug filter using xerces grammar pool for dtd and schema caching processing file d dita ot dita ot samples sample en topics en p programmingelements xml processing file d dita ot dita ot samples sample en topics en p longdescref xml processing file d dita ot dita ot samples sample en topics en glossary en ditamap processing file d dita ot dita ot samples sample en topics en c preface xml processing file d dita ot dita ot samples sample en topics en p typographic xml processing file d dita ot dita ot samples sample en topics en p table xml processing file d dita ot dita ot samples sample en topics en dita sample xml processing file d dita ot dita ot samples sample en topics en p reference to no print xml processing file d dita ot dita ot samples sample en topics en r xml processing file d dita ot dita ot samples sample en topics en p miscellaneouselements xml processing file d dita ot dita ot samples sample en sample en ditamap processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p fig xml processing file d dita ot dita ot samples sample en topics en p specializationelements xml processing file d dita ot dita ot samples sample en topics en gloss xslfo xml processing file d dita ot dita ot samples sample en topics en p xref xml processing file d dita ot dita ot samples sample en topics en p note xml processing file d dita ot dita ot samples sample en topics en r properties xml processing file d dita ot dita ot samples sample en topics en gloss xmlandhtml xml processing file d dita ot dita ot samples sample en topics en p utilityelements xml processing file d dita ot dita ot samples sample en topics en gloss dtd xml processing file d dita ot dita ot samples sample en topics en gloss dom xml processing file d dita ot dita ot samples sample en topics en p complecated index example xml processing file d dita ot dita ot samples sample en topics en gloss xmlschema xml processing file d dita ot dita ot samples sample en topics en p dir attribute xml processing file d dita ot dita ot samples sample en topics en m keydef ditamap processing file d dita ot dita ot samples sample en topics en gloss sax xml processing file d dita ot dita ot samples sample en topics en r xslt xml processing file d dita ot dita ot samples sample en topics en p abstract xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p no print xml processing file d dita ot dita ot samples sample en topics en p bodyelements xml processing file d dita ot dita ot samples sample en topics en p sectiondiv xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en gloss xslt xml processing file d dita ot dita ot samples sample en topics en p longquoteref xml processing file d dita ot dita ot samples sample en topics en c test introduction xml processing file d dita ot dita ot samples sample en topics en t logging in to client xml processing file d dita ot dita ot samples sample en topics en p backmatter xml processing file d dita ot dita ot samples sample en topics en p softwareelements xml processing d dita ot dita ot temp job xml to d dita ot dita ot temp outditafiles list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamapandtopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamap list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hrefditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conref list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp image list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp flagimage list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp html list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp canditopics list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp subjectscheme list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp copytosource list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp subtargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp resourceonly list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp user input file list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conref list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hrefditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamapandtopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp canditopics list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp resourceonly list loading stylesheet d dita ot dita ot xsl job helper xsl mapref check mapref transforming into d dita ot dita ot temp loading stylesheet d dita ot dita ot xsl preprocess mapref xsl processing d dita ot dita ot temp topics en m keydef ditamap processing d dita ot dita ot temp topics en glossary en ditamap processing d dita ot dita ot temp sample en ditamap branch filter processing file d dita ot dita ot temp sample en ditamap copy image copy html copy flag check copy flag copy files keyref reading file d dita ot dita ot temp sample en ditamap processing file d dita ot dita ot temp topics en p longquoteref xml build failed d dita ot dita ot build xml the following error occurred while executing this line d dita ot dita ot plugins org dita base build preprocess xml java lang nullpointerexception at org dita dost writer keyrefpaser processelement keyrefpaser java at org dita dost writer keyrefpaser startelement keyrefpaser java at org xml sax helpers xmlfilterimpl startelement unknown source at org dita dost writer topicfragmentfilter startelement topicfragmentfilter java at org dita dost writer conkeyreffilter startelement conkeyreffilter java at org apache xerces parsers abstractsaxparser startelement unknown source at org apache xerces parsers abstractxmldocumentparser emptyelement unknown source at org apache xerces impl xmlnsdocumentscannerimpl scanstartelement unknown source at org apache xerces impl xmldocumentfragmentscannerimpl fragmentcontentdispatcher dispatch unknown source at org apache xerces impl xmldocumentfragmentscannerimpl scandocument unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers abstractsaxparser parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at net sf saxon event sender sendsaxsource sender java at net sf saxon event sender send sender java at net sf saxon identitytransformer transform identitytransformer java at org dita dost util xmlutils transform xmlutils java at org dita dost util xmlutils transform xmlutils java at org dita dost module keyrefmodule processfile keyrefmodule java at org dita dost module keyrefmodule execute keyrefmodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant project executetarget project java at org apache tools ant helper defaultexecutor executetargets defaultexecutor java at org apache tools ant project executetargets project java at org apache tools ant main runbuild main java at org apache tools ant main startant main java at org apache tools ant launch launcher run launcher java at org apache tools ant launch launcher main launcher java total time seconds it seems that p longquoteref xml has no validation errors i attached used dita instance
1
2,608
5,367,284,967
IssuesEvent
2017-02-22 03:27:04
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [eng] Mélenchon - FAQ 30/11/16
Language: English Process: [6] Approved
# Video title MÉLENCHON - FAQ - 6E RÉPUBLIQUE, CANNABIS, BIO, YOUTUBE, SNOWDEN, ASSANGE… # URL https://www.youtube.com/watch?v=GkPZ3sIhMPg # Youtube subtitles language Anglais # Duration 48:34 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&action_mde_edit_form=1&ui=hd&v=GkPZ3sIhMPg&lang=en&ref=player&tab=captions
1.0
[subtitles] [eng] Mélenchon - FAQ 30/11/16 - # Video title MÉLENCHON - FAQ - 6E RÉPUBLIQUE, CANNABIS, BIO, YOUTUBE, SNOWDEN, ASSANGE… # URL https://www.youtube.com/watch?v=GkPZ3sIhMPg # Youtube subtitles language Anglais # Duration 48:34 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&action_mde_edit_form=1&ui=hd&v=GkPZ3sIhMPg&lang=en&ref=player&tab=captions
process
mélenchon faq video title mélenchon faq république cannabis bio youtube snowden assange… url youtube subtitles language anglais duration subtitles url
1
19,080
25,126,461,506
IssuesEvent
2022-11-09 12:07:52
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Only use buildjet action runner where needed
process/candidate kind/tech team/schema team/client
We use buildjet runner in multiple actions https://github.com/prisma/prisma-engines/search?q=buildjet Buildjet is great but we should only use it where it adds value, when it makes tests a lot faster. Overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue. Example situation https://prisma-company.slack.com/archives/C1FPU5FPT/p1667986934989559 Where a ~3 min long action was waiting for ~1h20 <img width="725" alt="Screenshot 2022-11-09 at 12 18 20" src="https://user-images.githubusercontent.com/1328733/200825862-af63c693-150a-4f33-9e32-43529a07bf3b.png"> https://github.com/prisma/prisma-engines/actions/runs/3426620655 - [ ] We should switch to GitHub linux runner for Formatting https://github.com/prisma/prisma-engines/blob/main/.github/workflows/formatting.yml - [ ] Check https://github.com/prisma/prisma-engines/search?q=buildjet and decide if we want to do the switch in other places as well
1.0
Only use buildjet action runner where needed - We use buildjet runner in multiple actions https://github.com/prisma/prisma-engines/search?q=buildjet Buildjet is great but we should only use it where it adds value, when it makes tests a lot faster. Overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue. Example situation https://prisma-company.slack.com/archives/C1FPU5FPT/p1667986934989559 Where a ~3 min long action was waiting for ~1h20 <img width="725" alt="Screenshot 2022-11-09 at 12 18 20" src="https://user-images.githubusercontent.com/1328733/200825862-af63c693-150a-4f33-9e32-43529a07bf3b.png"> https://github.com/prisma/prisma-engines/actions/runs/3426620655 - [ ] We should switch to GitHub linux runner for Formatting https://github.com/prisma/prisma-engines/blob/main/.github/workflows/formatting.yml - [ ] Check https://github.com/prisma/prisma-engines/search?q=buildjet and decide if we want to do the switch in other places as well
process
only use buildjet action runner where needed we use buildjet runner in multiple actions buildjet is great but we should only use it where it adds value when it makes tests a lot faster overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue example situation where a  min long action was waiting for img width alt screenshot at src we should switch to github linux runner for formatting check and decide if we want to do the switch in other places as well
1
13,804
16,564,730,393
IssuesEvent
2021-05-29 06:47:18
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Details Threshold has lost functionality
bug: pending scope: image processing
**Description** The Details Threshold appears inoperable with DT 3.5+2360. The Details Threshold option was working with my copy of DT 3.5+2226 **To Reproduce** 1. Activate a module such as local contrast and enable the parametric mask option and activate the mask. The display will become entirely yellow 2. Details Threshold slider has no effect on the mask when moved right or left 3. I also turned off the mask verified that the slider has no effect on the image, so it's not just an issue with the mask **Screenshots** See attached. 1. Original image with DT 3.5+2360 2. Parametric mask activated with mask displayed 3. Mask display with Details Threshold moved left 4. Mask display with Details Threshold moved right 5. Correctly operating Details Threshold using DT 3.5+2226 **Which commit introduced the error** As mentioned, this seems to be a new error introduced since DT 3.5+2226 **Platform** darktable version : 3.5.0+2360~ge734491a9 OS : Windows 10 Memory : 8 GB Graphics card : AMD A8-7410 APU with AMD Radeon R5 Graphics, 2.20 Ghz Graphics driver : Unk OpenCL installed : Yes OpenCL activated : No (Open CL always causes my system to lock up - different issue) Xorg : Unk Desktop : HP Pavalion GTK+ : Unk gcc : Unk cflags : Unk CMAKE_BUILD_TYPE : Unk ![1 Original Image](https://user-images.githubusercontent.com/61152378/119899887-beb7d500-bf11-11eb-8e13-3fa627230893.JPG) ![2 Mask activated](https://user-images.githubusercontent.com/61152378/119899888-beb7d500-bf11-11eb-9fcd-15444035b4cf.JPG) ![3 With Details Threshold slider adjusted to the left](https://user-images.githubusercontent.com/61152378/119899890-bf506b80-bf11-11eb-9c5e-3d3900cd740b.JPG) ![4 With Details Threshold slider adjusted to the right](https://user-images.githubusercontent.com/61152378/119899891-bf506b80-bf11-11eb-9733-a37a7cf82ba3.JPG) ![5 DT 3 5+2226 with Details Threshold slider adjusted to the left](https://user-images.githubusercontent.com/61152378/119899892-bf506b80-bf11-11eb-9573-a53abcac1502.JPG)
1.0
Details Threshold has lost functionality - **Description** The Details Threshold appears inoperable with DT 3.5+2360. The Details Threshold option was working with my copy of DT 3.5+2226 **To Reproduce** 1. Activate a module such as local contrast and enable the parametric mask option and activate the mask. The display will become entirely yellow 2. Details Threshold slider has no effect on the mask when moved right or left 3. I also turned off the mask verified that the slider has no effect on the image, so it's not just an issue with the mask **Screenshots** See attached. 1. Original image with DT 3.5+2360 2. Parametric mask activated with mask displayed 3. Mask display with Details Threshold moved left 4. Mask display with Details Threshold moved right 5. Correctly operating Details Threshold using DT 3.5+2226 **Which commit introduced the error** As mentioned, this seems to be a new error introduced since DT 3.5+2226 **Platform** darktable version : 3.5.0+2360~ge734491a9 OS : Windows 10 Memory : 8 GB Graphics card : AMD A8-7410 APU with AMD Radeon R5 Graphics, 2.20 Ghz Graphics driver : Unk OpenCL installed : Yes OpenCL activated : No (Open CL always causes my system to lock up - different issue) Xorg : Unk Desktop : HP Pavalion GTK+ : Unk gcc : Unk cflags : Unk CMAKE_BUILD_TYPE : Unk ![1 Original Image](https://user-images.githubusercontent.com/61152378/119899887-beb7d500-bf11-11eb-8e13-3fa627230893.JPG) ![2 Mask activated](https://user-images.githubusercontent.com/61152378/119899888-beb7d500-bf11-11eb-9fcd-15444035b4cf.JPG) ![3 With Details Threshold slider adjusted to the left](https://user-images.githubusercontent.com/61152378/119899890-bf506b80-bf11-11eb-9c5e-3d3900cd740b.JPG) ![4 With Details Threshold slider adjusted to the right](https://user-images.githubusercontent.com/61152378/119899891-bf506b80-bf11-11eb-9733-a37a7cf82ba3.JPG) ![5 DT 3 5+2226 with Details Threshold slider adjusted to the left](https://user-images.githubusercontent.com/61152378/119899892-bf506b80-bf11-11eb-9573-a53abcac1502.JPG)
process
details threshold has lost functionality description the details threshold appears inoperable with dt the details threshold option was working with my copy of dt to reproduce activate a module such as local contrast and enable the parametric mask option and activate the mask the display will become entirely yellow details threshold slider has no effect on the mask when moved right or left i also turned off the mask verified that the slider has no effect on the image so it s not just an issue with the mask screenshots see attached original image with dt parametric mask activated with mask displayed mask display with details threshold moved left mask display with details threshold moved right correctly operating details threshold using dt which commit introduced the error as mentioned this seems to be a new error introduced since dt platform darktable version os windows memory gb graphics card amd apu with amd radeon graphics ghz graphics driver unk opencl installed yes opencl activated no open cl always causes my system to lock up different issue xorg unk desktop hp pavalion gtk unk gcc unk cflags unk cmake build type unk
1
172,619
13,325,342,247
IssuesEvent
2020-08-27 09:49:08
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpcc/w=100/nodes=3/chaos=true failed
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
[(roachtest).tpcc/w=100/nodes=3/chaos=true failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=buildLog) on [release-19.2@603c352969903579f78c7a580afd729bb198de22](https://github.com/cockroachdb/cockroach/commits/603c352969903579f78c7a580afd729bb198de22): ``` | 40.0s 0 15.0 10.6 71.3 113.2 192.9 192.9 orderStatus | 40.0s 0 124.0 106.0 302.0 1140.9 2147.5 2818.6 payment | 40.0s 0 10.0 10.5 56.6 1610.6 1610.6 1610.6 stockLevel | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 41.0s 0 23.0 10.8 302.0 503.3 570.4 570.4 delivery | 41.0s 0 145.0 106.9 486.5 939.5 1409.3 1409.3 newOrder | 41.0s 0 17.0 10.8 48.2 113.2 134.2 134.2 orderStatus | 41.0s 0 142.0 106.9 302.0 1342.2 3221.2 3758.1 payment | 41.0s 0 16.0 10.7 54.5 302.0 318.8 318.8 stockLevel | 42.0s 0 16.0 10.9 453.0 838.9 838.9 838.9 delivery | 42.0s 0 130.0 107.5 520.1 939.5 1476.4 1610.6 newOrder | 42.0s 0 7.0 10.7 75.5 151.0 151.0 151.0 orderStatus | 42.0s 0 112.0 107.0 369.1 906.0 1275.1 1677.7 payment | 42.0s 0 11.0 10.7 88.1 209.7 260.0 260.0 stockLevel | 43.0s 0 14.0 11.0 402.7 838.9 1275.1 1275.1 delivery | 43.0s 0 129.0 108.0 570.4 1073.7 1275.1 1610.6 newOrder | 43.0s 0 11.0 10.7 54.5 130.0 167.8 167.8 orderStatus | 43.0s 0 123.0 107.4 369.1 1476.4 1811.9 2818.6 payment | 43.0s 0 11.0 10.7 134.2 469.8 1677.7 1677.7 stockLevel Wraps: (5) exit status 30 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2456,tpcc.go:168,tpcc.go:312,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452 | main.runTPCC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:168 | main.registerTPCC.func5 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:312 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5222 | runtime.main | /usr/local/go/src/runtime/proc.go:190 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (6) Goexit() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/tpcc/w=100/nodes=3/chaos=true](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true) Related: - #48199 roachtest: tpcc/w=100/nodes=3/chaos=true failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpcc%2Fw%3D100%2Fnodes%3D3%2Fchaos%3Dtrue.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: tpcc/w=100/nodes=3/chaos=true failed - [(roachtest).tpcc/w=100/nodes=3/chaos=true failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=buildLog) on [release-19.2@603c352969903579f78c7a580afd729bb198de22](https://github.com/cockroachdb/cockroach/commits/603c352969903579f78c7a580afd729bb198de22): ``` | 40.0s 0 15.0 10.6 71.3 113.2 192.9 192.9 orderStatus | 40.0s 0 124.0 106.0 302.0 1140.9 2147.5 2818.6 payment | 40.0s 0 10.0 10.5 56.6 1610.6 1610.6 1610.6 stockLevel | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 41.0s 0 23.0 10.8 302.0 503.3 570.4 570.4 delivery | 41.0s 0 145.0 106.9 486.5 939.5 1409.3 1409.3 newOrder | 41.0s 0 17.0 10.8 48.2 113.2 134.2 134.2 orderStatus | 41.0s 0 142.0 106.9 302.0 1342.2 3221.2 3758.1 payment | 41.0s 0 16.0 10.7 54.5 302.0 318.8 318.8 stockLevel | 42.0s 0 16.0 10.9 453.0 838.9 838.9 838.9 delivery | 42.0s 0 130.0 107.5 520.1 939.5 1476.4 1610.6 newOrder | 42.0s 0 7.0 10.7 75.5 151.0 151.0 151.0 orderStatus | 42.0s 0 112.0 107.0 369.1 906.0 1275.1 1677.7 payment | 42.0s 0 11.0 10.7 88.1 209.7 260.0 260.0 stockLevel | 43.0s 0 14.0 11.0 402.7 838.9 1275.1 1275.1 delivery | 43.0s 0 129.0 108.0 570.4 1073.7 1275.1 1610.6 newOrder | 43.0s 0 11.0 10.7 54.5 130.0 167.8 167.8 orderStatus | 43.0s 0 123.0 107.4 369.1 1476.4 1811.9 2818.6 payment | 43.0s 0 11.0 10.7 134.2 469.8 1677.7 1677.7 stockLevel Wraps: (5) exit status 30 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2456,tpcc.go:168,tpcc.go:312,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452 | main.runTPCC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:168 | main.registerTPCC.func5 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:312 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5222 | runtime.main | /usr/local/go/src/runtime/proc.go:190 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (6) Goexit() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/tpcc/w=100/nodes=3/chaos=true](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true) Related: - #48199 roachtest: tpcc/w=100/nodes=3/chaos=true failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpcc%2Fw%3D100%2Fnodes%3D3%2Fchaos%3Dtrue.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_process
roachtest tpcc w nodes chaos true failed on orderstatus payment stocklevel elapsed errors ops sec inst ops sec cum ms ms ms pmax ms delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go tpcc go tpcc go test runner go monitor failure monitor task failed goexit was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runtpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main registertpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps goexit was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring more artifacts related roachtest tpcc w nodes chaos true failed powered by
0
558,933
16,544,703,664
IssuesEvent
2021-05-27 21:54:06
lalitpagaria/obsei
https://api.github.com/repos/lalitpagaria/obsei
closed
Add Google news and website crawler as Source
high priority observer
Idea to add Google News as Source. Google News provide RSS feed and query support hence it is easy to crawl it. RSS link - ``` https://news.google.com/rss/search?q=[INPUT] ``` For now just add GoogleNews as source later we can add few other news sources. Google RSS feed give title, headlight, date and url. So inorder to fetch full article we need to use another library like https://github.com/adbar/trafilatura
1.0
Add Google news and website crawler as Source - Idea to add Google News as Source. Google News provide RSS feed and query support hence it is easy to crawl it. RSS link - ``` https://news.google.com/rss/search?q=[INPUT] ``` For now just add GoogleNews as source later we can add few other news sources. Google RSS feed give title, headlight, date and url. So inorder to fetch full article we need to use another library like https://github.com/adbar/trafilatura
non_process
add google news and website crawler as source idea to add google news as source google news provide rss feed and query support hence it is easy to crawl it rss link for now just add googlenews as source later we can add few other news sources google rss feed give title headlight date and url so inorder to fetch full article we need to use another library like
0
22,279
30,831,085,346
IssuesEvent
2023-08-02 01:48:50
mmattDonk/AI-TTS-Donations
https://api.github.com/repos/mmattDonk/AI-TTS-Donations
closed
[SOLP-14] "Crossing Paths"
@solrock/processor processor Bug Low priority
basically the \`backend\` server got that adam-driver message, and maybe a Dog_Pirate TTS was processing on the \`processor\` server. so adam-driver got processed first on the same \`processor\` server, so Dog_Pirate's Overlay ID was still on the \`processor\` server, however adam-driver got saved first, so it pushed the adam-driver message to Dog_Pirate's overlay, because their overlay ID was still there, and somehow xBunniBee's overlay ID just got lost ![image.png](https://uploads.linear.app/4832fe74-d6d1-406d-b0cd-97906aff6853/14e63c42-8f6c-455a-85b4-c710946cc1aa/ae70cd7a-29ad-48a3-bd8c-9da50ea669d3) <sub>From [SyncLinear.com](https://synclinear.com) | [SOLP-14](https://linear.app/donk/issue/SOLP-14/crossing-paths)</sub>
2.0
[SOLP-14] "Crossing Paths" - basically the \`backend\` server got that adam-driver message, and maybe a Dog_Pirate TTS was processing on the \`processor\` server. so adam-driver got processed first on the same \`processor\` server, so Dog_Pirate's Overlay ID was still on the \`processor\` server, however adam-driver got saved first, so it pushed the adam-driver message to Dog_Pirate's overlay, because their overlay ID was still there, and somehow xBunniBee's overlay ID just got lost ![image.png](https://uploads.linear.app/4832fe74-d6d1-406d-b0cd-97906aff6853/14e63c42-8f6c-455a-85b4-c710946cc1aa/ae70cd7a-29ad-48a3-bd8c-9da50ea669d3) <sub>From [SyncLinear.com](https://synclinear.com) | [SOLP-14](https://linear.app/donk/issue/SOLP-14/crossing-paths)</sub>
process
crossing paths basically the backend server got that adam driver message and maybe a dog pirate tts was processing on the processor server so adam driver got processed first on the same processor server so dog pirate s overlay id was still on the processor server however adam driver got saved first so it pushed the adam driver message to dog pirate s overlay because their overlay id was still there and somehow xbunnibee s overlay id just got lost from
1
22,559
31,777,927,522
IssuesEvent
2023-09-12 15:27:18
IMAP-Science-Operations-Center/imap_processing
https://api.github.com/repos/IMAP-Science-Operations-Center/imap_processing
opened
[L5] Produce L1B CDF formatted data products for Ultra
IMAP-Ultra L1 Requirement: Level 5 Untested Parent Req:SDC Execution of L1 Processing Software
### Summary of the L5 requirement The SDC needs to produce L1B CDF formatted data products for the Ultra instrument. The data products for L1B can be found in the Ultra algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement <!DOCTYPE html> SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks - [ ]
1.0
[L5] Produce L1B CDF formatted data products for Ultra - ### Summary of the L5 requirement The SDC needs to produce L1B CDF formatted data products for the Ultra instrument. The data products for L1B can be found in the Ultra algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement <!DOCTYPE html> SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks - [ ]
process
produce cdf formatted data products for ultra summary of the requirement the sdc needs to produce cdf formatted data products for the ultra instrument the data products for can be found in the ultra algorithm document here parent requirement soc sdc sdc execution of processing software the sdc shall be capable of executing processing software to produce data products within seven days of receipt of all requisite data inputs requirement the sdc produces data products via instrument team supplied algorithms and sdc produced processing code soc test sit processing sit tasks
1
23,983
5,008,429,050
IssuesEvent
2016-12-12 19:29:18
DjalmaPessoa/convey
https://api.github.com/repos/DjalmaPessoa/convey
opened
create a replication vignette for svygei
documentation
@guilhermejacob how do we hit the numbers `GE(-1)` ? flip back assignee to me once you finish or if we shouldn't bother? thanks library(convey) library(survey) library(foreign) tf <- tempfile() download.file( "http://econpapers.repec.org/scripts/redir.pf?u=http%3A%2F%2Frepec.org%2Fnasug2006%2Fnasug2006_jenkins.zip;h=repec:boc:asug06:16" , tf , mode = 'wb' ) z <- unzip( tf , exdir = tempdir() ) x81 <- read.dta( grep( "ifs81" , z , value = TRUE ) ) x85 <- read.dta( grep( "ifs85" , z , value = TRUE ) ) x91 <- read.dta( grep( "ifs91" , z , value = TRUE ) ) x <- rbind( x81 , x85 , x91 ) y <- svydesign( ~ hrn , data = x , weights = ~ wgt ) z <- convey_prep( y ) z81 <- subset( z , year == 1981 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 0 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 3 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 0 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 2 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 3 ) # . svygei x if year == 1981 # Warning: x has 20 values = 0. Not used in calculations # Complex survey estimates of Generalized Entropy inequality indices # pweight: wgt Number of obs = 9752 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7459 # Population size = 54766261 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # GE(-1) | .1902062 .02474921 7.69 0.000 .1416987 .2387138 # MLD | .1142851 .00275138 41.54 0.000 .1088925 .1196777 # Theil | .1116923 .00226489 49.31 0.000 .1072532 .1161314 # GE(2) | .128793 .00330774 38.94 0.000 .1223099 .135276 # GE(3) | .1739994 .00662015 26.28 0.000 .1610242 .1869747 # --------------------------------------------------------------------------- svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 1.5 ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2.5 ) # . * account for clustering within HHs # . version 8: svyset [pweight = wgt], psu(hrn) # pweight is wgt # psu is hrn # . svyatk x if year == 1981 # Warning: x has 20 values = 0. Not used in calculations # Complex survey estimates of Atkinson inequality indices # pweight: wgt Number of obs = 9752 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7459 # Population size = 54766261 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # A(0.5) | .0543239 .00107583 50.49 0.000 .0522153 .0564324 # A(1) | .1079964 .00245424 44.00 0.000 .1031862 .1128066 # A(1.5) | .1701794 .0066943 25.42 0.000 .1570588 .1833 # A(2) | .2755788 .02597608 10.61 0.000 .2246666 .326491 # A(2.5) | .4992701 .06754311 7.39 0.000 .366888 .6316522 # --------------------------------------------------------------------------- z81_two <- subset( z , year == 1981 & eybhc0 > 1 ) svyatk( ~ eybhc0 , z81_two , epsilon = 0.5 ) svyatk( ~ eybhc0 , z81_two ) svyatk( ~ eybhc0 , z81_two , epsilon = 1.5 ) svyatk( ~ eybhc0 , z81_two , epsilon = 2 ) svyatk( ~ eybhc0 , z81_two , epsilon = 2.5 ) # . svyatk x if year == 1981 & x >= 1 # Complex survey estimates of Atkinson inequality indices # pweight: wgt Number of obs = 9748 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7457 # Population size = 54744234 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # A(0.5) | .0540059 .00105011 51.43 0.000 .0519477 .0560641 # A(1) | .1066082 .00223318 47.74 0.000 .1022313 .1109852 # A(1.5) | .1638299 .00483069 33.91 0.000 .154362 .1732979 # A(2) | .2443206 .01425258 17.14 0.000 .2163861 .2722552 # A(2.5) | .394787 .04155221 9.50 0.000 .3133461 .4762278 # --------------------------------------------------------------------------- # . svygei x if year == 1985 & x >= 1 # Complex survey estimates of Generalized Entropy inequality indices # pweight: wgt Number of obs = 8969 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 6950 # Population size = 55042871 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # GE(-1) | .1602358 .00936931 17.10 0.000 .1418723 .1785993 # MLD | .127616 .00332187 38.42 0.000 .1211052 .1341267 # Theil | .1337177 .00406302 32.91 0.000 .1257543 .141681 # GE(2) | .1676393 .00730057 22.96 0.000 .1533304 .1819481 # GE(3) | .2609507 .01850689 14.10 0.000 .2246779 .2972235 # ---------------------------------------------------------------------------
1.0
create a replication vignette for svygei - @guilhermejacob how do we hit the numbers `GE(-1)` ? flip back assignee to me once you finish or if we shouldn't bother? thanks library(convey) library(survey) library(foreign) tf <- tempfile() download.file( "http://econpapers.repec.org/scripts/redir.pf?u=http%3A%2F%2Frepec.org%2Fnasug2006%2Fnasug2006_jenkins.zip;h=repec:boc:asug06:16" , tf , mode = 'wb' ) z <- unzip( tf , exdir = tempdir() ) x81 <- read.dta( grep( "ifs81" , z , value = TRUE ) ) x85 <- read.dta( grep( "ifs85" , z , value = TRUE ) ) x91 <- read.dta( grep( "ifs91" , z , value = TRUE ) ) x <- rbind( x81 , x85 , x91 ) y <- svydesign( ~ hrn , data = x , weights = ~ wgt ) z <- convey_prep( y ) z81 <- subset( z , year == 1981 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 0 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 3 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 0 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 2 ) svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 3 ) # . svygei x if year == 1981 # Warning: x has 20 values = 0. Not used in calculations # Complex survey estimates of Generalized Entropy inequality indices # pweight: wgt Number of obs = 9752 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7459 # Population size = 54766261 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # GE(-1) | .1902062 .02474921 7.69 0.000 .1416987 .2387138 # MLD | .1142851 .00275138 41.54 0.000 .1088925 .1196777 # Theil | .1116923 .00226489 49.31 0.000 .1072532 .1161314 # GE(2) | .128793 .00330774 38.94 0.000 .1223099 .135276 # GE(3) | .1739994 .00662015 26.28 0.000 .1610242 .1869747 # --------------------------------------------------------------------------- svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 1.5 ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 ) svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2.5 ) # . * account for clustering within HHs # . version 8: svyset [pweight = wgt], psu(hrn) # pweight is wgt # psu is hrn # . svyatk x if year == 1981 # Warning: x has 20 values = 0. Not used in calculations # Complex survey estimates of Atkinson inequality indices # pweight: wgt Number of obs = 9752 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7459 # Population size = 54766261 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # A(0.5) | .0543239 .00107583 50.49 0.000 .0522153 .0564324 # A(1) | .1079964 .00245424 44.00 0.000 .1031862 .1128066 # A(1.5) | .1701794 .0066943 25.42 0.000 .1570588 .1833 # A(2) | .2755788 .02597608 10.61 0.000 .2246666 .326491 # A(2.5) | .4992701 .06754311 7.39 0.000 .366888 .6316522 # --------------------------------------------------------------------------- z81_two <- subset( z , year == 1981 & eybhc0 > 1 ) svyatk( ~ eybhc0 , z81_two , epsilon = 0.5 ) svyatk( ~ eybhc0 , z81_two ) svyatk( ~ eybhc0 , z81_two , epsilon = 1.5 ) svyatk( ~ eybhc0 , z81_two , epsilon = 2 ) svyatk( ~ eybhc0 , z81_two , epsilon = 2.5 ) # . svyatk x if year == 1981 & x >= 1 # Complex survey estimates of Atkinson inequality indices # pweight: wgt Number of obs = 9748 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 7457 # Population size = 54744234 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # A(0.5) | .0540059 .00105011 51.43 0.000 .0519477 .0560641 # A(1) | .1066082 .00223318 47.74 0.000 .1022313 .1109852 # A(1.5) | .1638299 .00483069 33.91 0.000 .154362 .1732979 # A(2) | .2443206 .01425258 17.14 0.000 .2163861 .2722552 # A(2.5) | .394787 .04155221 9.50 0.000 .3133461 .4762278 # --------------------------------------------------------------------------- # . svygei x if year == 1985 & x >= 1 # Complex survey estimates of Generalized Entropy inequality indices # pweight: wgt Number of obs = 8969 # Strata: <one> Number of strata = 1 # PSU: hrn Number of PSUs = 6950 # Population size = 55042871 # --------------------------------------------------------------------------- # Index | Estimate Std. Err. z P>|z| [95% Conf. Interval] # ---------+----------------------------------------------------------------- # GE(-1) | .1602358 .00936931 17.10 0.000 .1418723 .1785993 # MLD | .127616 .00332187 38.42 0.000 .1211052 .1341267 # Theil | .1337177 .00406302 32.91 0.000 .1257543 .141681 # GE(2) | .1676393 .00730057 22.96 0.000 .1533304 .1819481 # GE(3) | .2609507 .01850689 14.10 0.000 .2246779 .2972235 # ---------------------------------------------------------------------------
non_process
create a replication vignette for svygei guilhermejacob how do we hit the numbers ge flip back assignee to me once you finish or if we shouldn t bother thanks library convey library survey library foreign tf tempfile download file tf mode wb z unzip tf exdir tempdir read dta grep z value true read dta grep z value true read dta grep z value true x rbind y svydesign hrn data x weights wgt z convey prep y subset z year svygei subset epsilon svygei subset svygei subset epsilon svygei subset epsilon svygei subset epsilon svygei subset svygei subset epsilon svygei subset epsilon svygei x if year warning x has values not used in calculations complex survey estimates of generalized entropy inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z ge mld theil ge ge svyatk subset svyatk subset epsilon svyatk subset epsilon svyatk subset epsilon account for clustering within hhs version svyset psu hrn pweight is wgt psu is hrn svyatk x if year warning x has values not used in calculations complex survey estimates of atkinson inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z a a a a a two svyatk two epsilon svyatk two svyatk two epsilon svyatk two epsilon svyatk two epsilon svyatk x if year x complex survey estimates of atkinson inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z a a a a a svygei x if year x complex survey estimates of generalized entropy inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z ge mld theil ge ge
0
3,005
6,006,945,885
IssuesEvent
2017-06-06 00:56:15
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Add ability to hide referrers from report (e.g., local sites).
add change log-processing
I've got 80...90% local referrers within a website. With goaccess my.log -a -o report.html I get a nice report including all requests. But in the referrer urls and referrer sites reports, I only find referrers within the same website, as there are so many. Works as intended. The report including the local referrers is correct, but not nice - I'd like to see only the interesting referrers directing from external websites to the local site. So I try a goaccess my.log -a --ignore-referer=www.example.com --ignore-referer=example.com -o report.html and get nice referrer urls and referrer sites reports without the local site referrers. Only problem: the ignore-referer option makes goaccess completely ignore the log lines containing these referrers (correctly behaviour as described in the .conf file: "Ignore referers from being counted"), resulting in a report that is missing many hits. An additional or changed option would be nice: counting the loglines as hits, but ignoring their referrer information
1.0
Add ability to hide referrers from report (e.g., local sites). - I've got 80...90% local referrers within a website. With goaccess my.log -a -o report.html I get a nice report including all requests. But in the referrer urls and referrer sites reports, I only find referrers within the same website, as there are so many. Works as intended. The report including the local referrers is correct, but not nice - I'd like to see only the interesting referrers directing from external websites to the local site. So I try a goaccess my.log -a --ignore-referer=www.example.com --ignore-referer=example.com -o report.html and get nice referrer urls and referrer sites reports without the local site referrers. Only problem: the ignore-referer option makes goaccess completely ignore the log lines containing these referrers (correctly behaviour as described in the .conf file: "Ignore referers from being counted"), resulting in a report that is missing many hits. An additional or changed option would be nice: counting the loglines as hits, but ignoring their referrer information
process
add ability to hide referrers from report e g local sites i ve got local referrers within a website with goaccess my log a o report html i get a nice report including all requests but in the referrer urls and referrer sites reports i only find referrers within the same website as there are so many works as intended the report including the local referrers is correct but not nice i d like to see only the interesting referrers directing from external websites to the local site so i try a goaccess my log a ignore referer ignore referer example com o report html and get nice referrer urls and referrer sites reports without the local site referrers only problem the ignore referer option makes goaccess completely ignore the log lines containing these referrers correctly behaviour as described in the conf file ignore referers from being counted resulting in a report that is missing many hits an additional or changed option would be nice counting the loglines as hits but ignoring their referrer information
1
2,084
4,912,510,406
IssuesEvent
2016-11-23 09:21:22
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Tabs do not display in completed tasks
browser: all bug comp: activiti-processList
1. Start a new process 2. Ensure user task within process has form attached that includes more than one tab 3. Go to tasklist 4. Complete task 5. Go to process list 6. Open task completed **Expected results** All tabs are displayed **Actual results** Only one tab is displayed
1.0
Tabs do not display in completed tasks - 1. Start a new process 2. Ensure user task within process has form attached that includes more than one tab 3. Go to tasklist 4. Complete task 5. Go to process list 6. Open task completed **Expected results** All tabs are displayed **Actual results** Only one tab is displayed
process
tabs do not display in completed tasks start a new process ensure user task within process has form attached that includes more than one tab go to tasklist complete task go to process list open task completed expected results all tabs are displayed actual results only one tab is displayed
1
83,493
10,369,211,431
IssuesEvent
2019-09-08 00:21:30
phetsims/vector-addition
https://api.github.com/repos/phetsims/vector-addition
closed
incorrect labeling of magnitudes in Base Vectors accordion box
design:polish
In 1.0.0-dev.11 (baseline), and still in master, the Base Vectors accordion box looks like this for the polar snap mode: <img width="176" alt="screenshot_1444" src="https://user-images.githubusercontent.com/3046552/64452639-c9cf2f80-d0b4-11e9-9a28-74caa25460a6.png"> @arouinfar shouldn't the magnitudes be labeled as |d&#8407;| and |e&#8407;|, with a vector arrow over each symbol?
1.0
incorrect labeling of magnitudes in Base Vectors accordion box - In 1.0.0-dev.11 (baseline), and still in master, the Base Vectors accordion box looks like this for the polar snap mode: <img width="176" alt="screenshot_1444" src="https://user-images.githubusercontent.com/3046552/64452639-c9cf2f80-d0b4-11e9-9a28-74caa25460a6.png"> @arouinfar shouldn't the magnitudes be labeled as |d&#8407;| and |e&#8407;|, with a vector arrow over each symbol?
non_process
incorrect labeling of magnitudes in base vectors accordion box in dev baseline and still in master the base vectors accordion box looks like this for the polar snap mode img width alt screenshot src arouinfar shouldn t the magnitudes be labeled as d and e with a vector arrow over each symbol
0
12,029
14,738,585,558
IssuesEvent
2021-01-07 05:10:59
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Session expiry functionality to be implemented
Bug P1 Participant manager Process: Dev Process: Release 2 Process: Tested QA
Once session get Expired (30 minutes Inactive), User should be logged out of the application and sign-in page should be displayed
3.0
[PM] Session expiry functionality to be implemented - Once session get Expired (30 minutes Inactive), User should be logged out of the application and sign-in page should be displayed
process
session expiry functionality to be implemented once session get expired minutes inactive user should be logged out of the application and sign in page should be displayed
1
16,216
20,742,611,722
IssuesEvent
2022-03-14 19:13:54
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
Regression test: no crash if sound is not found
Help Wanted Good First Issue! Stale Test process
See stack trace for the cause: https://github.com/ankidroid/Anki-Android/issues/9401 The deliverable will be a unit test which fails before PR https://github.com/ankidroid/Anki-Android/pull/9407 goes in, and passes afterwards ---- Thanks! Could you add a regression test here for when the sound is not in the collection? _Originally posted by @david-allison-1 in https://github.com/ankidroid/Anki-Android/issues/9407#issuecomment-898621819_
1.0
Regression test: no crash if sound is not found - See stack trace for the cause: https://github.com/ankidroid/Anki-Android/issues/9401 The deliverable will be a unit test which fails before PR https://github.com/ankidroid/Anki-Android/pull/9407 goes in, and passes afterwards ---- Thanks! Could you add a regression test here for when the sound is not in the collection? _Originally posted by @david-allison-1 in https://github.com/ankidroid/Anki-Android/issues/9407#issuecomment-898621819_
process
regression test no crash if sound is not found see stack trace for the cause the deliverable will be a unit test which fails before pr goes in and passes afterwards thanks could you add a regression test here for when the sound is not in the collection originally posted by david allison in
1
53,317
11,035,667,136
IssuesEvent
2019-12-07 15:19:46
paulevsGitch/BetterNether
https://api.github.com/repos/paulevsGitch/BetterNether
closed
Extreme load times
Fixed in dev code
Hello I added this to my modpack but I noticed a considerable load time for the Nether, after some tests these are the results: **Nether Performance Test** BOP + BetterNether _37,73s_ BetterNether _35,23s_ BOP _6,08s_ None _5,08s_ And this isn't even on a low-end machine.
1.0
Extreme load times - Hello I added this to my modpack but I noticed a considerable load time for the Nether, after some tests these are the results: **Nether Performance Test** BOP + BetterNether _37,73s_ BetterNether _35,23s_ BOP _6,08s_ None _5,08s_ And this isn't even on a low-end machine.
non_process
extreme load times hello i added this to my modpack but i noticed a considerable load time for the nether after some tests these are the results nether performance test bop betternether betternether bop none and this isn t even on a low end machine
0
53,431
11,053,284,860
IssuesEvent
2019-12-10 11:02:52
wahlgenial/wahlgenial-webapp
https://api.github.com/repos/wahlgenial/wahlgenial-webapp
opened
Teacher_apps Seite: Verlinkung auf die Detailseiten überlappt mit Dropdown
Code Priority: High bug
Die Verlinkung der Detailseiten auf der Lehrer-apps Seite überlappt die aus dem Dropdownmenü. So sind die Verlinkungen im Dropdown an manchen Stellen überschrieben. Das sollte man über ein Mouseover im Dropdownmenü erkennen.
1.0
Teacher_apps Seite: Verlinkung auf die Detailseiten überlappt mit Dropdown - Die Verlinkung der Detailseiten auf der Lehrer-apps Seite überlappt die aus dem Dropdownmenü. So sind die Verlinkungen im Dropdown an manchen Stellen überschrieben. Das sollte man über ein Mouseover im Dropdownmenü erkennen.
non_process
teacher apps seite verlinkung auf die detailseiten überlappt mit dropdown die verlinkung der detailseiten auf der lehrer apps seite überlappt die aus dem dropdownmenü so sind die verlinkungen im dropdown an manchen stellen überschrieben das sollte man über ein mouseover im dropdownmenü erkennen
0
20,234
26,839,596,123
IssuesEvent
2023-02-02 22:45:23
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Spotless Depends on NPM installed and available on $PATH
bug process
### Description When importing the project, the Intellij sync process done to import a new gradle project fails. Spotless attempts to resolve the NPM executable during the configuration phase which is prior to node gradle plugin installing node. ### Steps to reproduce 1. Create new intellij project via VCS or existing source 2. Go to the build tab to see gradle output 3. Notice spotless failures ### Additional context ```console FAILURE: Build failed with an exception. * What went wrong: Could not create task ':common:spotlessJavascriptApply'. Could not create task ':common:spotlessJavascript'. Can't automatically determine npm executable and none was specifically supplied! Spotless tries to find your npm executable automatically. It looks for npm in the following places: - An executable referenced by the java system property 'npm.exec' - if such a system property exists. - The environment variable 'NVM_BIN' - if such an environment variable exists. - The environment variable 'NVM_SYMLINK' - if such an environment variable exists. - The environment variable 'NODE_PATH' - if such an environment variable exists. - In your 'PATH' environment variable If autodiscovery fails for your system, try to set one of the environment variables correctly or try setting the system property 'npm.exec' in the build process to override autodiscovery. * Try: > Run with --info or --debug option to get more log output. * Exception is: com.intellij.openapi.externalSystem.model.ExternalSystemException: Could not create task ':common:spotlessJavascriptApply'. Could not create task ':common:spotlessJavascript'. Can't automatically determine npm executable and none was specifically supplied! Spotless tries to find your npm executable automatically. It looks for npm in the following places: - An executable referenced by the java system property 'npm.exec' - if such a system property exists. - The environment variable 'NVM_BIN' - if such an environment variable exists. - The environment variable 'NVM_SYMLINK' - if such an environment variable exists. - The environment variable 'NODE_PATH' - if such an environment variable exists. - In your 'PATH' environment variable If autodiscovery fails for your system, try to set one of the environment variables correctly or try setting the system property 'npm.exec' in the build process to override autodiscovery. at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:412) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42) at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84) at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43) at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53) at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35) at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49) at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65) at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136) at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49) org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascriptApply'. at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715) at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929) at org.gradle.api.internal.provider.AbstractMinimalProvider.getOrNull(AbstractMinimalProvider.java:97) at org.gradle.api.internal.DefaultNamedDomainObjectCollection.findByName(DefaultNamedDomainObjectCollection.java:295) at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:558) at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:75) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.tasks(GradleProjectBuilder.java:105) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:80) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:58) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildAll(GradleProjectBuilder.java:52) at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:78) at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:60) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithNoParameter.build(DefaultToolingModelBuilderRegistry.java:264) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383) at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334) at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82) at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106) at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:634) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:655) at org.jetbrains.plugins.gradle.model.ClassSetImportModelProvider.populateBuildModels(ClassSetImportModelProvider.java:27) at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:400) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42) at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84) at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43) at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53) at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35) at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49) at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65) at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136) at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascript'. at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715) at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929) at org.gradle.api.internal.provider.AbstractMinimalProvider.calculateValue(AbstractMinimalProvider.java:107) at org.gradle.api.internal.provider.Collectors$ElementFromProvider.collectEntries(Collectors.java:100) at org.gradle.api.internal.provider.Collectors$TypedCollector.collectEntries(Collectors.java:334) at org.gradle.api.internal.provider.Collectors$TypedCollector.collectInto(Collectors.java:329) at org.gradle.api.internal.collections.DefaultPendingSource.realize(DefaultPendingSource.java:62) at org.gradle.api.internal.collections.DefaultPendingSource.realizePending(DefaultPendingSource.java:39) at org.gradle.api.internal.collections.SortedSetElementSource.iterator(SortedSetElementSource.java:63) at org.gradle.api.internal.DefaultDomainObjectCollection.iterator(DefaultDomainObjectCollection.java:128) at org.gradle.api.internal.tasks.DefaultTaskContainer.iterator(DefaultTaskContainer.java:620) at java.base/java.util.AbstractCollection.addAll(AbstractCollection.java:335) at java.base/java.util.TreeSet.addAll(TreeSet.java:309) at java.base/java.util.TreeSet.<init>(TreeSet.java:160) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getAllTasks(TasksFactory.java:56) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.collectTasks(TasksFactory.java:46) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getTasks(TasksFactory.java:78) at org.jetbrains.gradle.profiler.builder.JavaTaskModelBuilderImpl.buildAll(JavaTaskModelBuilderImpl.java:34) at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:107) at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:76) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithParameter.build(DefaultToolingModelBuilderRegistry.java:287) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:374) at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication.reapply(DefaultUserCodeApplicationContext.java:109) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.build(DefaultToolingModelBuilderRegistry.java:374) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383) at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334) at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82) at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106) at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:658) at org.jetbrains.plugins.gradle.model.ClassSetProjectImportModelProvider.populateProjectModels(ClassSetProjectImportModelProvider.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction.getProjectModels(ProjectImportAction.java:338) at org.jetbrains.plugins.gradle.model.ProjectImportAction.access$500(ProjectImportAction.java:42) at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:249) at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:246) at org.jetbrains.plugins.gradle.model.ProjectImportAction.fetchProjectBuildModels(ProjectImportAction.java:262) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:137) ... 87 more Caused by: java.lang.IllegalStateException: Can't automatically determine npm executable and none was specifically supplied! ``` ### Hedera network other ### Version 0.71.0-SNAPSHOT ### Operating system macOS
1.0
Spotless Depends on NPM installed and available on $PATH - ### Description When importing the project, the Intellij sync process done to import a new gradle project fails. Spotless attempts to resolve the NPM executable during the configuration phase which is prior to node gradle plugin installing node. ### Steps to reproduce 1. Create new intellij project via VCS or existing source 2. Go to the build tab to see gradle output 3. Notice spotless failures ### Additional context ```console FAILURE: Build failed with an exception. * What went wrong: Could not create task ':common:spotlessJavascriptApply'. Could not create task ':common:spotlessJavascript'. Can't automatically determine npm executable and none was specifically supplied! Spotless tries to find your npm executable automatically. It looks for npm in the following places: - An executable referenced by the java system property 'npm.exec' - if such a system property exists. - The environment variable 'NVM_BIN' - if such an environment variable exists. - The environment variable 'NVM_SYMLINK' - if such an environment variable exists. - The environment variable 'NODE_PATH' - if such an environment variable exists. - In your 'PATH' environment variable If autodiscovery fails for your system, try to set one of the environment variables correctly or try setting the system property 'npm.exec' in the build process to override autodiscovery. * Try: > Run with --info or --debug option to get more log output. * Exception is: com.intellij.openapi.externalSystem.model.ExternalSystemException: Could not create task ':common:spotlessJavascriptApply'. Could not create task ':common:spotlessJavascript'. Can't automatically determine npm executable and none was specifically supplied! Spotless tries to find your npm executable automatically. It looks for npm in the following places: - An executable referenced by the java system property 'npm.exec' - if such a system property exists. - The environment variable 'NVM_BIN' - if such an environment variable exists. - The environment variable 'NVM_SYMLINK' - if such an environment variable exists. - The environment variable 'NODE_PATH' - if such an environment variable exists. - In your 'PATH' environment variable If autodiscovery fails for your system, try to set one of the environment variables correctly or try setting the system property 'npm.exec' in the build process to override autodiscovery. at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:412) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42) at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84) at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43) at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53) at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35) at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49) at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65) at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136) at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49) org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascriptApply'. at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715) at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929) at org.gradle.api.internal.provider.AbstractMinimalProvider.getOrNull(AbstractMinimalProvider.java:97) at org.gradle.api.internal.DefaultNamedDomainObjectCollection.findByName(DefaultNamedDomainObjectCollection.java:295) at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:558) at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:75) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.tasks(GradleProjectBuilder.java:105) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:80) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:58) at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildAll(GradleProjectBuilder.java:52) at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:78) at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:60) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithNoParameter.build(DefaultToolingModelBuilderRegistry.java:264) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383) at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334) at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82) at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106) at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:634) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:655) at org.jetbrains.plugins.gradle.model.ClassSetImportModelProvider.populateBuildModels(ClassSetImportModelProvider.java:27) at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:400) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42) at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84) at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81) at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43) at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53) at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35) at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49) at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65) at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136) at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascript'. at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715) at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676) at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929) at org.gradle.api.internal.provider.AbstractMinimalProvider.calculateValue(AbstractMinimalProvider.java:107) at org.gradle.api.internal.provider.Collectors$ElementFromProvider.collectEntries(Collectors.java:100) at org.gradle.api.internal.provider.Collectors$TypedCollector.collectEntries(Collectors.java:334) at org.gradle.api.internal.provider.Collectors$TypedCollector.collectInto(Collectors.java:329) at org.gradle.api.internal.collections.DefaultPendingSource.realize(DefaultPendingSource.java:62) at org.gradle.api.internal.collections.DefaultPendingSource.realizePending(DefaultPendingSource.java:39) at org.gradle.api.internal.collections.SortedSetElementSource.iterator(SortedSetElementSource.java:63) at org.gradle.api.internal.DefaultDomainObjectCollection.iterator(DefaultDomainObjectCollection.java:128) at org.gradle.api.internal.tasks.DefaultTaskContainer.iterator(DefaultTaskContainer.java:620) at java.base/java.util.AbstractCollection.addAll(AbstractCollection.java:335) at java.base/java.util.TreeSet.addAll(TreeSet.java:309) at java.base/java.util.TreeSet.<init>(TreeSet.java:160) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getAllTasks(TasksFactory.java:56) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.collectTasks(TasksFactory.java:46) at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getTasks(TasksFactory.java:78) at org.jetbrains.gradle.profiler.builder.JavaTaskModelBuilderImpl.buildAll(JavaTaskModelBuilderImpl.java:34) at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:107) at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:76) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithParameter.build(DefaultToolingModelBuilderRegistry.java:287) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:374) at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication.reapply(DefaultUserCodeApplicationContext.java:109) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.build(DefaultToolingModelBuilderRegistry.java:374) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383) at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345) at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334) at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82) at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106) at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81) at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:658) at org.jetbrains.plugins.gradle.model.ClassSetProjectImportModelProvider.populateProjectModels(ClassSetProjectImportModelProvider.java:31) at org.jetbrains.plugins.gradle.model.ProjectImportAction.getProjectModels(ProjectImportAction.java:338) at org.jetbrains.plugins.gradle.model.ProjectImportAction.access$500(ProjectImportAction.java:42) at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:249) at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:246) at org.jetbrains.plugins.gradle.model.ProjectImportAction.fetchProjectBuildModels(ProjectImportAction.java:262) at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:137) ... 87 more Caused by: java.lang.IllegalStateException: Can't automatically determine npm executable and none was specifically supplied! ``` ### Hedera network other ### Version 0.71.0-SNAPSHOT ### Operating system macOS
process
spotless depends on npm installed and available on path description when importing the project the intellij sync process done to import a new gradle project fails spotless attempts to resolve the npm executable during the configuration phase which is prior to node gradle plugin installing node steps to reproduce create new intellij project via vcs or existing source go to the build tab to see gradle output notice spotless failures additional context console failure build failed with an exception what went wrong could not create task common spotlessjavascriptapply could not create task common spotlessjavascript can t automatically determine npm executable and none was specifically supplied spotless tries to find your npm executable automatically it looks for npm in the following places an executable referenced by the java system property npm exec if such a system property exists the environment variable nvm bin if such an environment variable exists the environment variable nvm symlink if such an environment variable exists the environment variable node path if such an environment variable exists in your path environment variable if autodiscovery fails for your system try to set one of the environment variables correctly or try setting the system property npm exec in the build process to override autodiscovery try run with info or debug option to get more log output exception is com intellij openapi externalsystem model externalsystemexception could not create task common spotlessjavascriptapply could not create task common spotlessjavascript can t automatically determine npm executable and none was specifically supplied spotless tries to find your npm executable automatically it looks for npm in the following places an executable referenced by the java system property npm exec if such a system property exists the environment variable nvm bin if such an environment variable exists the environment variable nvm symlink if such an environment variable exists the environment variable node path if such an environment variable exists in your path environment variable if autodiscovery fails for your system try to set one of the environment variables correctly or try setting the system property npm exec in the build process to override autodiscovery at org jetbrains plugins gradle model projectimportaction addbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org gradle tooling internal consumer connection internalbuildactionadapter execute internalbuildactionadapter java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter runaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle internal buildtree defaultbuildtreemodelcreator frombuildmodel defaultbuildtreemodelcreator java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda runbuild defaultbuildtreelifecyclecontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transition statetransitioncontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller runbuild defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner runclientaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner clientprovidedphasedactionrunner run clientprovidedphasedactionrunner java at org gradle launcher exec chainingbuildactionrunner run chainingbuildactionrunner java at org gradle internal buildtree problemreportingbuildactionrunner run problemreportingbuildactionrunner java at org gradle launcher exec buildoutcomereportingbuildactionrunner run buildoutcomereportingbuildactionrunner java at org gradle tooling internal provider filesystemwatchingbuildactionrunner run filesystemwatchingbuildactionrunner java at org gradle launcher exec buildcompletionnotifyingbuildactionrunner run buildcompletionnotifyingbuildactionrunner java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor lambda execute rootbuildlifecyclebuildactionexecutor java at org gradle composite internal defaultrootbuildstate run defaultrootbuildstate java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor execute rootbuildlifecyclebuildactionexecutor java at org gradle internal buildtree defaultbuildtreecontext execute defaultbuildtreecontext java at org gradle launcher exec buildtreelifecyclebuildactionexecutor lambda execute buildtreelifecyclebuildactionexecutor java at org gradle internal buildtree buildtreestate run buildtreestate java at org gradle launcher exec buildtreelifecyclebuildactionexecutor execute buildtreelifecyclebuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor execute runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasworkerthreadbuildactionexecutor lambda execute runasworkerthreadbuildactionexecutor java at org gradle internal work defaultworkerleaseservice withlocks defaultworkerleaseservice java at org gradle internal work defaultworkerleaseservice runasworkerthread defaultworkerleaseservice java at org gradle launcher exec runasworkerthreadbuildactionexecutor execute runasworkerthreadbuildactionexecutor java at org gradle tooling internal provider continuous continuousbuildactionexecutor execute continuousbuildactionexecutor java at org gradle tooling internal provider subscribablebuildactionexecutor execute subscribablebuildactionexecutor java at org gradle internal session defaultbuildsessioncontext execute defaultbuildsessioncontext java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle internal session buildsessionstate run buildsessionstate java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle launcher daemon server exec executebuild dobuild executebuild java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec watchfordisconnection execute watchfordisconnection java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec resetdeprecationlogger execute resetdeprecationlogger java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec requeststopifsingleuseddaemon execute requeststopifsingleuseddaemon java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle util internal swapper swap swapper java at org gradle launcher daemon server exec forwardclientinput execute forwardclientinput java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logandcheckhealth execute logandcheckhealth java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logtoclient dobuild logtoclient java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec establishbuildenvironment dobuild establishbuildenvironment java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec startbuildorrespondwithbusy run startbuildorrespondwithbusy java at org gradle launcher daemon server daemonstatecoordinator run daemonstatecoordinator java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java org gradle api internal tasks defaulttaskcontainer taskcreationexception could not create task common spotlessjavascriptapply at org gradle api internal tasks defaulttaskcontainer taskcreationexception defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider domainobjectcreationexception defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider trycreate defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider run defaulttaskcontainer java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider trycreate defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider calculateownvalue defaultnameddomainobjectcollection java at org gradle api internal provider abstractminimalprovider getornull abstractminimalprovider java at org gradle api internal defaultnameddomainobjectcollection findbyname defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer findbyname defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer findbyname defaulttaskcontainer java at org gradle plugins ide internal tooling gradleprojectbuilder tasks gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildhierarchy gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildhierarchy gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildall gradleprojectbuilder java at org gradle plugins ide internal tooling ideamodelbuilder buildall ideamodelbuilder java at org gradle plugins ide internal tooling ideamodelbuilder buildall ideamodelbuilder java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry builderwithnoparameter build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle api internal project defaultprojectstateregistry projectstateimpl lambda frommutablestate defaultprojectstateregistry java at org gradle internal work defaultworkerleaseservice withreplacedlocks defaultworkerleaseservice java at org gradle api internal project defaultprojectstateregistry projectstateimpl frommutablestate defaultprojectstateregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder call defaulttoolingmodelbuilderregistry java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder build defaulttoolingmodelbuilderregistry java at org gradle internal build defaultbuildtoolingmodelcontroller abstracttoolingscope getmodel defaultbuildtoolingmodelcontroller java at org gradle tooling internal provider runner defaultbuildcontroller getmodel defaultbuildcontroller java at org gradle tooling internal consumer connection parameterawarebuildcontrolleradapter getmodel parameterawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller getmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter getmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model classsetimportmodelprovider populatebuildmodels classsetimportmodelprovider java at org jetbrains plugins gradle model projectimportaction addbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org gradle tooling internal consumer connection internalbuildactionadapter execute internalbuildactionadapter java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter runaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle internal buildtree defaultbuildtreemodelcreator frombuildmodel defaultbuildtreemodelcreator java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda runbuild defaultbuildtreelifecyclecontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transition statetransitioncontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller runbuild defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner runclientaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner clientprovidedphasedactionrunner run clientprovidedphasedactionrunner java at org gradle launcher exec chainingbuildactionrunner run chainingbuildactionrunner java at org gradle internal buildtree problemreportingbuildactionrunner run problemreportingbuildactionrunner java at org gradle launcher exec buildoutcomereportingbuildactionrunner run buildoutcomereportingbuildactionrunner java at org gradle tooling internal provider filesystemwatchingbuildactionrunner run filesystemwatchingbuildactionrunner java at org gradle launcher exec buildcompletionnotifyingbuildactionrunner run buildcompletionnotifyingbuildactionrunner java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor lambda execute rootbuildlifecyclebuildactionexecutor java at org gradle composite internal defaultrootbuildstate run defaultrootbuildstate java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor execute rootbuildlifecyclebuildactionexecutor java at org gradle internal buildtree defaultbuildtreecontext execute defaultbuildtreecontext java at org gradle launcher exec buildtreelifecyclebuildactionexecutor lambda execute buildtreelifecyclebuildactionexecutor java at org gradle internal buildtree buildtreestate run buildtreestate java at org gradle launcher exec buildtreelifecyclebuildactionexecutor execute buildtreelifecyclebuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor execute runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasworkerthreadbuildactionexecutor lambda execute runasworkerthreadbuildactionexecutor java at org gradle internal work defaultworkerleaseservice withlocks defaultworkerleaseservice java at org gradle internal work defaultworkerleaseservice runasworkerthread defaultworkerleaseservice java at org gradle launcher exec runasworkerthreadbuildactionexecutor execute runasworkerthreadbuildactionexecutor java at org gradle tooling internal provider continuous continuousbuildactionexecutor execute continuousbuildactionexecutor java at org gradle tooling internal provider subscribablebuildactionexecutor execute subscribablebuildactionexecutor java at org gradle internal session defaultbuildsessioncontext execute defaultbuildsessioncontext java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle internal session buildsessionstate run buildsessionstate java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle launcher daemon server exec executebuild dobuild executebuild java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec watchfordisconnection execute watchfordisconnection java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec resetdeprecationlogger execute resetdeprecationlogger java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec requeststopifsingleuseddaemon execute requeststopifsingleuseddaemon java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle util internal swapper swap swapper java at org gradle launcher daemon server exec forwardclientinput execute forwardclientinput java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logandcheckhealth execute logandcheckhealth java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logtoclient dobuild logtoclient java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec establishbuildenvironment dobuild establishbuildenvironment java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec startbuildorrespondwithbusy run startbuildorrespondwithbusy java at org gradle launcher daemon server daemonstatecoordinator run daemonstatecoordinator java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by org gradle api internal tasks defaulttaskcontainer taskcreationexception could not create task common spotlessjavascript at org gradle api internal tasks defaulttaskcontainer taskcreationexception defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider domainobjectcreationexception defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider trycreate defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider run defaulttaskcontainer java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider trycreate defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider calculateownvalue defaultnameddomainobjectcollection java at org gradle api internal provider abstractminimalprovider calculatevalue abstractminimalprovider java at org gradle api internal provider collectors elementfromprovider collectentries collectors java at org gradle api internal provider collectors typedcollector collectentries collectors java at org gradle api internal provider collectors typedcollector collectinto collectors java at org gradle api internal collections defaultpendingsource realize defaultpendingsource java at org gradle api internal collections defaultpendingsource realizepending defaultpendingsource java at org gradle api internal collections sortedsetelementsource iterator sortedsetelementsource java at org gradle api internal defaultdomainobjectcollection iterator defaultdomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer iterator defaulttaskcontainer java at java base java util abstractcollection addall abstractcollection java at java base java util treeset addall treeset java at java base java util treeset treeset java at org jetbrains plugins gradle tooling builder tasksfactory getalltasks tasksfactory java at org jetbrains plugins gradle tooling builder tasksfactory collecttasks tasksfactory java at org jetbrains plugins gradle tooling builder tasksfactory gettasks tasksfactory java at org jetbrains gradle profiler builder javataskmodelbuilderimpl buildall javataskmodelbuilderimpl java at org jetbrains plugins gradle tooling internal extramodelbuilder buildall extramodelbuilder java at org jetbrains plugins gradle tooling internal extramodelbuilder buildall extramodelbuilder java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry builderwithparameter build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry usercodeassigningbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle configuration internal defaultusercodeapplicationcontext currentapplication reapply defaultusercodeapplicationcontext java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry usercodeassigningbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle api internal project defaultprojectstateregistry projectstateimpl lambda frommutablestate defaultprojectstateregistry java at org gradle internal work defaultworkerleaseservice withreplacedlocks defaultworkerleaseservice java at org gradle api internal project defaultprojectstateregistry projectstateimpl frommutablestate defaultprojectstateregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder call defaulttoolingmodelbuilderregistry java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder build defaulttoolingmodelbuilderregistry java at org gradle internal build defaultbuildtoolingmodelcontroller abstracttoolingscope getmodel defaultbuildtoolingmodelcontroller java at org gradle tooling internal provider runner defaultbuildcontroller getmodel defaultbuildcontroller java at org gradle tooling internal consumer connection parameterawarebuildcontrolleradapter getmodel parameterawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller getmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter getmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model classsetprojectimportmodelprovider populateprojectmodels classsetprojectimportmodelprovider java at org jetbrains plugins gradle model projectimportaction getprojectmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction access projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction fetchprojectbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java more caused by java lang illegalstateexception can t automatically determine npm executable and none was specifically supplied hedera network other version snapshot operating system macos
1
10,201
13,065,958,398
IssuesEvent
2020-07-30 20:44:00
keep-network/keep-ecdsa
https://api.github.com/repos/keep-network/keep-ecdsa
closed
Allow the application to configure expectations about the minimum bond
process & client team ⛓chain
Right now, bonded sortition pool updates the minimum bondable value for each successful signer set selection. This way, the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application's expectations. Instead of doing it automatically, we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane, balancing signer selection cost, to the application. For tBTC, this value could be the minimum lot size.
1.0
Allow the application to configure expectations about the minimum bond - Right now, bonded sortition pool updates the minimum bondable value for each successful signer set selection. This way, the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application's expectations. Instead of doing it automatically, we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane, balancing signer selection cost, to the application. For tBTC, this value could be the minimum lot size.
process
allow the application to configure expectations about the minimum bond right now bonded sortition pool updates the minimum bondable value for each successful signer set selection this way the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application s expectations instead of doing it automatically we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane balancing signer selection cost to the application for tbtc this value could be the minimum lot size
1
416,437
12,146,458,777
IssuesEvent
2020-04-24 11:11:37
PoltyPoltou/CACAO2020
https://api.github.com/repos/PoltyPoltou/CACAO2020
opened
Implémenter IActeur
High Priority
Modification obligatoire : getIndicateurs() getParametres() getJournaux() getDescription() next() getNomsFilieresProposees() getFiliere(string) Modification possible : initialiser()
1.0
Implémenter IActeur - Modification obligatoire : getIndicateurs() getParametres() getJournaux() getDescription() next() getNomsFilieresProposees() getFiliere(string) Modification possible : initialiser()
non_process
implémenter iacteur modification obligatoire getindicateurs getparametres getjournaux getdescription next getnomsfilieresproposees getfiliere string modification possible initialiser
0
35,466
12,339,815,099
IssuesEvent
2020-05-14 18:47:38
wrbejar/JavaVulnerableLabIcatu
https://api.github.com/repos/wrbejar/JavaVulnerableLabIcatu
opened
CVE-2015-6420 (High) detected in commons-collections-3.2.1.jar
security vulnerability
## CVE-2015-6420 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to vulnerable library: /JavaVulnerableLabIcatu/target/JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar,/root/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wrbejar/JavaVulnerableLabIcatu/commit/68aef85b943fe5994864ad4b1984daaa93ad6530">68aef85b943fe5994864ad4b1984daaa93ad6530</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2015-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420>CVE-2015-6420</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420</a></p> <p>Release Date: 2015-12-15</p> <p>Fix Resolution: org.apache.commons:commons-collections4:4.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","isTransitiveDependency":false,"dependencyTree":"commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-collections4:4.1"}],"vulnerabilityIdentifier":"CVE-2015-6420","vulnerabilityDetails":"Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
True
CVE-2015-6420 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-6420 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to vulnerable library: /JavaVulnerableLabIcatu/target/JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar,/root/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wrbejar/JavaVulnerableLabIcatu/commit/68aef85b943fe5994864ad4b1984daaa93ad6530">68aef85b943fe5994864ad4b1984daaa93ad6530</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2015-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420>CVE-2015-6420</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420</a></p> <p>Release Date: 2015-12-15</p> <p>Fix Resolution: org.apache.commons:commons-collections4:4.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","isTransitiveDependency":false,"dependencyTree":"commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-collections4:4.1"}],"vulnerabilityIdentifier":"CVE-2015-6420","vulnerabilityDetails":"Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
non_process
cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to vulnerable library javavulnerablelabicatu target javavulnerablelab web inf lib commons collections jar root repository commons collections commons collections commons collections jar dependency hierarchy x commons collections jar vulnerable library found in head commit a href vulnerability details serialized object interfaces in certain cisco collaboration and social media endpoint clients and client software network application service and acceleration network and content security devices network management and provisioning routing and switching enterprise and service provider unified computing voice and unified communications devices video streaming telepresence and transcoding devices wireless and cisco hosted services products allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution org apache commons commons rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails serialized object interfaces in certain cisco collaboration and social media endpoint clients and client software network application service and acceleration network and content security devices network management and provisioning routing and switching enterprise and service provider unified computing voice and unified communications devices video streaming telepresence and transcoding devices wireless and cisco hosted services products allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library vulnerabilityurl
0
16,250
20,799,284,015
IssuesEvent
2022-03-17 12:26:26
streamnative/flink
https://api.github.com/repos/streamnative/flink
opened
[FLINK-26445] Fix When users specify only consume from 1 partition but the source consumes from all partitions.
compute/data-processing type/bug
https://issues.apache.org/jira/browse/FLINK-26645
1.0
[FLINK-26445] Fix When users specify only consume from 1 partition but the source consumes from all partitions. - https://issues.apache.org/jira/browse/FLINK-26645
process
fix when users specify only consume from partition but the source consumes from all partitions
1
34,783
6,376,761,091
IssuesEvent
2017-08-02 08:22:45
apache/bookkeeper
https://api.github.com/repos/apache/bookkeeper
opened
[documentation] weight-based placement policy
area/documentation area/release type/task
- general introduction of placement policies - weight-based placement policy
1.0
[documentation] weight-based placement policy - - general introduction of placement policies - weight-based placement policy
non_process
weight based placement policy general introduction of placement policies weight based placement policy
0
19,335
25,472,614,005
IssuesEvent
2022-11-25 11:32:06
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP] [PM] Organizational admin is able to login with invalid password
Bug P1 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
**Pre-condition:** IDP and MFA should be enabled in the PM **Steps:** 1. Add organizational user in the participant manager 2. Complete set up your account process 3. Try to sign in with an invalid password and Verify **AR:** Organizational admin is able to sign in with invalid password **ER:** Organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
3.0
[IDP] [PM] Organizational admin is able to login with invalid password - **Pre-condition:** IDP and MFA should be enabled in the PM **Steps:** 1. Add organizational user in the participant manager 2. Complete set up your account process 3. Try to sign in with an invalid password and Verify **AR:** Organizational admin is able to sign in with invalid password **ER:** Organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
process
organizational admin is able to login with invalid password pre condition idp and mfa should be enabled in the pm steps add organizational user in the participant manager complete set up your account process try to sign in with an invalid password and verify ar organizational admin is able to sign in with invalid password er organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
1
106,869
16,735,432,893
IssuesEvent
2021-06-11 00:08:39
AlexRogalskiy/qiitos
https://api.github.com/repos/AlexRogalskiy/qiitos
closed
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: qiitos/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: qiitos/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/qiitos/commit/9ea0c7ba4083c92f216516728105e2b17415771e">9ea0c7ba4083c92f216516728105e2b17415771e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: qiitos/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: qiitos/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/qiitos/commit/9ea0c7ba4083c92f216516728105e2b17415771e">9ea0c7ba4083c92f216516728105e2b17415771e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file qiitos node modules redeyed examples browser index html path to vulnerable library qiitos node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
184,141
14,272,717,105
IssuesEvent
2020-11-21 18:14:48
cybersemics/em
https://api.github.com/repos/cybersemics/em
opened
Upgrade to react-scripts v4
test
The [pullQueue tests](https://github.com/cybersemics/em/blob/dev/src/redux-middleware/__tests__/pullQueue.ts) fail on `react-scripts` v4. It is related to the fake timer and mock debounce logic, but I can't figure out how to fix it. We should be able to use `useFakeTimers('modern')` now and not have to mock lodash's debounce, but that causes an error. Here are the CHANGELOGS although I didn't find anything relevant: - create-react-app CHANGELOG: https://github.com/facebook/create-react-app/blob/master/CHANGELOG.md - jest CHANGELOG: https://github.com/facebook/jest/blob/master/CHANGELOG.md
1.0
Upgrade to react-scripts v4 - The [pullQueue tests](https://github.com/cybersemics/em/blob/dev/src/redux-middleware/__tests__/pullQueue.ts) fail on `react-scripts` v4. It is related to the fake timer and mock debounce logic, but I can't figure out how to fix it. We should be able to use `useFakeTimers('modern')` now and not have to mock lodash's debounce, but that causes an error. Here are the CHANGELOGS although I didn't find anything relevant: - create-react-app CHANGELOG: https://github.com/facebook/create-react-app/blob/master/CHANGELOG.md - jest CHANGELOG: https://github.com/facebook/jest/blob/master/CHANGELOG.md
non_process
upgrade to react scripts the fail on react scripts it is related to the fake timer and mock debounce logic but i can t figure out how to fix it we should be able to use usefaketimers modern now and not have to mock lodash s debounce but that causes an error here are the changelogs although i didn t find anything relevant create react app changelog jest changelog
0
185,914
14,384,565,803
IssuesEvent
2020-12-02 10:37:57
hashicorp/nomad
https://api.github.com/repos/hashicorp/nomad
opened
e2e: flakey TestMetricsLinux test
theme/flaky-tests type/bug
The [Linux alloc metrics e2e test](https://github.com/hashicorp/nomad/blob/5fbcf82ceefe5facacaff31f45e04ccd89fed3ac/e2e/metrics/metrics.go#L89) occasinally fails with the following error: ``` prometheus query failed (nomad_client_allocs_memory_usage{exported_job="memstress"}): expected non-zero metrics, got: nomad_client_allocs_memory_usage{alloc_id="70ed194c-e5e5-099b-d000-17cc0ad6920e", exported_job="memstress", instance="172.31.13.159:4646", job="nomad_metrics", namespace="default", task="memstress", task_group="memstress"} => 0 @[1606903456.227] ```
1.0
e2e: flakey TestMetricsLinux test - The [Linux alloc metrics e2e test](https://github.com/hashicorp/nomad/blob/5fbcf82ceefe5facacaff31f45e04ccd89fed3ac/e2e/metrics/metrics.go#L89) occasinally fails with the following error: ``` prometheus query failed (nomad_client_allocs_memory_usage{exported_job="memstress"}): expected non-zero metrics, got: nomad_client_allocs_memory_usage{alloc_id="70ed194c-e5e5-099b-d000-17cc0ad6920e", exported_job="memstress", instance="172.31.13.159:4646", job="nomad_metrics", namespace="default", task="memstress", task_group="memstress"} => 0 @[1606903456.227] ```
non_process
flakey testmetricslinux test the occasinally fails with the following error prometheus query failed nomad client allocs memory usage exported job memstress expected non zero metrics got nomad client allocs memory usage alloc id exported job memstress instance job nomad metrics namespace default task memstress task group memstress
0
146,474
19,404,088,479
IssuesEvent
2021-12-19 17:49:57
vincenzodistasio97/BenchmarkJava
https://api.github.com/repos/vincenzodistasio97/BenchmarkJava
opened
CVE-2021-22096 (Medium) detected in multiple libraries
security vulnerability
## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-core-4.3.30.RELEASE.jar</b>, <b>spring-webmvc-4.3.30.RELEASE.jar</b>, <b>spring-web-4.3.30.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-core-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.30.RELEASE/spring-core-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-context-4.3.30.RELEASE.jar (Root Library) - :x: **spring-core-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-webmvc-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: ository/org/springframework/spring-webmvc/4.3.30.RELEASE/spring-webmvc-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-webmvc-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-web-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: ository/org/springframework/spring-web/4.3.30.RELEASE/spring-web-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/BenchmarkJava/commit/0f7ddc18e36494a62123563264c65e3344dae7af">0f7ddc18e36494a62123563264c65e3344dae7af</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22096 (Medium) detected in multiple libraries - ## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-core-4.3.30.RELEASE.jar</b>, <b>spring-webmvc-4.3.30.RELEASE.jar</b>, <b>spring-web-4.3.30.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-core-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.30.RELEASE/spring-core-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-context-4.3.30.RELEASE.jar (Root Library) - :x: **spring-core-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-webmvc-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: ository/org/springframework/spring-webmvc/4.3.30.RELEASE/spring-webmvc-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-webmvc-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-web-4.3.30.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: BenchmarkJava/pom.xml</p> <p>Path to vulnerable library: ository/org/springframework/spring-web/4.3.30.RELEASE/spring-web-4.3.30.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-4.3.30.RELEASE.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/BenchmarkJava/commit/0f7ddc18e36494a62123563264c65e3344dae7af">0f7ddc18e36494a62123563264c65e3344dae7af</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries spring core release jar spring webmvc release jar spring web release jar spring core release jar spring core library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library home wss scanner repository org springframework spring core release spring core release jar dependency hierarchy spring context release jar root library x spring core release jar vulnerable library spring webmvc release jar spring web mvc library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library ository org springframework spring webmvc release spring webmvc release jar dependency hierarchy x spring webmvc release jar vulnerable library spring web release jar spring web library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library ository org springframework spring web release spring web release jar dependency hierarchy x spring web release jar vulnerable library found in head commit a href found in base branch master vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release org springframework spring web release org springframework spring webmvc release org springframework spring webflux release step up your open source security game with whitesource
0
23,645
22,441,322,602
IssuesEvent
2022-06-21 01:39:16
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
opened
Add a method to get a `ComponentId` from a `Component`
A-ECS C-Usability
## What problem does this solve or what need does it fill? Being able to extract the `ComponentId` for a given component of type `C` is surprisingly unintuitive. ## What solution would you like? Create a public method on `Component` that mirrors [`Bundle::component_ids()`](https://dev-docs.bevyengine.org/bevy/ecs/bundle/trait.Bundle.html#tymethod.component_ids). ## What alternative(s) have you considered? 1. Use `world.init_component::<C>()`. This is unintuitive, but basically works. 2. Convert the component type into a bundle, and call `(C, )::component_ids().iter().next().unwrap()`. Which is more than a little cursed. ## Additional context Discovered while working on #1481 with @plof27.
True
Add a method to get a `ComponentId` from a `Component` - ## What problem does this solve or what need does it fill? Being able to extract the `ComponentId` for a given component of type `C` is surprisingly unintuitive. ## What solution would you like? Create a public method on `Component` that mirrors [`Bundle::component_ids()`](https://dev-docs.bevyengine.org/bevy/ecs/bundle/trait.Bundle.html#tymethod.component_ids). ## What alternative(s) have you considered? 1. Use `world.init_component::<C>()`. This is unintuitive, but basically works. 2. Convert the component type into a bundle, and call `(C, )::component_ids().iter().next().unwrap()`. Which is more than a little cursed. ## Additional context Discovered while working on #1481 with @plof27.
non_process
add a method to get a componentid from a component what problem does this solve or what need does it fill being able to extract the componentid for a given component of type c is surprisingly unintuitive what solution would you like create a public method on component that mirrors what alternative s have you considered use world init component this is unintuitive but basically works convert the component type into a bundle and call c component ids iter next unwrap which is more than a little cursed additional context discovered while working on with
0
37,298
8,343,517,833
IssuesEvent
2018-09-30 05:44:03
liuxuewei/bluebee-accounting-system
https://api.github.com/repos/liuxuewei/bluebee-accounting-system
closed
反馈
Priority-Medium Type-Defect auto-migrated
``` 软件还是比较好的,试用了一下午,发现了不上问题,有时�� �会响应比较慢,添加的商品要重启软件之后才能出来,进货� ��保存无返回值,继续点保存会保存多条数据,总之还是有不 少bug的,希望开发者再接再励!! ``` Original issue reported on code.google.com by `wsjta...@gmail.com` on 18 Aug 2013 at 5:06
1.0
反馈 - ``` 软件还是比较好的,试用了一下午,发现了不上问题,有时�� �会响应比较慢,添加的商品要重启软件之后才能出来,进货� ��保存无返回值,继续点保存会保存多条数据,总之还是有不 少bug的,希望开发者再接再励!! ``` Original issue reported on code.google.com by `wsjta...@gmail.com` on 18 Aug 2013 at 5:06
non_process
反馈 软件还是比较好的,试用了一下午,发现了不上问题,有时�� �会响应比较慢,添加的商品要重启软件之后才能出来,进货� ��保存无返回值,继续点保存会保存多条数据,总之还是有不 少bug的,希望开发者再接再励!! original issue reported on code google com by wsjta gmail com on aug at
0
779,638
27,360,959,923
IssuesEvent
2023-02-27 15:50:29
rangav/thunder-client-support
https://api.github.com/repos/rangav/thunder-client-support
closed
Ability to Format Graphql Body
feature request Priority
Could you add the feature to format a Graphql body code snippet, please. That featured is available for JSON and XML.
1.0
Ability to Format Graphql Body - Could you add the feature to format a Graphql body code snippet, please. That featured is available for JSON and XML.
non_process
ability to format graphql body could you add the feature to format a graphql body code snippet please that featured is available for json and xml
0
8,770
11,886,457,510
IssuesEvent
2020-03-27 21:59:20
nodejs/node
https://api.github.com/repos/nodejs/node
opened
child_process.exec/execFile docs have some inconsistencies and inaccuracies
child_process doc
I don't have time to fix this now, but while looking at sec issues related to these APIs, I found some oddities. https://nodejs.org/api/child_process.html#child_process_child_process_execfile_file_args_options_callback says > shell <boolean> | <string> If true, runs command inside of a shell. But execFile() doesn't have a `command` argument... this was pasted from exec(), it seems. Probably what happens is that if there is a shell, then `file` and `args` are all concatenated together, `' '` seperated, and passed to the shell. Since the shell option AFAICT ends up following the same path as from exec(), it suggests that the exec docs: > shell <string> Shell to execute the command with. See Shell Requirements and Default Windows Shell. Default: '/bin/sh' on Unix, process.env.ComSpec on Windows. are incomplete, probably `false` would work just fine as an arg there, making exec() behave exactly like execFile(). This seems to be a bit legacy as well: > The child_process.execFile() function is similar to child_process.exec() except that it does not spawn a shell by default. Rather, the specified executable file is spawned directly as a new process making it slightly more efficient than child_process.exec(). Now that both exec and execFile() have a shell, differing only be the default value, its probably more accurate to say the difference is that one takes an array of strings as an argument `execFile(file, argv, ..` and the other takes a single string `exec(command, ...)`. The text following is now wrong: > The same options as child_process.exec() are supported. Since a shell is not spawned, behaviors such as I/O redirection and file globbing are not supported. It can't both *support* the same options, and *not support* some of the options. It should probably say "If a shell is ..." (only one word different, but its important). exec should probably have docs saying the same thing, shell behaviours are not supported when shell is `false`. And execFile() should probably include the warnings from exec about how shell special chars vary by platform. Some of these issues are shared with the "sync" versions of the APIs.
1.0
child_process.exec/execFile docs have some inconsistencies and inaccuracies - I don't have time to fix this now, but while looking at sec issues related to these APIs, I found some oddities. https://nodejs.org/api/child_process.html#child_process_child_process_execfile_file_args_options_callback says > shell <boolean> | <string> If true, runs command inside of a shell. But execFile() doesn't have a `command` argument... this was pasted from exec(), it seems. Probably what happens is that if there is a shell, then `file` and `args` are all concatenated together, `' '` seperated, and passed to the shell. Since the shell option AFAICT ends up following the same path as from exec(), it suggests that the exec docs: > shell <string> Shell to execute the command with. See Shell Requirements and Default Windows Shell. Default: '/bin/sh' on Unix, process.env.ComSpec on Windows. are incomplete, probably `false` would work just fine as an arg there, making exec() behave exactly like execFile(). This seems to be a bit legacy as well: > The child_process.execFile() function is similar to child_process.exec() except that it does not spawn a shell by default. Rather, the specified executable file is spawned directly as a new process making it slightly more efficient than child_process.exec(). Now that both exec and execFile() have a shell, differing only be the default value, its probably more accurate to say the difference is that one takes an array of strings as an argument `execFile(file, argv, ..` and the other takes a single string `exec(command, ...)`. The text following is now wrong: > The same options as child_process.exec() are supported. Since a shell is not spawned, behaviors such as I/O redirection and file globbing are not supported. It can't both *support* the same options, and *not support* some of the options. It should probably say "If a shell is ..." (only one word different, but its important). exec should probably have docs saying the same thing, shell behaviours are not supported when shell is `false`. And execFile() should probably include the warnings from exec about how shell special chars vary by platform. Some of these issues are shared with the "sync" versions of the APIs.
process
child process exec execfile docs have some inconsistencies and inaccuracies i don t have time to fix this now but while looking at sec issues related to these apis i found some oddities says shell if true runs command inside of a shell but execfile doesn t have a command argument this was pasted from exec it seems probably what happens is that if there is a shell then file and args are all concatenated together seperated and passed to the shell since the shell option afaict ends up following the same path as from exec it suggests that the exec docs shell shell to execute the command with see shell requirements and default windows shell default bin sh on unix process env comspec on windows are incomplete probably false would work just fine as an arg there making exec behave exactly like execfile this seems to be a bit legacy as well the child process execfile function is similar to child process exec except that it does not spawn a shell by default rather the specified executable file is spawned directly as a new process making it slightly more efficient than child process exec now that both exec and execfile have a shell differing only be the default value its probably more accurate to say the difference is that one takes an array of strings as an argument execfile file argv and the other takes a single string exec command the text following is now wrong the same options as child process exec are supported since a shell is not spawned behaviors such as i o redirection and file globbing are not supported it can t both support the same options and not support some of the options it should probably say if a shell is only one word different but its important exec should probably have docs saying the same thing shell behaviours are not supported when shell is false and execfile should probably include the warnings from exec about how shell special chars vary by platform some of these issues are shared with the sync versions of the apis
1
15,437
19,651,448,236
IssuesEvent
2022-01-10 07:44:59
linuxdeepin/developer-center
https://api.github.com/repos/linuxdeepin/developer-center
closed
Terminal blur is displaced with deepin-kwin
bug | functional behavior other | delay processing
Bug: with deepin-kwin, the blur on the terminal seems to be displaced. ![DeepinScreenshot_dde-desktop_20200322083648](https://user-images.githubusercontent.com/62511778/77253559-5045fb80-6c18-11ea-9db5-ddeb3ce8c199.png) Update: issue is not present when the terminal is fullscreen.
1.0
Terminal blur is displaced with deepin-kwin - Bug: with deepin-kwin, the blur on the terminal seems to be displaced. ![DeepinScreenshot_dde-desktop_20200322083648](https://user-images.githubusercontent.com/62511778/77253559-5045fb80-6c18-11ea-9db5-ddeb3ce8c199.png) Update: issue is not present when the terminal is fullscreen.
process
terminal blur is displaced with deepin kwin bug with deepin kwin the blur on the terminal seems to be displaced update issue is not present when the terminal is fullscreen
1
472,778
13,631,234,447
IssuesEvent
2020-09-24 17:40:56
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[0.9.0.2 beta develop-62] Government Bank permissions further refinement
Priority: Medium Type: Exploit
When a player is missing a title with permissions "Has Government Banking privileges", it should disable government tagged account access regardless of the players manager/use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an "offshore tax free" account.
1.0
[0.9.0.2 beta develop-62] Government Bank permissions further refinement - When a player is missing a title with permissions "Has Government Banking privileges", it should disable government tagged account access regardless of the players manager/use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an "offshore tax free" account.
non_process
government bank permissions further refinement when a player is missing a title with permissions has government banking privileges it should disable government tagged account access regardless of the players manager use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an offshore tax free account
0
197,776
6,963,758,753
IssuesEvent
2017-12-08 18:43:48
elementary/gala
https://api.github.com/repos/elementary/gala
opened
Send a persistent notification after taking a screenshot
Priority: Wishlist
This is a migration of https://github.com/elementary/screenshot-tool/issues/30 After taking a screenshot, Gala should send a persistent notification using the screenshot preview as the notification image. Clicking the notification should open the image in the default image viewer. This would give: 1. A visual queue that the screenshot was completed, useful when sound is disabled or the user is not able to hear 2. A way to quickly view/edit/share the screenshot without having to know where screenshots are stored on the filesystem
1.0
Send a persistent notification after taking a screenshot - This is a migration of https://github.com/elementary/screenshot-tool/issues/30 After taking a screenshot, Gala should send a persistent notification using the screenshot preview as the notification image. Clicking the notification should open the image in the default image viewer. This would give: 1. A visual queue that the screenshot was completed, useful when sound is disabled or the user is not able to hear 2. A way to quickly view/edit/share the screenshot without having to know where screenshots are stored on the filesystem
non_process
send a persistent notification after taking a screenshot this is a migration of after taking a screenshot gala should send a persistent notification using the screenshot preview as the notification image clicking the notification should open the image in the default image viewer this would give a visual queue that the screenshot was completed useful when sound is disabled or the user is not able to hear a way to quickly view edit share the screenshot without having to know where screenshots are stored on the filesystem
0
7,470
10,566,758,795
IssuesEvent
2019-10-05 21:13:21
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
opened
Decide if macOS support is still feasible/worthwhile with Apples new notarization requirements
area: software process platform: macos type: discussion type: help wanted
[Reddit discussion](https://www.reddit.com/r/programming/comments/ddbxmw/sorry_macos_users_but_apple_has_gone_too_far_for/)
1.0
Decide if macOS support is still feasible/worthwhile with Apples new notarization requirements - [Reddit discussion](https://www.reddit.com/r/programming/comments/ddbxmw/sorry_macos_users_but_apple_has_gone_too_far_for/)
process
decide if macos support is still feasible worthwhile with apples new notarization requirements
1
450,822
13,019,845,030
IssuesEvent
2020-07-27 00:52:32
momentum-mod/game
https://api.github.com/repos/momentum-mod/game
closed
Stickybomb particles depend on TF2
Priority: High Size: Small Type: Enhancement
**What feature is your improvement idea related to? Please describe.** Currently, the stickybomb particle depends on the TF2 material `effects/circle2`. Not having TF2 installed when using the sticky launcher results in this error: `C_OP_RenderSprites::RenderUnsorted: Attempting to use an unimplemented sprite renderer for system "mom_stickypulse_blue"!` **Describe the solution you'd like** Either add the material from TF2 or replace it with another that we do have. **Additional context** Our base Momentum assets shouldn't depend on any other game being installed. In the future when adding assets should make sure that all materials used are included in the game.
1.0
Stickybomb particles depend on TF2 - **What feature is your improvement idea related to? Please describe.** Currently, the stickybomb particle depends on the TF2 material `effects/circle2`. Not having TF2 installed when using the sticky launcher results in this error: `C_OP_RenderSprites::RenderUnsorted: Attempting to use an unimplemented sprite renderer for system "mom_stickypulse_blue"!` **Describe the solution you'd like** Either add the material from TF2 or replace it with another that we do have. **Additional context** Our base Momentum assets shouldn't depend on any other game being installed. In the future when adding assets should make sure that all materials used are included in the game.
non_process
stickybomb particles depend on what feature is your improvement idea related to please describe currently the stickybomb particle depends on the material effects not having installed when using the sticky launcher results in this error c op rendersprites renderunsorted attempting to use an unimplemented sprite renderer for system mom stickypulse blue describe the solution you d like either add the material from or replace it with another that we do have additional context our base momentum assets shouldn t depend on any other game being installed in the future when adding assets should make sure that all materials used are included in the game
0
52,399
13,224,714,621
IssuesEvent
2020-08-17 19:41:52
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
segments in wimpsimreader.py (Trac #2161)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2161">https://code.icecube.wisc.edu/projects/icecube/ticket/2161</a>, reported by grenziand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "_ts": "1550067323910946", "description": "Both segments at [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/python/wimpsimreader.py] are supposed to launch I3WimpSimReader module, but they are set with non acceptable values of variable InjectionRadius. The real default is NaN, and not 0 as visible here [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/private/wimpsim-reader/I3WimpSimReader.cxx]. The variable has this function (from docs):\n\n''InjectionRadius [Default=NAN] If >0, events will be injected in cylinder with specified radius and [zmin, zmax] height instead of rectangular box''\n\nIf the original intention was to avoid the cylinder, this value should be left to NaN, as default. (Or the value 0 should be accepted in I3 module at lines 180-181)\n\nFor WimpSimReaderEarth there is an issue with GCD, the code isn't able to read time in given format:\n\n{{{No registered converter was able to produce a C++ rvalue of type std::string from this Python object of type I3Time}}}\n\nTo avoid it, time should be taken as julian day (or another \"good\" format), so for example\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time}}}\n\ncould be:\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time.mod_julian_day_double}}}", "reporter": "grenzi", "cc": "", "resolution": "fixed", "time": "2018-06-12T10:09:52", "component": "combo simulation", "summary": "segments in wimpsimreader.py", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
segments in wimpsimreader.py (Trac #2161) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2161">https://code.icecube.wisc.edu/projects/icecube/ticket/2161</a>, reported by grenziand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "_ts": "1550067323910946", "description": "Both segments at [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/python/wimpsimreader.py] are supposed to launch I3WimpSimReader module, but they are set with non acceptable values of variable InjectionRadius. The real default is NaN, and not 0 as visible here [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/private/wimpsim-reader/I3WimpSimReader.cxx]. The variable has this function (from docs):\n\n''InjectionRadius [Default=NAN] If >0, events will be injected in cylinder with specified radius and [zmin, zmax] height instead of rectangular box''\n\nIf the original intention was to avoid the cylinder, this value should be left to NaN, as default. (Or the value 0 should be accepted in I3 module at lines 180-181)\n\nFor WimpSimReaderEarth there is an issue with GCD, the code isn't able to read time in given format:\n\n{{{No registered converter was able to produce a C++ rvalue of type std::string from this Python object of type I3Time}}}\n\nTo avoid it, time should be taken as julian day (or another \"good\" format), so for example\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time}}}\n\ncould be:\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time.mod_julian_day_double}}}", "reporter": "grenzi", "cc": "", "resolution": "fixed", "time": "2018-06-12T10:09:52", "component": "combo simulation", "summary": "segments in wimpsimreader.py", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
non_process
segments in wimpsimreader py trac migrated from json status closed changetime ts description both segments at are supposed to launch module but they are set with non acceptable values of variable injectionradius the real default is nan and not as visible here the variable has this function from docs n n injectionradius if events will be injected in cylinder with specified radius and height instead of rectangular box n nif the original intention was to avoid the cylinder this value should be left to nan as default or the value should be accepted in module at lines n nfor wimpsimreaderearth there is an issue with gcd the code isn t able to read time in given format n n no registered converter was able to produce a c rvalue of type std string from this python object of type n nto avoid it time should be taken as julian day or another good format so for example n n line return frame get start time n ncould be n n line return frame get start time mod julian day double reporter grenzi cc resolution fixed time component combo simulation summary segments in wimpsimreader py priority normal keywords milestone owner nega type defect
0
142,757
5,477,062,516
IssuesEvent
2017-03-12 03:27:36
NCEAS/eml
https://api.github.com/repos/NCEAS/eml
closed
DataManager cuts last row off data_table
Category: datamanager Component: Bugzilla-Id Priority: Normal Status: Resolved Tracker: Bug
--- Author Name: **Chad Burt** (Chad Burt) Original Redmine Issue: 2775, https://projects.ecoinformatics.org/ecoinfo/issues/2775 Original Date: 2007-02-13 Original Assignee: ben leinfelder --- Had a problem putting this data into a database: http://data.piscoweb.org/catalog/metacat?action=read&docid=cburt.5.2&qformat=pisco It looks like the last row, (2006-02-01,3,10,"test fixtures are fun") is getting cut off. I haven't tested a dataset without a header file. I suspect either the header is being counted against it in a loop or there is a < where a <= would be more appropriate.
1.0
DataManager cuts last row off data_table - --- Author Name: **Chad Burt** (Chad Burt) Original Redmine Issue: 2775, https://projects.ecoinformatics.org/ecoinfo/issues/2775 Original Date: 2007-02-13 Original Assignee: ben leinfelder --- Had a problem putting this data into a database: http://data.piscoweb.org/catalog/metacat?action=read&docid=cburt.5.2&qformat=pisco It looks like the last row, (2006-02-01,3,10,"test fixtures are fun") is getting cut off. I haven't tested a dataset without a header file. I suspect either the header is being counted against it in a loop or there is a < where a <= would be more appropriate.
non_process
datamanager cuts last row off data table author name chad burt chad burt original redmine issue original date original assignee ben leinfelder had a problem putting this data into a database it looks like the last row test fixtures are fun is getting cut off i haven t tested a dataset without a header file i suspect either the header is being counted against it in a loop or there is a where a would be more appropriate
0
174,897
6,544,245,362
IssuesEvent
2017-09-03 13:45:10
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Minor issue : compiling with doc2 throws Error: ambiguous call
Easy Low Priority
Simple example: ```nim # test.nim import os ## Documentation here echo("User home dir " , getHomeDir()) ``` Compile with nim doc2 test ```nim test.nim(3, 36) Error: ambiguous call; both ospaths.getHomeDir() and os.getHomeDir() match for: () ``` The error seems to have no effect and the file test.html is created fine.
1.0
Minor issue : compiling with doc2 throws Error: ambiguous call - Simple example: ```nim # test.nim import os ## Documentation here echo("User home dir " , getHomeDir()) ``` Compile with nim doc2 test ```nim test.nim(3, 36) Error: ambiguous call; both ospaths.getHomeDir() and os.getHomeDir() match for: () ``` The error seems to have no effect and the file test.html is created fine.
non_process
minor issue compiling with throws error ambiguous call simple example nim test nim import os documentation here echo user home dir gethomedir compile with nim test nim test nim error ambiguous call both ospaths gethomedir and os gethomedir match for the error seems to have no effect and the file test html is created fine
0
50,316
6,354,218,497
IssuesEvent
2017-07-29 07:13:20
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
C# Pure Function Keyword to Mark No Side Effects Or External Dependencies
Area-Language Design Discussion
Be able to mark a method as having no side effects or external dependencies, ie: it does not change any state outside the inputs or outputs. Any code that attempts to do this would throw an exception. My thought was that the keyword could be "functional", "pure" (as in a "pure functions" mentioned in some Msdn documentation ), "purefunction", or even "nosideffects". See https://msdn.microsoft.com/en-us/library/bb669139.aspx for some current naming conventions and reasons for this feature.
1.0
C# Pure Function Keyword to Mark No Side Effects Or External Dependencies - Be able to mark a method as having no side effects or external dependencies, ie: it does not change any state outside the inputs or outputs. Any code that attempts to do this would throw an exception. My thought was that the keyword could be "functional", "pure" (as in a "pure functions" mentioned in some Msdn documentation ), "purefunction", or even "nosideffects". See https://msdn.microsoft.com/en-us/library/bb669139.aspx for some current naming conventions and reasons for this feature.
non_process
c pure function keyword to mark no side effects or external dependencies be able to mark a method as having no side effects or external dependencies ie it does not change any state outside the inputs or outputs any code that attempts to do this would throw an exception my thought was that the keyword could be functional pure as in a pure functions mentioned in some msdn documentation purefunction or even nosideffects see for some current naming conventions and reasons for this feature
0
61,698
7,495,240,600
IssuesEvent
2018-04-07 18:41:36
NYU-Shopcarts/shopcarts
https://api.github.com/repos/NYU-Shopcarts/shopcarts
closed
Create the design for all CRUD routes
design
**As a** developler **I need** to know the what to name the API routes **So that** I can create a standardized API **Assumptions:** * We can create and publish a documentation for the Shopcarts API **Acceptance Criteria:** ``` Given that I'm a user of the Shopcarts API When my client calls the CRUD routes on the Shopcarts API Then my products gets created, updated, deleted, or fetched from the Shopcarts DB ```
1.0
Create the design for all CRUD routes - **As a** developler **I need** to know the what to name the API routes **So that** I can create a standardized API **Assumptions:** * We can create and publish a documentation for the Shopcarts API **Acceptance Criteria:** ``` Given that I'm a user of the Shopcarts API When my client calls the CRUD routes on the Shopcarts API Then my products gets created, updated, deleted, or fetched from the Shopcarts DB ```
non_process
create the design for all crud routes as a developler i need to know the what to name the api routes so that i can create a standardized api assumptions we can create and publish a documentation for the shopcarts api acceptance criteria given that i m a user of the shopcarts api when my client calls the crud routes on the shopcarts api then my products gets created updated deleted or fetched from the shopcarts db
0
12,553
14,976,935,706
IssuesEvent
2021-01-28 08:48:52
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
failed to deploy taiga
process_wontfix
Failed to initialize peertube, please contact support with this information: Domain: sashaastiadi-peertube-sasha11.webg2test.grid.tf VDC Name: vdc_vdctest4_sashaastiadi Farm name: freefarm Reason: Couldn't find resources in the cluster for the solution . Use the refresh button on the upper right to restart Peertube creation
1.0
failed to deploy taiga - Failed to initialize peertube, please contact support with this information: Domain: sashaastiadi-peertube-sasha11.webg2test.grid.tf VDC Name: vdc_vdctest4_sashaastiadi Farm name: freefarm Reason: Couldn't find resources in the cluster for the solution . Use the refresh button on the upper right to restart Peertube creation
process
failed to deploy taiga failed to initialize peertube please contact support with this information domain sashaastiadi peertube grid tf vdc name vdc sashaastiadi farm name freefarm reason couldn t find resources in the cluster for the solution use the refresh button on the upper right to restart peertube creation
1
17,347
23,171,755,230
IssuesEvent
2022-07-30 20:40:47
MartinBruun/P6
https://api.github.com/repos/MartinBruun/P6
closed
(CI) Setup self-hosted github runner, migrating to private repo
Need PR 2: Must have Need Tests Process
**Who wants it?** Product Owner: Martin BM **What is the proposed change?** Setup a self-hosted runner (maybe on a Raspberry Pi) to run the github actions, so the project can be made possible to migrate to a private repo, in case this will be a business idea. Follow these guidelines: https://docs.github.com/en/actions/hosting-your-own-runners/about-self-hosted-runners
1.0
(CI) Setup self-hosted github runner, migrating to private repo - **Who wants it?** Product Owner: Martin BM **What is the proposed change?** Setup a self-hosted runner (maybe on a Raspberry Pi) to run the github actions, so the project can be made possible to migrate to a private repo, in case this will be a business idea. Follow these guidelines: https://docs.github.com/en/actions/hosting-your-own-runners/about-self-hosted-runners
process
ci setup self hosted github runner migrating to private repo who wants it product owner martin bm what is the proposed change setup a self hosted runner maybe on a raspberry pi to run the github actions so the project can be made possible to migrate to a private repo in case this will be a business idea follow these guidelines
1
489,623
14,109,019,520
IssuesEvent
2020-11-06 18:50:03
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Python wheel missing for grpcio version > 1.27.2
kind/bug lang/Python priority/P2
<!-- PLEASE DO NOT POST A QUESTION HERE. This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? Python 3.6.8 grpcio==1.33.2 ### What operating system (Linux, Windows,...) and version? Centos 7 Ubuntu 18..04 ### What runtime / compiler are you using (e.g. python version or version of gcc) gcc (GCC) 4.8.5 20150623 (Red Hat 4.8.5-39) gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 ### What did you do? pip3 install grpcio==1.33.2 --only-binary ":all:" ### What did you expect to see? Should simply install the package using an existing python wheel ### What did you see instead? Based on the output below, which happens on both Centos 7 and Ubuntu 18.04, there isn't a python wheel for anything above grpcio 1.27.2. ``` sudo pip3 install grpcio==1.33.2 --only-binary ":all:" WARNING: Running pip install with root privileges is generally not a good idea. Try `pip3 install --user` instead. Collecting grpcio==1.33.2 Could not find a version that satisfies the requirement grpcio==1.33.2 (from versions: 1.1.3, 1.2.0rc1, 1.2.0rc2, 1.2.0, 1.2.1, 1.3.0, 1.3.3, 1.3.5, 1.4.0rc1, 1.4.0, 1.6.0rc1, 1.6.0, 1.6.3, 1.7.0rc1, 1.7.0, 1.7.3, 1.8.0rc1, 1.8.0rc2, 1.8.0rc3, 1.8.1, 1.8.2, 1.8.3, 1.8.4, 1.8.6, 1.9.0rc3, 1.9.0, 1.9.1, 1.10.0rc1, 1.10.0rc2, 1.10.0, 1.10.1rc1, 1.10.1rc2, 1.10.1, 1.11.0rc1, 1.11.0rc2, 1.11.0, 1.11.1rc1, 1.11.1, 1.12.0rc1, 1.12.0, 1.12.1, 1.13.0rc1, 1.13.0rc2, 1.13.0rc3, 1.13.0, 1.14.0rc1, 1.14.0rc2, 1.14.0, 1.14.1, 1.14.2rc1, 1.14.2, 1.15.0rc1, 1.15.0, 1.16.0rc1, 1.16.0, 1.16.1rc1, 1.16.1, 1.17.0rc1, 1.17.0, 1.17.1rc1, 1.17.1, 1.18.0rc1, 1.18.0, 1.19.0rc1, 1.19.0, 1.20.0rc1, 1.20.0rc2, 1.20.0rc3, 1.20.0, 1.20.1, 1.21.0rc1, 1.21.1rc1, 1.21.1, 1.22.0rc1, 1.22.0, 1.22.1, 1.23.0rc1, 1.23.0, 1.23.1, 1.24.0rc1, 1.24.0, 1.24.1, 1.24.3, 1.25.0rc1, 1.25.0, 1.26.0rc1, 1.26.0, 1.27.0rc1, 1.27.1, 1.27.2) No matching distribution found for grpcio==1.33.2 ``` ### Anything else we should know about your project / environment? We have to deploy this to multiple systems. If we don't use `--only-binary`, our deployments have to wait for grpci to compile on every system and it causes deployments to take forever.
1.0
Python wheel missing for grpcio version > 1.27.2 - <!-- PLEASE DO NOT POST A QUESTION HERE. This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? Python 3.6.8 grpcio==1.33.2 ### What operating system (Linux, Windows,...) and version? Centos 7 Ubuntu 18..04 ### What runtime / compiler are you using (e.g. python version or version of gcc) gcc (GCC) 4.8.5 20150623 (Red Hat 4.8.5-39) gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 ### What did you do? pip3 install grpcio==1.33.2 --only-binary ":all:" ### What did you expect to see? Should simply install the package using an existing python wheel ### What did you see instead? Based on the output below, which happens on both Centos 7 and Ubuntu 18.04, there isn't a python wheel for anything above grpcio 1.27.2. ``` sudo pip3 install grpcio==1.33.2 --only-binary ":all:" WARNING: Running pip install with root privileges is generally not a good idea. Try `pip3 install --user` instead. Collecting grpcio==1.33.2 Could not find a version that satisfies the requirement grpcio==1.33.2 (from versions: 1.1.3, 1.2.0rc1, 1.2.0rc2, 1.2.0, 1.2.1, 1.3.0, 1.3.3, 1.3.5, 1.4.0rc1, 1.4.0, 1.6.0rc1, 1.6.0, 1.6.3, 1.7.0rc1, 1.7.0, 1.7.3, 1.8.0rc1, 1.8.0rc2, 1.8.0rc3, 1.8.1, 1.8.2, 1.8.3, 1.8.4, 1.8.6, 1.9.0rc3, 1.9.0, 1.9.1, 1.10.0rc1, 1.10.0rc2, 1.10.0, 1.10.1rc1, 1.10.1rc2, 1.10.1, 1.11.0rc1, 1.11.0rc2, 1.11.0, 1.11.1rc1, 1.11.1, 1.12.0rc1, 1.12.0, 1.12.1, 1.13.0rc1, 1.13.0rc2, 1.13.0rc3, 1.13.0, 1.14.0rc1, 1.14.0rc2, 1.14.0, 1.14.1, 1.14.2rc1, 1.14.2, 1.15.0rc1, 1.15.0, 1.16.0rc1, 1.16.0, 1.16.1rc1, 1.16.1, 1.17.0rc1, 1.17.0, 1.17.1rc1, 1.17.1, 1.18.0rc1, 1.18.0, 1.19.0rc1, 1.19.0, 1.20.0rc1, 1.20.0rc2, 1.20.0rc3, 1.20.0, 1.20.1, 1.21.0rc1, 1.21.1rc1, 1.21.1, 1.22.0rc1, 1.22.0, 1.22.1, 1.23.0rc1, 1.23.0, 1.23.1, 1.24.0rc1, 1.24.0, 1.24.1, 1.24.3, 1.25.0rc1, 1.25.0, 1.26.0rc1, 1.26.0, 1.27.0rc1, 1.27.1, 1.27.2) No matching distribution found for grpcio==1.33.2 ``` ### Anything else we should know about your project / environment? We have to deploy this to multiple systems. If we don't use `--only-binary`, our deployments have to wait for grpci to compile on every system and it causes deployments to take forever.
non_process
python wheel missing for grpcio version please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using python grpcio what operating system linux windows and version centos ubuntu what runtime compiler are you using e g python version or version of gcc gcc gcc red hat gcc ubuntu what did you do install grpcio only binary all what did you expect to see should simply install the package using an existing python wheel what did you see instead based on the output below which happens on both centos and ubuntu there isn t a python wheel for anything above grpcio sudo install grpcio only binary all warning running pip install with root privileges is generally not a good idea try install user instead collecting grpcio could not find a version that satisfies the requirement grpcio from versions no matching distribution found for grpcio anything else we should know about your project environment we have to deploy this to multiple systems if we don t use only binary our deployments have to wait for grpci to compile on every system and it causes deployments to take forever
0
10,793
13,609,065,976
IssuesEvent
2020-09-23 04:10:29
googleapis/java-compute
https://api.github.com/repos/googleapis/java-compute
closed
Dependency Dashboard
api: compute type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10.1.0 - [ ] <!-- rebase-branch=renovate/com.google.apis-google-api-services-compute-1.x -->deps: update dependency com.google.apis:google-api-services-compute to v1-rev20200811-1.30.10 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1 - [ ] <!-- rebase-branch=renovate/org.easymock-easymock-4.x -->deps: update dependency org.easymock:easymock to v4 - [ ] <!-- rebase-branch=renovate/org.objenesis-objenesis-3.x -->deps: update dependency org.objenesis:objenesis to v3 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10.1.0 - [ ] <!-- rebase-branch=renovate/com.google.apis-google-api-services-compute-1.x -->deps: update dependency com.google.apis:google-api-services-compute to v1-rev20200811-1.30.10 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1 - [ ] <!-- rebase-branch=renovate/org.easymock-easymock-4.x -->deps: update dependency org.easymock:easymock to v4 - [ ] <!-- rebase-branch=renovate/org.objenesis-objenesis-3.x -->deps: update dependency org.objenesis:objenesis to v3 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud libraries bom to deps update dependency com google apis google api services compute to deps update dependency com google cloud google cloud storage to deps update dependency org easymock easymock to deps update dependency org objenesis objenesis to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
1
8,192
11,392,260,727
IssuesEvent
2020-01-30 01:51:39
aodn/imos-toolbox
https://api.github.com/repos/aodn/imos-toolbox
opened
Validate DB entries
Type:enhancement Unit:Processing
A common issue with users is unrecognizable errors in the toolbox when some database entry is wrong. The pain point here is that the toolbox debugging msg is, most of the time, unclear for those cases. Moreover, it's daunting to detect this kind of problems in large/lot of csv files. As the DB field size increases, so does the chance of errors and failure because of a single entry. We should try to validate the entry(ies) at read time, and provide a more useful error message.
1.0
Validate DB entries - A common issue with users is unrecognizable errors in the toolbox when some database entry is wrong. The pain point here is that the toolbox debugging msg is, most of the time, unclear for those cases. Moreover, it's daunting to detect this kind of problems in large/lot of csv files. As the DB field size increases, so does the chance of errors and failure because of a single entry. We should try to validate the entry(ies) at read time, and provide a more useful error message.
process
validate db entries a common issue with users is unrecognizable errors in the toolbox when some database entry is wrong the pain point here is that the toolbox debugging msg is most of the time unclear for those cases moreover it s daunting to detect this kind of problems in large lot of csv files as the db field size increases so does the chance of errors and failure because of a single entry we should try to validate the entry ies at read time and provide a more useful error message
1
45,454
5,716,261,609
IssuesEvent
2017-04-19 14:48:45
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
Fix batch attestation calculation for lineage
fixed (test & close)
this will now be for flow_diagram and not for lineage
1.0
Fix batch attestation calculation for lineage - this will now be for flow_diagram and not for lineage
non_process
fix batch attestation calculation for lineage this will now be for flow diagram and not for lineage
0
15,650
19,846,704,406
IssuesEvent
2022-01-21 07:30:48
ooi-data/RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:30:47.907585. ## Details Flow name: `RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:30:47.907585. ## Details Flow name: `RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst dpc ctd instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
194,505
22,262,009,611
IssuesEvent
2022-06-10 01:58:50
nanopathi/linux-4.19.72_CVE-2021-3347
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-3347
reopened
WS-2021-0524 (Medium) detected in linuxlinux-4.19.239
security vulnerability
## WS-2021-0524 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-3347/commit/8b7c061438f230c475fd8cd97a0917f6ebb9fbe0">8b7c061438f230c475fd8cd97a0917f6ebb9fbe0</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Linux/Kernel is vulnerable to crash with CONFIG_FORTIRY_SOURCE in arch/arm/mach-socfpga/platsmp.c <p>Publish Date: 2021-12-01 <p>URL: <a href=https://github.com/gregkh/linux/commit/20a09c8b25a21506cf8b60bdef78f31a6549ebfc>WS-2021-0524</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002532">https://osv.dev/vulnerability/GSD-2021-1002532</a></p> <p>Release Date: 2021-12-01</p> <p>Fix Resolution: Linux/Kernel -v4.14.257, v4.19.219, v5.4.163, v5.10.83, v5.15.6, v5.16-rc3 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0524 (Medium) detected in linuxlinux-4.19.239 - ## WS-2021-0524 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-3347/commit/8b7c061438f230c475fd8cd97a0917f6ebb9fbe0">8b7c061438f230c475fd8cd97a0917f6ebb9fbe0</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Linux/Kernel is vulnerable to crash with CONFIG_FORTIRY_SOURCE in arch/arm/mach-socfpga/platsmp.c <p>Publish Date: 2021-12-01 <p>URL: <a href=https://github.com/gregkh/linux/commit/20a09c8b25a21506cf8b60bdef78f31a6549ebfc>WS-2021-0524</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002532">https://osv.dev/vulnerability/GSD-2021-1002532</a></p> <p>Release Date: 2021-12-01</p> <p>Fix Resolution: Linux/Kernel -v4.14.257, v4.19.219, v5.4.163, v5.10.83, v5.15.6, v5.16-rc3 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in linuxlinux ws medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details in linux kernel is vulnerable to crash with config fortiry source in arch arm mach socfpga platsmp c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux kernel step up your open source security game with whitesource
0
600,302
18,293,040,110
IssuesEvent
2021-10-05 17:18:04
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
Dependency error for onezone command
Category: CLI Type: Bug Status: Accepted Priority: Normal
**Description** When installing the CLI in a node different than the Frontend, `onezone` command will throw an error because of an unmet dependency: ``` $ onezone list Traceback (most recent call last): 5: from /usr/bin/onezone:39:in `<main>' 4: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' 3: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' 2: from /usr/lib/one/ruby/cli/one_helper/onezone_helper.rb:19:in `<top (required)>' 1: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require': cannot load such file -- CommandManager (LoadError) ``` **Expected behavior** Every command dependency should be correctly installed. **Details** - Affected Component: CLI - Version:5.12 & master <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Dependency error for onezone command - **Description** When installing the CLI in a node different than the Frontend, `onezone` command will throw an error because of an unmet dependency: ``` $ onezone list Traceback (most recent call last): 5: from /usr/bin/onezone:39:in `<main>' 4: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' 3: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' 2: from /usr/lib/one/ruby/cli/one_helper/onezone_helper.rb:19:in `<top (required)>' 1: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require' /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require': cannot load such file -- CommandManager (LoadError) ``` **Expected behavior** Every command dependency should be correctly installed. **Details** - Affected Component: CLI - Version:5.12 & master <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
non_process
dependency error for onezone command description when installing the cli in a node different than the frontend onezone command will throw an error because of an unmet dependency onezone list traceback most recent call last from usr bin onezone in from usr lib ruby rubygems core ext kernel require rb in require from usr lib ruby rubygems core ext kernel require rb in require from usr lib one ruby cli one helper onezone helper rb in from usr lib ruby rubygems core ext kernel require rb in require usr lib ruby rubygems core ext kernel require rb in require cannot load such file commandmanager loaderror expected behavior every command dependency should be correctly installed details affected component cli version master progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
0
10,115
13,044,162,218
IssuesEvent
2020-07-29 03:47:30
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `WeekWithoutMode` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `WeekWithoutMode` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `WeekWithoutMode` from TiDB - ## Description Port the scalar function `WeekWithoutMode` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function weekwithoutmode from tidb description port the scalar function weekwithoutmode from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
181,326
30,668,353,371
IssuesEvent
2023-07-25 20:09:14
readthedocs/readthedocs.org
https://api.github.com/repos/readthedocs/readthedocs.org
closed
Support searching with facets.
Feature Needed: design decision
Search facets will be useful in many ways. They can be used to target the search inside a subproject. For example: Celery has a subproject Kombu, so we can search in Celery docs, we can search inside Kombu docs, like: `subproject: kombu serializers`
1.0
Support searching with facets. - Search facets will be useful in many ways. They can be used to target the search inside a subproject. For example: Celery has a subproject Kombu, so we can search in Celery docs, we can search inside Kombu docs, like: `subproject: kombu serializers`
non_process
support searching with facets search facets will be useful in many ways they can be used to target the search inside a subproject for example celery has a subproject kombu so we can search in celery docs we can search inside kombu docs like subproject kombu serializers
0
9,562
12,518,751,418
IssuesEvent
2020-06-03 13:26:48
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Reevaluate ASP.NET Core analyzer value
type: process
The analyzer complicates project generation - we should check that it still adds value. (It may well be that as the installation approach has changed, many of the things it guards against are irrelevant.) Even if we keep it, we could potentially not generate the project file - or just generate certain parts of it (e.g. version and copyright), making the rest manual.
1.0
Reevaluate ASP.NET Core analyzer value - The analyzer complicates project generation - we should check that it still adds value. (It may well be that as the installation approach has changed, many of the things it guards against are irrelevant.) Even if we keep it, we could potentially not generate the project file - or just generate certain parts of it (e.g. version and copyright), making the rest manual.
process
reevaluate asp net core analyzer value the analyzer complicates project generation we should check that it still adds value it may well be that as the installation approach has changed many of the things it guards against are irrelevant even if we keep it we could potentially not generate the project file or just generate certain parts of it e g version and copyright making the rest manual
1
8,299
8,222,993,130
IssuesEvent
2018-09-06 09:14:28
a2develop/bugTracker
https://api.github.com/repos/a2develop/bugTracker
closed
Свойства ФОПа. Отчество -- неточная подсказка.
c_beta c_service done! misspelling
Для поля "По батькові" подсказка "Введить прізвище та ініціали" Наверное, нужна типа "Введить по батькові" ![2018-08-30](https://user-images.githubusercontent.com/35918141/44994899-6abe7280-afa9-11e8-8129-fed2eeaff887.png)
1.0
Свойства ФОПа. Отчество -- неточная подсказка. - Для поля "По батькові" подсказка "Введить прізвище та ініціали" Наверное, нужна типа "Введить по батькові" ![2018-08-30](https://user-images.githubusercontent.com/35918141/44994899-6abe7280-afa9-11e8-8129-fed2eeaff887.png)
non_process
свойства фопа отчество неточная подсказка для поля по батькові подсказка введить прізвище та ініціали наверное нужна типа введить по батькові
0
10,326
13,162,135,300
IssuesEvent
2020-08-10 20:57:21
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
vsphere post processor lacks roles documentation
docs help wanted post-processor/vsphere
The only way I can get the vSphere post-processor to work, is when I add the configured user to the administrators group. This will cause lots of red flags in a security audit. Please include in the documentation which vSphere roles by Packer are required so we can fine-tune access. I tried the recommendations here: https://github.com/TotalJobsGroup/Packer/blob/master/README.md But it seems this is not complete, as I still get permissions in Packer when I only enable those roles.
1.0
vsphere post processor lacks roles documentation - The only way I can get the vSphere post-processor to work, is when I add the configured user to the administrators group. This will cause lots of red flags in a security audit. Please include in the documentation which vSphere roles by Packer are required so we can fine-tune access. I tried the recommendations here: https://github.com/TotalJobsGroup/Packer/blob/master/README.md But it seems this is not complete, as I still get permissions in Packer when I only enable those roles.
process
vsphere post processor lacks roles documentation the only way i can get the vsphere post processor to work is when i add the configured user to the administrators group this will cause lots of red flags in a security audit please include in the documentation which vsphere roles by packer are required so we can fine tune access i tried the recommendations here but it seems this is not complete as i still get permissions in packer when i only enable those roles
1
121,881
26,046,645,130
IssuesEvent
2022-12-22 14:56:08
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Language String Problem with Menu Heading Menu Type
No Code Attached Yet
### Steps to reproduce the issue Joomla 4.2.6 create a new menu item select System Links >> Menu Heading type the language string of Options Tab does not display properly it displays constant as "COM_MENUS_PAGE-OPTIONS_FIELDSET_LABEL" see attached image. I have searched for language string whole package with text-search tool and found nothing. image https://prnt.sc/snrVNXAs7bCG ### Expected result It should be Options ### Actual result ### System information (as much as possible) ### Additional comments
1.0
Language String Problem with Menu Heading Menu Type - ### Steps to reproduce the issue Joomla 4.2.6 create a new menu item select System Links >> Menu Heading type the language string of Options Tab does not display properly it displays constant as "COM_MENUS_PAGE-OPTIONS_FIELDSET_LABEL" see attached image. I have searched for language string whole package with text-search tool and found nothing. image https://prnt.sc/snrVNXAs7bCG ### Expected result It should be Options ### Actual result ### System information (as much as possible) ### Additional comments
non_process
language string problem with menu heading menu type steps to reproduce the issue joomla create a new menu item select system links menu heading type the language string of options tab does not display properly it displays constant as com menus page options fieldset label see attached image i have searched for language string whole package with text search tool and found nothing image expected result it should be options actual result system information as much as possible additional comments
0
2,524
5,288,207,339
IssuesEvent
2017-02-08 14:35:52
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Catch error in case there is no RDMA device and show meaningful error
process_wontfix type_enhancement
``` 2016-12-17 08:13:18 931793 -0800 - ubuntu-16 - 24199/0x00007fcd2ef50740 - dtl_server//usr/bin/failovercache - 0000000000000002 - info - run: starting, server path: "/mnt/hdd4/pstorage_write_dtl_1", address to bind to: 125.212.218.15, port: 26217, transport type: RSocket, busy-loop usecs: 0 [...] dtl_server//usr/bin/failovercache - 0000000000000005 - fatal - operator(): caught exception: socket No such device (19) ``` DTL is configured to use RSocket, but it looks like there's no RDMA device / the wrong one is specified.
1.0
Catch error in case there is no RDMA device and show meaningful error - ``` 2016-12-17 08:13:18 931793 -0800 - ubuntu-16 - 24199/0x00007fcd2ef50740 - dtl_server//usr/bin/failovercache - 0000000000000002 - info - run: starting, server path: "/mnt/hdd4/pstorage_write_dtl_1", address to bind to: 125.212.218.15, port: 26217, transport type: RSocket, busy-loop usecs: 0 [...] dtl_server//usr/bin/failovercache - 0000000000000005 - fatal - operator(): caught exception: socket No such device (19) ``` DTL is configured to use RSocket, but it looks like there's no RDMA device / the wrong one is specified.
process
catch error in case there is no rdma device and show meaningful error ubuntu dtl server usr bin failovercache info run starting server path mnt pstorage write dtl address to bind to port transport type rsocket busy loop usecs dtl server usr bin failovercache fatal operator caught exception socket no such device dtl is configured to use rsocket but it looks like there s no rdma device the wrong one is specified
1
356,944
25,176,294,191
IssuesEvent
2022-11-11 09:33:25
ongwenpin/pe
https://api.github.com/repos/ongwenpin/pe
opened
Extra "G" in sequence diagram of DG
severity.VeryLow type.DocumentationBug
![Screenshot 2022-11-11 at 5.31.10 PM.png](https://raw.githubusercontent.com/ongwenpin/pe/main/files/7866ea89-23f3-4055-9ca0-e37dca61aea2.png) On page 30 of the DG, the sequence diagram has extra letter "G" which may lead to confusion for readers. <!--session: 1668153030050-2e911811-a668-4ac5-b895-28e31f18355d--> <!--Version: Web v3.4.4-->
1.0
Extra "G" in sequence diagram of DG - ![Screenshot 2022-11-11 at 5.31.10 PM.png](https://raw.githubusercontent.com/ongwenpin/pe/main/files/7866ea89-23f3-4055-9ca0-e37dca61aea2.png) On page 30 of the DG, the sequence diagram has extra letter "G" which may lead to confusion for readers. <!--session: 1668153030050-2e911811-a668-4ac5-b895-28e31f18355d--> <!--Version: Web v3.4.4-->
non_process
extra g in sequence diagram of dg on page of the dg the sequence diagram has extra letter g which may lead to confusion for readers
0
15,807
20,008,700,617
IssuesEvent
2022-02-01 01:48:26
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Resources should be processed in Web Workers
TYPE: bug AREA: client SYSTEM: resource processing health-monitor SYSTEM: workers STATE: Stale
The Service Workers feature is usually used for resource-caching. Currently, when the user's service worker uses [Caches](https://w3c.github.io/ServiceWorker/#cache-objects) objects, all resources are cached unprocessed. So, in the Web Workers we should process resources before they will be placed into the cache. Example in which the service worker puts a script into the cache after the installation is finished: ```js require('http') .createServer((req, res) => { if (req.url === '/') { res.writeHead(200, { 'content-type': 'text/html' }); res.end(` <!DOCTYPE html> <head> <link rel="shortcut icon" href="#" /> </head> <body> <script> navigator.serviceWorker.register('sw.js'); </script> </body> `); } else if (req.url === '/sw.js') { res.writeHead(200, { 'content-type': 'application/javascript' }); res.end(` self.addEventListener('install', function (event) { const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js'; event.waitUntil( caches.open('v1').then(function (cache) { return cache.add(url); }) ); }); `); } else res.destroy(); }) .listen(2020, () => console.log('http://localhost:2020')); ``` also, `fetch` could be used: ```js self.addEventListener('install', function (event) { const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js'; event.waitUntil( caches.open('v1').then(function (cache) { return fetch(url).then(function(response) { if (!response.ok) throw new TypeError('bad response status'); return cache.put(url, response); }) }) ); }); ``` or the fetch event handler could respond with the Response object, containing a script: ```js self.addEventListener('fetch', function(event) { if(event.request.url.match(/some\-script\.js/)) { event.respondWith( new Response(`console.log('mocked some-script.js');`, { headers: {'Content-Type': 'text/javascript'} }) ); } }); ``` This bug was found during analysis of the Health-Monitor run issues on the twitter.com website.
1.0
Resources should be processed in Web Workers - The Service Workers feature is usually used for resource-caching. Currently, when the user's service worker uses [Caches](https://w3c.github.io/ServiceWorker/#cache-objects) objects, all resources are cached unprocessed. So, in the Web Workers we should process resources before they will be placed into the cache. Example in which the service worker puts a script into the cache after the installation is finished: ```js require('http') .createServer((req, res) => { if (req.url === '/') { res.writeHead(200, { 'content-type': 'text/html' }); res.end(` <!DOCTYPE html> <head> <link rel="shortcut icon" href="#" /> </head> <body> <script> navigator.serviceWorker.register('sw.js'); </script> </body> `); } else if (req.url === '/sw.js') { res.writeHead(200, { 'content-type': 'application/javascript' }); res.end(` self.addEventListener('install', function (event) { const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js'; event.waitUntil( caches.open('v1').then(function (cache) { return cache.add(url); }) ); }); `); } else res.destroy(); }) .listen(2020, () => console.log('http://localhost:2020')); ``` also, `fetch` could be used: ```js self.addEventListener('install', function (event) { const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js'; event.waitUntil( caches.open('v1').then(function (cache) { return fetch(url).then(function(response) { if (!response.ok) throw new TypeError('bad response status'); return cache.put(url, response); }) }) ); }); ``` or the fetch event handler could respond with the Response object, containing a script: ```js self.addEventListener('fetch', function(event) { if(event.request.url.match(/some\-script\.js/)) { event.respondWith( new Response(`console.log('mocked some-script.js');`, { headers: {'Content-Type': 'text/javascript'} }) ); } }); ``` This bug was found during analysis of the Health-Monitor run issues on the twitter.com website.
process
resources should be processed in web workers the service workers feature is usually used for resource caching currently when the user s service worker uses objects all resources are cached unprocessed so in the web workers we should process resources before they will be placed into the cache example in which the service worker puts a script into the cache after the installation is finished js require http createserver req res if req url res writehead content type text html res end navigator serviceworker register sw js else if req url sw js res writehead content type application javascript res end self addeventlistener install function event const url event waituntil caches open then function cache return cache add url else res destroy listen console log also fetch could be used js self addeventlistener install function event const url event waituntil caches open then function cache return fetch url then function response if response ok throw new typeerror bad response status return cache put url response or the fetch event handler could respond with the response object containing a script js self addeventlistener fetch function event if event request url match some script js event respondwith new response console log mocked some script js headers content type text javascript this bug was found during analysis of the health monitor run issues on the twitter com website
1
675
3,145,705,071
IssuesEvent
2015-09-14 19:17:03
K0zka/kerub
https://api.github.com/repos/K0zka/kerub
opened
Add exception mapping for Jackson object mapping exceptions
component:data processing enhancement priority: normal
A partial vm json for example should not create an uggly 500, but some meaningful error.
1.0
Add exception mapping for Jackson object mapping exceptions - A partial vm json for example should not create an uggly 500, but some meaningful error.
process
add exception mapping for jackson object mapping exceptions a partial vm json for example should not create an uggly but some meaningful error
1
2,566
5,316,106,015
IssuesEvent
2017-02-13 19:02:06
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [fr] MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI »
Language: French Process: [7] Confirmation sent to Antoine
# Video title MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI » # URL https://youtu.be/EDC1-D4qYV8 # Youtube subtitles language Langue des sous-titres : Français # Duration 7'39 # Subtitles URL https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&bl=watch&ref=wt&ui=hd&v=EDC1-D4qYV8&lang=fr&tab=captions
1.0
[subtitles] [fr] MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI » - # Video title MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI » # URL https://youtu.be/EDC1-D4qYV8 # Youtube subtitles language Langue des sous-titres : Français # Duration 7'39 # Subtitles URL https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&bl=watch&ref=wt&ui=hd&v=EDC1-D4qYV8&lang=fr&tab=captions
process
mélenchon « mon bulletin de vote est un coup de balai » video title mélenchon « mon bulletin de vote est un coup de balai » url youtube subtitles language langue des sous titres français duration subtitles url
1
18,164
24,201,554,491
IssuesEvent
2022-09-24 16:41:22
fertadeo/ISPC-2do-Cuat-Proyecto
https://api.github.com/repos/fertadeo/ISPC-2do-Cuat-Proyecto
closed
FrontEnd - #US02 - Crear archivos iniciales del proyecto
in process
Crear archivo index.html style.css main.js
1.0
FrontEnd - #US02 - Crear archivos iniciales del proyecto - Crear archivo index.html style.css main.js
process
frontend crear archivos iniciales del proyecto crear archivo index html style css main js
1
17,287
23,095,899,302
IssuesEvent
2022-07-26 19:30:48
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
opened
Traduzir/remover códigos remanescentes de módulos que utilizem o Pyppeteer
[0] Desenvolvimento [2] Média Prioridade [1] Aprimoramento [3] Processamento Dinâmico
## Comportamento Esperado Com a introdução do Playwright no sistema, qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca. Códigos específicos do Pyppeteer que não estejam em uso devem ser removidos, e códigos que estejam em uso devem ser corretamente traduzidos. ## Comportamento Atual Algumas partes do código ainda utilizam a biblioteca Pyppeteer para realização de interações dinâmicas, entre elas: - Módulo `antiblock_drivers` (menciona o Pyppeteer no README, mas não utiliza internamente) - View `load_iframe` (utiliza o módulo `scrapy_puppeteer`, que deve ser removido) - Módulo `step-by-step` (menciona o Pyppeteer no código e utiliza em alguns exemplos) - Módulo `entry_probing` - Módulo `scrapy_puppeteer` (obsoleto e deve ser removido)
1.0
Traduzir/remover códigos remanescentes de módulos que utilizem o Pyppeteer - ## Comportamento Esperado Com a introdução do Playwright no sistema, qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca. Códigos específicos do Pyppeteer que não estejam em uso devem ser removidos, e códigos que estejam em uso devem ser corretamente traduzidos. ## Comportamento Atual Algumas partes do código ainda utilizam a biblioteca Pyppeteer para realização de interações dinâmicas, entre elas: - Módulo `antiblock_drivers` (menciona o Pyppeteer no README, mas não utiliza internamente) - View `load_iframe` (utiliza o módulo `scrapy_puppeteer`, que deve ser removido) - Módulo `step-by-step` (menciona o Pyppeteer no código e utiliza em alguns exemplos) - Módulo `entry_probing` - Módulo `scrapy_puppeteer` (obsoleto e deve ser removido)
process
traduzir remover códigos remanescentes de módulos que utilizem o pyppeteer comportamento esperado com a introdução do playwright no sistema qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca códigos específicos do pyppeteer que não estejam em uso devem ser removidos e códigos que estejam em uso devem ser corretamente traduzidos comportamento atual algumas partes do código ainda utilizam a biblioteca pyppeteer para realização de interações dinâmicas entre elas módulo antiblock drivers menciona o pyppeteer no readme mas não utiliza internamente view load iframe utiliza o módulo scrapy puppeteer que deve ser removido módulo step by step menciona o pyppeteer no código e utiliza em alguns exemplos módulo entry probing módulo scrapy puppeteer obsoleto e deve ser removido
1
276,847
30,554,827,461
IssuesEvent
2023-07-20 10:57:30
CliffCrerar/bohemian-rhapsody
https://api.github.com/repos/CliffCrerar/bohemian-rhapsody
opened
CVE-2023-26136 (Medium) detected in tough-cookie-2.5.0.tgz, tough-cookie-3.0.1.tgz
Mend: dependency security vulnerability
## CVE-2023-26136 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tough-cookie-2.5.0.tgz</b>, <b>tough-cookie-3.0.1.tgz</b></p></summary> <p> <details><summary><b>tough-cookie-2.5.0.tgz</b></p></summary> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - ng-deploy-0.2.3.tgz (Root Library) - arm-resources-1.1.0.tgz - ms-rest-js-1.8.14.tgz - :x: **tough-cookie-2.5.0.tgz** (Vulnerable Library) </details> <details><summary><b>tough-cookie-3.0.1.tgz</b></p></summary> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/@azure/storage-blob/node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - ng-deploy-0.2.3.tgz (Root Library) - storage-blob-10.5.0.tgz - ms-rest-js-2.0.5.tgz - :x: **tough-cookie-3.0.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bohemian-rhapsody/commit/ecce634609a1383c732a17c20e59f30a68158b81">ecce634609a1383c732a17c20e59f30a68158b81</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package tough-cookie before 4.1.3 are vulnerable to Prototype Pollution due to improper handling of Cookies when using CookieJar in rejectPublicSuffixes=false mode. This issue arises from the manner in which the objects are initialized. <p>Publish Date: 2023-07-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26136>CVE-2023-26136</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-26136">https://www.cve.org/CVERecord?id=CVE-2023-26136</a></p> <p>Release Date: 2023-07-01</p> <p>Fix Resolution: tough-cookie - 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-26136 (Medium) detected in tough-cookie-2.5.0.tgz, tough-cookie-3.0.1.tgz - ## CVE-2023-26136 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tough-cookie-2.5.0.tgz</b>, <b>tough-cookie-3.0.1.tgz</b></p></summary> <p> <details><summary><b>tough-cookie-2.5.0.tgz</b></p></summary> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - ng-deploy-0.2.3.tgz (Root Library) - arm-resources-1.1.0.tgz - ms-rest-js-1.8.14.tgz - :x: **tough-cookie-2.5.0.tgz** (Vulnerable Library) </details> <details><summary><b>tough-cookie-3.0.1.tgz</b></p></summary> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/@azure/storage-blob/node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - ng-deploy-0.2.3.tgz (Root Library) - storage-blob-10.5.0.tgz - ms-rest-js-2.0.5.tgz - :x: **tough-cookie-3.0.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bohemian-rhapsody/commit/ecce634609a1383c732a17c20e59f30a68158b81">ecce634609a1383c732a17c20e59f30a68158b81</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package tough-cookie before 4.1.3 are vulnerable to Prototype Pollution due to improper handling of Cookies when using CookieJar in rejectPublicSuffixes=false mode. This issue arises from the manner in which the objects are initialized. <p>Publish Date: 2023-07-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26136>CVE-2023-26136</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-26136">https://www.cve.org/CVERecord?id=CVE-2023-26136</a></p> <p>Release Date: 2023-07-01</p> <p>Fix Resolution: tough-cookie - 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in tough cookie tgz tough cookie tgz cve medium severity vulnerability vulnerable libraries tough cookie tgz tough cookie tgz tough cookie tgz cookies and cookie jar for node js library home page a href path to dependency file package json path to vulnerable library node modules tough cookie package json dependency hierarchy ng deploy tgz root library arm resources tgz ms rest js tgz x tough cookie tgz vulnerable library tough cookie tgz cookies and cookie jar for node js library home page a href path to dependency file package json path to vulnerable library node modules azure storage blob node modules tough cookie package json dependency hierarchy ng deploy tgz root library storage blob tgz ms rest js tgz x tough cookie tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package tough cookie before are vulnerable to prototype pollution due to improper handling of cookies when using cookiejar in rejectpublicsuffixes false mode this issue arises from the manner in which the objects are initialized publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tough cookie step up your open source security game with mend
0
11,210
13,959,170,566
IssuesEvent
2020-10-24 15:18:21
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
QA ISO images
0.kind: bug 6.topic: GNOME 6.topic: nixos 6.topic: qt/kde 6.topic: release process
**Describe the bug** Some amount of manual QA is needed to verify these images are in good shape. In 20.03 we had a pretty silly bug slip through and got dragged on HN :rofl: Even without that as encouragement this is a good idea. The main goal is to determine the following: > It needs to be verified that all ISO's that are using a release blocking desktop boot cleanly in their supported configurations. They must boot live when written to a USB stick with **any** of our supported methods documented in the NixOS manual. I've drafted some criteria to look out for when testing a desktop. The version for this exact issue can be seen at this [gist](https://gist.github.com/worldofpeace/019e0b77020affbe3c3623d807f73f22). After having testing please comment with a checklist on the items like ```markdown - [x] Shutdown, reboot, login, logout ... ``` The following images need testing: - [x] Plasma5 ISO - [x] GNOME ISO Here are links to the latest images: https://channels.nixos.org/nixos-20.09/latest-nixos-plasma5-x86_64-linux.iso ~https://channels.nixos.org/nixos-20.09/latest-nixos-gnome-x86_64-linux.iso~ The gnome iso isn't added to the channel atm: You can get it from an eval of it on hydra: https://hydra.nixos.org/job/nixos/release-20.09/nixos.iso_gnome.x86_64-linux picking a build https://hydra.nixos.org/build/128080348 and selecting `Link` from `Build Products`. # Remaining Q.A tasks ## Plasma - [x] Printing https://github.com/NixOS/nixpkgs/issues/98536 - [ ] Keyboard layout configuration - [ ] unlocking encrypted storage volumes - [x] console - this is tested automatically - [ ] When logging in via the default login manager for a release-blocking desktop - [ ] After logging in to a release-blocking desktop, if the user account does not have its own keyboard layout configuration for that desktop (if there is such a user/desktop-specific configuration, it must be used when that user logs in to that desktop) - [ ] Desktop Keyring - [ ] User Switching? ## GNOME - [x] configured keyboard layout (using nixos options) is used when unlocking encrypted storage volumes - [x] Desktop Keyring
1.0
QA ISO images - **Describe the bug** Some amount of manual QA is needed to verify these images are in good shape. In 20.03 we had a pretty silly bug slip through and got dragged on HN :rofl: Even without that as encouragement this is a good idea. The main goal is to determine the following: > It needs to be verified that all ISO's that are using a release blocking desktop boot cleanly in their supported configurations. They must boot live when written to a USB stick with **any** of our supported methods documented in the NixOS manual. I've drafted some criteria to look out for when testing a desktop. The version for this exact issue can be seen at this [gist](https://gist.github.com/worldofpeace/019e0b77020affbe3c3623d807f73f22). After having testing please comment with a checklist on the items like ```markdown - [x] Shutdown, reboot, login, logout ... ``` The following images need testing: - [x] Plasma5 ISO - [x] GNOME ISO Here are links to the latest images: https://channels.nixos.org/nixos-20.09/latest-nixos-plasma5-x86_64-linux.iso ~https://channels.nixos.org/nixos-20.09/latest-nixos-gnome-x86_64-linux.iso~ The gnome iso isn't added to the channel atm: You can get it from an eval of it on hydra: https://hydra.nixos.org/job/nixos/release-20.09/nixos.iso_gnome.x86_64-linux picking a build https://hydra.nixos.org/build/128080348 and selecting `Link` from `Build Products`. # Remaining Q.A tasks ## Plasma - [x] Printing https://github.com/NixOS/nixpkgs/issues/98536 - [ ] Keyboard layout configuration - [ ] unlocking encrypted storage volumes - [x] console - this is tested automatically - [ ] When logging in via the default login manager for a release-blocking desktop - [ ] After logging in to a release-blocking desktop, if the user account does not have its own keyboard layout configuration for that desktop (if there is such a user/desktop-specific configuration, it must be used when that user logs in to that desktop) - [ ] Desktop Keyring - [ ] User Switching? ## GNOME - [x] configured keyboard layout (using nixos options) is used when unlocking encrypted storage volumes - [x] Desktop Keyring
process
qa iso images describe the bug some amount of manual qa is needed to verify these images are in good shape in we had a pretty silly bug slip through and got dragged on hn rofl even without that as encouragement this is a good idea the main goal is to determine the following it needs to be verified that all iso s that are using a release blocking desktop boot cleanly in their supported configurations they must boot live when written to a usb stick with any of our supported methods documented in the nixos manual i ve drafted some criteria to look out for when testing a desktop the version for this exact issue can be seen at this after having testing please comment with a checklist on the items like markdown shutdown reboot login logout the following images need testing iso gnome iso here are links to the latest images the gnome iso isn t added to the channel atm you can get it from an eval of it on hydra picking a build and selecting link from build products remaining q a tasks plasma printing keyboard layout configuration unlocking encrypted storage volumes console this is tested automatically when logging in via the default login manager for a release blocking desktop after logging in to a release blocking desktop if the user account does not have its own keyboard layout configuration for that desktop if there is such a user desktop specific configuration it must be used when that user logs in to that desktop desktop keyring user switching gnome configured keyboard layout using nixos options is used when unlocking encrypted storage volumes desktop keyring
1
43,897
2,893,895,318
IssuesEvent
2015-06-15 20:23:15
Naoghuman/Dream-Better-Worlds
https://api.github.com/repos/Naoghuman/Dream-Better-Worlds
closed
Refactoring logger level in all onActionXy methods from info to debug.
development refactoring issue fixed priority low
- Refactoring logger level in all onActionXy methods from info to debug. - (v) Add list from all classes which have onActionXy methods.
1.0
Refactoring logger level in all onActionXy methods from info to debug. - - Refactoring logger level in all onActionXy methods from info to debug. - (v) Add list from all classes which have onActionXy methods.
non_process
refactoring logger level in all onactionxy methods from info to debug refactoring logger level in all onactionxy methods from info to debug v add list from all classes which have onactionxy methods
0
16,873
22,152,749,303
IssuesEvent
2022-06-03 18:44:53
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
closed
Brainstorm ways to reduce Cloud Shell dependency
type: process priority: p3
Backend changes can break sandbox in ways we can't control. We should take some time to brainstorm ways we can prevent these issues in the future. Related to https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/issues/754, where we discuss moving sandboxctl into the Cloud Shell base image, rather than maintaining a custom image. But even using the base image, we may encounter issues related to Cloud Shell. We should consider more substantial re-designs as well
1.0
Brainstorm ways to reduce Cloud Shell dependency - Backend changes can break sandbox in ways we can't control. We should take some time to brainstorm ways we can prevent these issues in the future. Related to https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/issues/754, where we discuss moving sandboxctl into the Cloud Shell base image, rather than maintaining a custom image. But even using the base image, we may encounter issues related to Cloud Shell. We should consider more substantial re-designs as well
process
brainstorm ways to reduce cloud shell dependency backend changes can break sandbox in ways we can t control we should take some time to brainstorm ways we can prevent these issues in the future related to where we discuss moving sandboxctl into the cloud shell base image rather than maintaining a custom image but even using the base image we may encounter issues related to cloud shell we should consider more substantial re designs as well
1
7,624
10,729,241,390
IssuesEvent
2019-10-28 15:16:37
usgs/libcomcat
https://api.github.com/repos/usgs/libcomcat
closed
Implement nightly builds
process
Choose a method for building nightly (i.e. Jenkins, Azure and cron job, gitlab runners, etc.).
1.0
Implement nightly builds - Choose a method for building nightly (i.e. Jenkins, Azure and cron job, gitlab runners, etc.).
process
implement nightly builds choose a method for building nightly i e jenkins azure and cron job gitlab runners etc
1
56,801
15,373,966,970
IssuesEvent
2021-03-02 13:16:27
Questie/Questie
https://api.github.com/repos/Questie/Questie
closed
Questie tracker in invalid location, resetting....
Type - Defect
<!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> My questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly. When I open the game it says that is is resetting but never does. When I manually reset it to the center of the screen I can see it (though it is not populated) but when I go to move it anywhere else on the screen it disappears immediately ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". -->v6.2.5
1.0
Questie tracker in invalid location, resetting.... - <!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> My questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly. When I open the game it says that is is resetting but never does. When I manually reset it to the center of the screen I can see it (though it is not populated) but when I go to move it anywhere else on the screen it disappears immediately ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". -->v6.2.5
non_process
questie tracker in invalid location resetting read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description my questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly when i open the game it says that is is resetting but never does when i manually reset it to the center of the screen i can see it though it is not populated but when i go to move it anywhere else on the screen it disappears immediately screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version
0
432,449
12,493,516,066
IssuesEvent
2020-06-01 09:25:10
mesg-foundation/aragon
https://api.github.com/repos/mesg-foundation/aragon
closed
Delete connection
high priority
1. Aragon app mark as removed the connection on the aragon smart contract 2. A process listen for this removed event from the smart contract and trigger the service process-deployer to delete the process 3. the process-deployer service using the processUrl, download it and calculate its hash (may need to use calculate a hash endpoint). with the hash, it can remove the process from firebase using the delete process endpoint (or something like this).
1.0
Delete connection - 1. Aragon app mark as removed the connection on the aragon smart contract 2. A process listen for this removed event from the smart contract and trigger the service process-deployer to delete the process 3. the process-deployer service using the processUrl, download it and calculate its hash (may need to use calculate a hash endpoint). with the hash, it can remove the process from firebase using the delete process endpoint (or something like this).
non_process
delete connection aragon app mark as removed the connection on the aragon smart contract a process listen for this removed event from the smart contract and trigger the service process deployer to delete the process the process deployer service using the processurl download it and calculate its hash may need to use calculate a hash endpoint with the hash it can remove the process from firebase using the delete process endpoint or something like this
0
219,829
17,114,104,913
IssuesEvent
2021-07-11 00:42:29
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Java Backend Engineer Specialist na AgileProcess
AWS CI Docker Git Java MySQL Remoto SQL Scrum Stale Testes Unitários
## Descrição da vaga Buscamos um(a) **Backend Engineer Specialist** que será responsável, junto ao nossos squads de desenvolvimento, por prover a melhor experiência para nossos clientes através de nossas soluções. RESPONSABILIDADES E ATRIBUIÇÕES \- Desafiar o status quo e desenvolver soluções inovadoras para problemas complexos; \- Desenvolver e manter nossos Microserviços de forma ágil, aplicando boas práticas de Engenharia de Software; \- Contribuir com o desenvolvimento e arquitetura da plataforma, preparando-a para um crescimento acelerado; \- Construir uma base sólida para o desenvolvimento de novos produtos; \- Desenvolver sistemas escaláveis, sustentáveis e orientados ao usuário; \- Trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência; \- Ajudar o crescimento do time de tecnologia e engenharia. REQUISITOS E QUALIFICAÇÕES \- Experiência e conhecimento profundo com desenvolvimento Java 8; \- Experiência e conhecimento em GitFlow; \- Experiência e conhecimento profundo em Docker; \- Ter atuado na construção de testes unitários e integrados; \- Ter atuado na construção de testes de comportamento (BDD); \- Conhecimentos em Design Patterns, arquitetura e engenharia de software; \- Conhecimentos em GitLab CI; \- Conhecimentos em metodologias ágeis de desenvolvimento (Scrum, Kanban). INFORMAÇÕES ADICIONAIS **Nosso Stack:** \- Java; \- MySQL; \- AWS; \- Git (GitLab). **DNA Agile <3** Faz parte do DNA de todo **AGILE** ser ambicioso e buscar entregar sempre mais e melhor. O respeito e a tolerância são indispensáveis para nós, e somos apaixonados pela nossa diversidade. Crescemos como time e celebrando juntos todos os resultados atingidos. Somos Customer Centric e buscamos criar conexão com nossos clientes, com simplicidade e empatia. Compartilhamos de forma direta e honesta todas as informações e opiniões possíveis. Reconhecemos nossas fraquezas e assumimos nossos erros. Somos apaixonados por inovação e usamos da nossa criatividade, autonomia e voz para propor novas formas de fazer as coisas. Somos representados por cada um e nadamos todos na mesma direção.​ **O que você vai encontrar:** * Onboarding de boas-vindas! * “All Hands”: nosso encontro semanal com o CEO; * Dress Code: seja você mesmo(a); * Flexibilidade de horário; * VR/VA Flex: R$ 550,00 (mês); * Plano de saúde (para você e quem você ama); * Plano odontológico; * TotalPass; * Clube de Descontos - NewValue; * PLR; * Parceria com ZenKlub; * e muito mais! **Redes Sociais**: Venha conferir nossas redes sociais e revolucionar a logística com a gente! * Instagram: @agileprocess * LinkedIn: /agileprocess * Twitter: @agileprocess _\*Todas as vagas da AgileProcess são abertas para pessoas com deficiência (PcD)._ _\*Nos processos seletivos da AgileProcess, não há qualquer tipo de distinção em virtude de gênero, orientação sexual, religião, crença, condição social ou raça._ ## Salário R$12500 - R$14000 ## Local REMOTO ## Como se candidatar Accesse a [Vaga de Java Backend Engineer Specialist](https://vaga.contratar.dev/apply/full/06686ffa-44e4-4505-a874-f719b0532a21) para se candidatar. ## Labels * Remoto * SQL * Scrum * Redes * AWS * Docker * Java * Git ## Sobre a contratar.dev A contratar.dev é uma plataforma de recrutamento e seleção automatizada de Desenvolvedores. Ao se candidatar a uma vaga da contratar.dev, você estará participando automaticamente de dezenas de processos seletivos para vagas similares, aumentando em MUITO as suas chances de mudar de vida 😉
1.0
[Remoto] Java Backend Engineer Specialist na AgileProcess - ## Descrição da vaga Buscamos um(a) **Backend Engineer Specialist** que será responsável, junto ao nossos squads de desenvolvimento, por prover a melhor experiência para nossos clientes através de nossas soluções. RESPONSABILIDADES E ATRIBUIÇÕES \- Desafiar o status quo e desenvolver soluções inovadoras para problemas complexos; \- Desenvolver e manter nossos Microserviços de forma ágil, aplicando boas práticas de Engenharia de Software; \- Contribuir com o desenvolvimento e arquitetura da plataforma, preparando-a para um crescimento acelerado; \- Construir uma base sólida para o desenvolvimento de novos produtos; \- Desenvolver sistemas escaláveis, sustentáveis e orientados ao usuário; \- Trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência; \- Ajudar o crescimento do time de tecnologia e engenharia. REQUISITOS E QUALIFICAÇÕES \- Experiência e conhecimento profundo com desenvolvimento Java 8; \- Experiência e conhecimento em GitFlow; \- Experiência e conhecimento profundo em Docker; \- Ter atuado na construção de testes unitários e integrados; \- Ter atuado na construção de testes de comportamento (BDD); \- Conhecimentos em Design Patterns, arquitetura e engenharia de software; \- Conhecimentos em GitLab CI; \- Conhecimentos em metodologias ágeis de desenvolvimento (Scrum, Kanban). INFORMAÇÕES ADICIONAIS **Nosso Stack:** \- Java; \- MySQL; \- AWS; \- Git (GitLab). **DNA Agile <3** Faz parte do DNA de todo **AGILE** ser ambicioso e buscar entregar sempre mais e melhor. O respeito e a tolerância são indispensáveis para nós, e somos apaixonados pela nossa diversidade. Crescemos como time e celebrando juntos todos os resultados atingidos. Somos Customer Centric e buscamos criar conexão com nossos clientes, com simplicidade e empatia. Compartilhamos de forma direta e honesta todas as informações e opiniões possíveis. Reconhecemos nossas fraquezas e assumimos nossos erros. Somos apaixonados por inovação e usamos da nossa criatividade, autonomia e voz para propor novas formas de fazer as coisas. Somos representados por cada um e nadamos todos na mesma direção.​ **O que você vai encontrar:** * Onboarding de boas-vindas! * “All Hands”: nosso encontro semanal com o CEO; * Dress Code: seja você mesmo(a); * Flexibilidade de horário; * VR/VA Flex: R$ 550,00 (mês); * Plano de saúde (para você e quem você ama); * Plano odontológico; * TotalPass; * Clube de Descontos - NewValue; * PLR; * Parceria com ZenKlub; * e muito mais! **Redes Sociais**: Venha conferir nossas redes sociais e revolucionar a logística com a gente! * Instagram: @agileprocess * LinkedIn: /agileprocess * Twitter: @agileprocess _\*Todas as vagas da AgileProcess são abertas para pessoas com deficiência (PcD)._ _\*Nos processos seletivos da AgileProcess, não há qualquer tipo de distinção em virtude de gênero, orientação sexual, religião, crença, condição social ou raça._ ## Salário R$12500 - R$14000 ## Local REMOTO ## Como se candidatar Accesse a [Vaga de Java Backend Engineer Specialist](https://vaga.contratar.dev/apply/full/06686ffa-44e4-4505-a874-f719b0532a21) para se candidatar. ## Labels * Remoto * SQL * Scrum * Redes * AWS * Docker * Java * Git ## Sobre a contratar.dev A contratar.dev é uma plataforma de recrutamento e seleção automatizada de Desenvolvedores. Ao se candidatar a uma vaga da contratar.dev, você estará participando automaticamente de dezenas de processos seletivos para vagas similares, aumentando em MUITO as suas chances de mudar de vida 😉
non_process
java backend engineer specialist na agileprocess descrição da vaga buscamos um a backend engineer specialist que será responsável junto ao nossos squads de desenvolvimento por prover a melhor experiência para nossos clientes através de nossas soluções responsabilidades e atribuições desafiar o status quo e desenvolver soluções inovadoras para problemas complexos desenvolver e manter nossos microserviços de forma ágil aplicando boas práticas de engenharia de software contribuir com o desenvolvimento e arquitetura da plataforma preparando a para um crescimento acelerado construir uma base sólida para o desenvolvimento de novos produtos desenvolver sistemas escaláveis sustentáveis e orientados ao usuário trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência ajudar o crescimento do time de tecnologia e engenharia requisitos e qualificações experiência e conhecimento profundo com desenvolvimento java experiência e conhecimento em gitflow experiência e conhecimento profundo em docker ter atuado na construção de testes unitários e integrados ter atuado na construção de testes de comportamento bdd conhecimentos em design patterns arquitetura e engenharia de software conhecimentos em gitlab ci conhecimentos em metodologias ágeis de desenvolvimento scrum kanban informações adicionais nosso stack java mysql aws git gitlab dna agile faz parte do dna de todo agile ser ambicioso e buscar entregar sempre mais e melhor o respeito e a tolerância são indispensáveis para nós e somos apaixonados pela nossa diversidade crescemos como time e celebrando juntos todos os resultados atingidos somos customer centric e buscamos criar conexão com nossos clientes com simplicidade e empatia compartilhamos de forma direta e honesta todas as informações e opiniões possíveis reconhecemos nossas fraquezas e assumimos nossos erros somos apaixonados por inovação e usamos da nossa criatividade autonomia e voz para propor novas formas de fazer as coisas somos representados por cada um e nadamos todos na mesma direção ​ o que você vai encontrar onboarding de boas vindas “all hands” nosso encontro semanal com o ceo dress code seja você mesmo a flexibilidade de horário vr va flex r mês plano de saúde para você e quem você ama plano odontológico totalpass clube de descontos newvalue plr parceria com zenklub e muito mais redes sociais venha conferir nossas redes sociais e revolucionar a logística com a gente instagram agileprocess linkedin agileprocess twitter agileprocess todas as vagas da agileprocess são abertas para pessoas com deficiência pcd nos processos seletivos da agileprocess não há qualquer tipo de distinção em virtude de gênero orientação sexual religião crença condição social ou raça salário r r local remoto como se candidatar accesse a para se candidatar labels remoto sql scrum redes aws docker java git sobre a contratar dev a contratar dev é uma plataforma de recrutamento e seleção automatizada de desenvolvedores ao se candidatar a uma vaga da contratar dev você estará participando automaticamente de dezenas de processos seletivos para vagas similares aumentando em muito as suas chances de mudar de vida 😉
0
2,937
5,920,860,911
IssuesEvent
2017-05-22 21:20:43
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
closed
HIV: fails to parse
in progress ontology processing problem
Submission 2.6 of the [HIV ontology](http://bioportal.bioontology.org/ontologies/HIV), uploaded on March 20, 2017 failed to parse. BioPortal UI shows status of "Error Rdf". Preliminary investigation shows several issues: 1. The user incorrectly designated the submission file format as OWL when it should be OBO. This shouldn't affect the parsing step though. The format should be corrected when the next submission is created. 2. Line 2 of their ontology source file uses an incorrect date format. The [OBO Flat File Format Guide](https://owlcollab.github.io/oboformat/doc/GO.format.obo-1_4.html) states that dates must be in a format of dd:MM:yyyy HH:mm. If you parse this ontology with version 4.2.5 of the OWL API, this exception is thrown: ``` Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@3888cf9 Stack trace: LINENO: 2 - Could not parse date from string: December 27, 2016 LINE: date: December 27, 2016 org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60) ``` To fix this exception, change line 2 of the ontology source file to this: ``` date: 27:12:2016 00:00 ``` 3. After fixing line 2 of the ontology, re-parsing results in further errors, e.g.: ``` Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@15563bcf Stack trace: LINENO: 38 - Could not find tag for: part_of LINE: part_of: HIV:29 ! retrovirus org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60) uk.ac.manchester.cs.owl.owlapi.OWLOntologyFactoryImpl.loadOWLOntology(OWLOntologyFactoryImpl.java:197) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.actualParse(OWLOntologyManagerImpl.java:1061) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntology(OWLOntologyManagerImpl.java:1017) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntologyFromOntologyDocument(OWLOntologyManagerImpl.java:967) edu.stanford.ncbo.bioportal.SimpleLoadTest.testLoadHIVOntology_WithDocumentFormat(SimpleLoadTest.java:50) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:606) ``` These exceptions are thrown with version 4.2.5 of the OWL API, which BioPortal currently uses internally. Version 4.2.6 (and above) of the OWL API successfully parse the ontology with no exceptions. The user reported that they could open the ontology in the Protege ontology editor without issues. The latest version of Protege (5.2.0) uses version 4.2.8 of the OWL API, thus the lack of issue opening the HIV ontology in their environment.
1.0
HIV: fails to parse - Submission 2.6 of the [HIV ontology](http://bioportal.bioontology.org/ontologies/HIV), uploaded on March 20, 2017 failed to parse. BioPortal UI shows status of "Error Rdf". Preliminary investigation shows several issues: 1. The user incorrectly designated the submission file format as OWL when it should be OBO. This shouldn't affect the parsing step though. The format should be corrected when the next submission is created. 2. Line 2 of their ontology source file uses an incorrect date format. The [OBO Flat File Format Guide](https://owlcollab.github.io/oboformat/doc/GO.format.obo-1_4.html) states that dates must be in a format of dd:MM:yyyy HH:mm. If you parse this ontology with version 4.2.5 of the OWL API, this exception is thrown: ``` Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@3888cf9 Stack trace: LINENO: 2 - Could not parse date from string: December 27, 2016 LINE: date: December 27, 2016 org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60) ``` To fix this exception, change line 2 of the ontology source file to this: ``` date: 27:12:2016 00:00 ``` 3. After fixing line 2 of the ontology, re-parsing results in further errors, e.g.: ``` Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@15563bcf Stack trace: LINENO: 38 - Could not find tag for: part_of LINE: part_of: HIV:29 ! retrovirus org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60) uk.ac.manchester.cs.owl.owlapi.OWLOntologyFactoryImpl.loadOWLOntology(OWLOntologyFactoryImpl.java:197) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.actualParse(OWLOntologyManagerImpl.java:1061) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntology(OWLOntologyManagerImpl.java:1017) uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntologyFromOntologyDocument(OWLOntologyManagerImpl.java:967) edu.stanford.ncbo.bioportal.SimpleLoadTest.testLoadHIVOntology_WithDocumentFormat(SimpleLoadTest.java:50) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:606) ``` These exceptions are thrown with version 4.2.5 of the OWL API, which BioPortal currently uses internally. Version 4.2.6 (and above) of the OWL API successfully parse the ontology with no exceptions. The user reported that they could open the ontology in the Protege ontology editor without issues. The latest version of Protege (5.2.0) uses version 4.2.8 of the OWL API, thus the lack of issue opening the HIV ontology in their environment.
process
hiv fails to parse submission of the uploaded on march failed to parse bioportal ui shows status of error rdf preliminary investigation shows several issues the user incorrectly designated the submission file format as owl when it should be obo this shouldn t affect the parsing step though the format should be corrected when the next submission is created line of their ontology source file uses an incorrect date format the states that dates must be in a format of dd mm yyyy hh mm if you parse this ontology with version of the owl api this exception is thrown parser org semanticweb owlapi oboformat oboformatowlapiparser stack trace lineno could not parse date from string december line date december org semanticweb owlapi oboformat oboformatowlapiparser parse oboformatowlapiparser java to fix this exception change line of the ontology source file to this date after fixing line of the ontology re parsing results in further errors e g parser org semanticweb owlapi oboformat oboformatowlapiparser stack trace lineno could not find tag for part of line part of hiv retrovirus org semanticweb owlapi oboformat oboformatowlapiparser parse oboformatowlapiparser java uk ac manchester cs owl owlapi owlontologyfactoryimpl loadowlontology owlontologyfactoryimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl actualparse owlontologymanagerimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl loadontology owlontologymanagerimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl loadontologyfromontologydocument owlontologymanagerimpl java edu stanford ncbo bioportal simpleloadtest testloadhivontology withdocumentformat simpleloadtest java sun reflect nativemethodaccessorimpl native method sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java java lang reflect method invoke method java these exceptions are thrown with version of the owl api which bioportal currently uses internally version and above of the owl api successfully parse the ontology with no exceptions the user reported that they could open the ontology in the protege ontology editor without issues the latest version of protege uses version of the owl api thus the lack of issue opening the hiv ontology in their environment
1
9,277
12,302,870,443
IssuesEvent
2020-05-11 17:42:35
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
RPM for openSUSE / SUSE Enterprise Linux / Need for dependency listing.
team-EngProd type: process under investigation untriaged
* Operating System: Linux * Bazel version (output of `bazel info release`): 0.8.0 #2701 is related I'm developing an rpm for openSUSE / SUSE which will be built on https://build.opensuse.org and hosted in a repo with a pretty permanent address on download.opensuse.org Before getting it officially in the distro, we need to be able to build bazel without the bundled third party bits. To do so, I would need a full list of the dependencies either as jars or libraries. Is there a way to get bazel to use the distro version of protobuf ? That is already in the distro and I suspect most of the bundled jars.
1.0
RPM for openSUSE / SUSE Enterprise Linux / Need for dependency listing. - * Operating System: Linux * Bazel version (output of `bazel info release`): 0.8.0 #2701 is related I'm developing an rpm for openSUSE / SUSE which will be built on https://build.opensuse.org and hosted in a repo with a pretty permanent address on download.opensuse.org Before getting it officially in the distro, we need to be able to build bazel without the bundled third party bits. To do so, I would need a full list of the dependencies either as jars or libraries. Is there a way to get bazel to use the distro version of protobuf ? That is already in the distro and I suspect most of the bundled jars.
process
rpm for opensuse suse enterprise linux need for dependency listing operating system linux bazel version output of bazel info release is related i m developing an rpm for opensuse suse which will be built on and hosted in a repo with a pretty permanent address on download opensuse org before getting it officially in the distro we need to be able to build bazel without the bundled third party bits to do so i would need a full list of the dependencies either as jars or libraries is there a way to get bazel to use the distro version of protobuf that is already in the distro and i suspect most of the bundled jars
1
15,696
19,848,160,629
IssuesEvent
2022-01-21 09:16:26
ooi-data/CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered
https://api.github.com/repos/ooi-data/CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:16:25.469693. ## Details Flow name: `CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:16:25.469693. ## Details Flow name: `CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp dofst k wfp instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
153,202
12,137,397,699
IssuesEvent
2020-04-23 15:40:25
deathlyrage/pot-demo-bugs
https://api.github.com/repos/deathlyrage/pot-demo-bugs
closed
Deinonychus slides when it gets up from rest
needs testing
Deinonychus slides when it gets up from rest/movement doesn't match animation.
1.0
Deinonychus slides when it gets up from rest - Deinonychus slides when it gets up from rest/movement doesn't match animation.
non_process
deinonychus slides when it gets up from rest deinonychus slides when it gets up from rest movement doesn t match animation
0
6,066
8,902,729,898
IssuesEvent
2019-01-17 08:33:08
Juris-M/citeproc-js
https://api.github.com/repos/Juris-M/citeproc-js
closed
Bug: creator label formatting
fix in process
I am working with this style: https://gist.github.com/bwiernik/f33d5ebf8793fbe663f826d478298c76 In the `"format"` macro, I use `names` to call `reviewed-author`: ``` <names variable="reviewed-author"> <label form="verb-short" suffix=" "/> <name and="symbol" initialize-with=". " delimiter=", "/> </names> ``` The expected output for this item: ``` [ { "id": "http://zotero.org/groups/2205533/items/IJBGS4BQ", "type": "post-weblog", "title": "The future for Deaf individuals is not that bleak", "URL": "http://www.personalityresearch.org/papers/hall.html", "note": "reviewed-title: Decrease of Deaf potential in a mainstreamed environment\ngenre: Peer commentary\nreviewed-author: Hall || W.C.", "language": "en", "author": [ { "family": "Wolf", "given": "K.S." } ], "issued": { "date-parts": [ [ "2005" ] ] } } ] ``` is: Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, by W. C. Hall]. Retrieved from http://www.personalityresearch.org/papers/hall.html However, what renders instead is: Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, W. C. Hallby ]. Retrieved from http://www.personalityresearch.org/papers/hall.html With the `label` coming after `name`, even though it is specified to come before. If I move the `names` element out of the `"format"` macro (e.g., directly into the bibliography layout or into the `"title"` macro), it renders correctly. If I change `reviewed-author` to another creator variable, the formatting of `label` still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable (or following the `name` if it is not specified elsewhere). The cause is that `<text macro="format">` is included in the `substitute` list of the `"author-short"` macro. If I delete the three instances of `<text macro="format">` from the `"author-short"` macro (all three must be deleted, even the two in conditional branches that do not apply to the item), the `label` formatting is rendered correctly. @adam3smith
1.0
Bug: creator label formatting - I am working with this style: https://gist.github.com/bwiernik/f33d5ebf8793fbe663f826d478298c76 In the `"format"` macro, I use `names` to call `reviewed-author`: ``` <names variable="reviewed-author"> <label form="verb-short" suffix=" "/> <name and="symbol" initialize-with=". " delimiter=", "/> </names> ``` The expected output for this item: ``` [ { "id": "http://zotero.org/groups/2205533/items/IJBGS4BQ", "type": "post-weblog", "title": "The future for Deaf individuals is not that bleak", "URL": "http://www.personalityresearch.org/papers/hall.html", "note": "reviewed-title: Decrease of Deaf potential in a mainstreamed environment\ngenre: Peer commentary\nreviewed-author: Hall || W.C.", "language": "en", "author": [ { "family": "Wolf", "given": "K.S." } ], "issued": { "date-parts": [ [ "2005" ] ] } } ] ``` is: Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, by W. C. Hall]. Retrieved from http://www.personalityresearch.org/papers/hall.html However, what renders instead is: Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, W. C. Hallby ]. Retrieved from http://www.personalityresearch.org/papers/hall.html With the `label` coming after `name`, even though it is specified to come before. If I move the `names` element out of the `"format"` macro (e.g., directly into the bibliography layout or into the `"title"` macro), it renders correctly. If I change `reviewed-author` to another creator variable, the formatting of `label` still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable (or following the `name` if it is not specified elsewhere). The cause is that `<text macro="format">` is included in the `substitute` list of the `"author-short"` macro. If I delete the three instances of `<text macro="format">` from the `"author-short"` macro (all three must be deleted, even the two in conditional branches that do not apply to the item), the `label` formatting is rendered correctly. @adam3smith
process
bug creator label formatting i am working with this style in the format macro i use names to call reviewed author the expected output for this item id type post weblog title the future for deaf individuals is not that bleak url note reviewed title decrease of deaf potential in a mainstreamed environment ngenre peer commentary nreviewed author hall w c language en author family wolf given k s issued date parts is wolf k s the future for deaf individuals is not that bleak retrieved from however what renders instead is wolf k s the future for deaf individuals is not that bleak retrieved from with the label coming after name even though it is specified to come before if i move the names element out of the format macro e g directly into the bibliography layout or into the title macro it renders correctly if i change reviewed author to another creator variable the formatting of label still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable or following the name if it is not specified elsewhere the cause is that is included in the substitute list of the author short macro if i delete the three instances of from the author short macro all three must be deleted even the two in conditional branches that do not apply to the item the label formatting is rendered correctly
1
439,241
12,679,747,385
IssuesEvent
2020-06-19 12:25:49
nokazn/spotify-player
https://api.github.com/repos/nokazn/spotify-player
closed
公式のランキングのプレイリストの追加日
0. low-priority 2. improvement 💪
## 問題 spotifycharts が作成したプレイリスト (バイラルチャートなど) では追加日が `1970-01-01-T00:00:00Z`になっている。
1.0
公式のランキングのプレイリストの追加日 - ## 問題 spotifycharts が作成したプレイリスト (バイラルチャートなど) では追加日が `1970-01-01-T00:00:00Z`になっている。
non_process
公式のランキングのプレイリストの追加日 問題 spotifycharts が作成したプレイリスト バイラルチャートなど では追加日が になっている。
0
3,696
3,222,840,851
IssuesEvent
2015-10-09 05:26:35
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
closed
strange drgui release build error with gcc 5.1+
Component-Build OpSys-Linux
Here's another release build error with gcc 5.1.1 (xref #1796), first mentioned in #1788: ``` /work/dr/git/src/ext/drgui/drgui_main_window.cpp: In destructor ‘drgui_main_window_t::~drgui_main_window_t()’: /work/dr/git/src/ext/drgui/drgui_main_window.cpp:100:1: error: assuming signed overflow does not occur when assuming that (X - c) <= X is always true [-Werror=strict-overflow] drgui_main_window_t::~drgui_main_window_t(void) ``` It took a while to figure out where this was coming from: it's the line "plugins.pop_back()" and it's QVector<T>::reallocData() being inlined (pop_back() calls removeLast() which calls reallocData()). This is the line in reallocData() in /usr/include/qt5/QtCore/qvector.h: ``` if (asize <= d->size) { ``` Changing it to this removes the warning: ``` if ((uint)asize <= d->size) { ``` Of course we don't want everyone to go edit their system headers so we'll turn off this warning.
1.0
strange drgui release build error with gcc 5.1+ - Here's another release build error with gcc 5.1.1 (xref #1796), first mentioned in #1788: ``` /work/dr/git/src/ext/drgui/drgui_main_window.cpp: In destructor ‘drgui_main_window_t::~drgui_main_window_t()’: /work/dr/git/src/ext/drgui/drgui_main_window.cpp:100:1: error: assuming signed overflow does not occur when assuming that (X - c) <= X is always true [-Werror=strict-overflow] drgui_main_window_t::~drgui_main_window_t(void) ``` It took a while to figure out where this was coming from: it's the line "plugins.pop_back()" and it's QVector<T>::reallocData() being inlined (pop_back() calls removeLast() which calls reallocData()). This is the line in reallocData() in /usr/include/qt5/QtCore/qvector.h: ``` if (asize <= d->size) { ``` Changing it to this removes the warning: ``` if ((uint)asize <= d->size) { ``` Of course we don't want everyone to go edit their system headers so we'll turn off this warning.
non_process
strange drgui release build error with gcc here s another release build error with gcc xref first mentioned in work dr git src ext drgui drgui main window cpp in destructor ‘drgui main window t drgui main window t ’ work dr git src ext drgui drgui main window cpp error assuming signed overflow does not occur when assuming that x c x is always true drgui main window t drgui main window t void it took a while to figure out where this was coming from it s the line plugins pop back and it s qvector reallocdata being inlined pop back calls removelast which calls reallocdata this is the line in reallocdata in usr include qtcore qvector h if asize size changing it to this removes the warning if uint asize size of course we don t want everyone to go edit their system headers so we ll turn off this warning
0
8,496
11,659,800,299
IssuesEvent
2020-03-03 01:11:49
MicrosoftDocs/vsts-docs
https://api.github.com/repos/MicrosoftDocs/vsts-docs
closed
dependencies for stages?
Pri1 devops-cicd-process/tech devops/prod doc-bug
How does dependencies work between stages? Can we use output variables from jobs in previous stages? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#type-casting) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
dependencies for stages? - How does dependencies work between stages? Can we use output variables from jobs in previous stages? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#type-casting) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
dependencies for stages how does dependencies work between stages can we use output variables from jobs in previous stages document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
20,347
27,002,946,217
IssuesEvent
2023-02-10 09:23:27
inmanta/inmanta-core
https://api.github.com/repos/inmanta/inmanta-core
opened
configure dependabot to use the widen strategy
process
dependabot/dependabot-core#6632 implements the `widen` strategy for Python packages. The behavior for this strategy matches the behavior we want and have gotten up till now. However, the default behavior wasn't necessarily meant to be like this (dependabot/dependabot-core#6631) and might change in the future. Once the next version of `dependabot` is released (`>0.215.0`) we'd best configure dependabot explicitly to use the desired strategy for all relevant repos (at least product and component repos, probably all our tools (pytest-inmanta, vscode-inmanta, ...), I'm not sure about modules).
1.0
configure dependabot to use the widen strategy - dependabot/dependabot-core#6632 implements the `widen` strategy for Python packages. The behavior for this strategy matches the behavior we want and have gotten up till now. However, the default behavior wasn't necessarily meant to be like this (dependabot/dependabot-core#6631) and might change in the future. Once the next version of `dependabot` is released (`>0.215.0`) we'd best configure dependabot explicitly to use the desired strategy for all relevant repos (at least product and component repos, probably all our tools (pytest-inmanta, vscode-inmanta, ...), I'm not sure about modules).
process
configure dependabot to use the widen strategy dependabot dependabot core implements the widen strategy for python packages the behavior for this strategy matches the behavior we want and have gotten up till now however the default behavior wasn t necessarily meant to be like this dependabot dependabot core and might change in the future once the next version of dependabot is released we d best configure dependabot explicitly to use the desired strategy for all relevant repos at least product and component repos probably all our tools pytest inmanta vscode inmanta i m not sure about modules
1
336,254
24,491,011,260
IssuesEvent
2022-10-10 01:51:21
sytone/obsidian-remote
https://api.github.com/repos/sytone/obsidian-remote
closed
Missing TZ variable
documentation
**Is your feature request related to a problem? Please describe.** There are plugins that use the current time of the system to work. Right now it shows an hour less than my Timezone. **Describe the solution you'd like** It would be awesome to set it as a TZ variable in the docker container.
1.0
Missing TZ variable - **Is your feature request related to a problem? Please describe.** There are plugins that use the current time of the system to work. Right now it shows an hour less than my Timezone. **Describe the solution you'd like** It would be awesome to set it as a TZ variable in the docker container.
non_process
missing tz variable is your feature request related to a problem please describe there are plugins that use the current time of the system to work right now it shows an hour less than my timezone describe the solution you d like it would be awesome to set it as a tz variable in the docker container
0
25,454
7,708,972,621
IssuesEvent
2018-05-22 07:40:44
xamarin/xamarin-android
https://api.github.com/repos/xamarin/xamarin-android
closed
Getting build failure for Cheesesquare sample in both debug and release mode
Area: App+Library Build
Moving VSTS bug [BUG 614180](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/614180) to this repo Affected Branch: master Affected Build: Version 7.6 Preview (7.6 build 750) Affected Product Language: While building "Cheesesquare" sample in VSFM , build failure is observed with below error "/Users/globallogic/Downloads/monodroid-samples-master/android5.0/Cheesesquare/obj/Debug/generated/Resources/layout/list_item.xml(38,38): Error CS0234: The type or namespace name 'Android' does not exist in the namespace 'Cheesesquare' (are you missing an assembly reference?) (CS0234) (Cheesesquare)" Build JSON : [https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json](https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json) Sample link : [https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare](https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare) Steps to reproduce: Download or clone the above sample and open it in VS for MAC Set build config as debug and build the project Repeat same for Release mode Expected : App should built successfully Actual : Build failure is observed with error mentioned above Build info : [https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0](https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0) Reproducibility : 100% Failure is observed in Automation run as well [http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt](http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt ) Screen-recording : [http://g.recordit.co/LK1X0YUlvd.gif](http://g.recordit.co/LK1X0YUlvd.gif) [build and Ide logs.zip](https://github.com/xamarin/xamarin-android/files/1991562/build.and.Ide.logs.zip) Attaching IDE and Build logs too
1.0
Getting build failure for Cheesesquare sample in both debug and release mode - Moving VSTS bug [BUG 614180](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/614180) to this repo Affected Branch: master Affected Build: Version 7.6 Preview (7.6 build 750) Affected Product Language: While building "Cheesesquare" sample in VSFM , build failure is observed with below error "/Users/globallogic/Downloads/monodroid-samples-master/android5.0/Cheesesquare/obj/Debug/generated/Resources/layout/list_item.xml(38,38): Error CS0234: The type or namespace name 'Android' does not exist in the namespace 'Cheesesquare' (are you missing an assembly reference?) (CS0234) (Cheesesquare)" Build JSON : [https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json](https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json) Sample link : [https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare](https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare) Steps to reproduce: Download or clone the above sample and open it in VS for MAC Set build config as debug and build the project Repeat same for Release mode Expected : App should built successfully Actual : Build failure is observed with error mentioned above Build info : [https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0](https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0) Reproducibility : 100% Failure is observed in Automation run as well [http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt](http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt ) Screen-recording : [http://g.recordit.co/LK1X0YUlvd.gif](http://g.recordit.co/LK1X0YUlvd.gif) [build and Ide logs.zip](https://github.com/xamarin/xamarin-android/files/1991562/build.and.Ide.logs.zip) Attaching IDE and Build logs too
non_process
getting build failure for cheesesquare sample in both debug and release mode moving vsts bug to this repo affected branch master affected build version preview build affected product language while building cheesesquare sample in vsfm build failure is observed with below error users globallogic downloads monodroid samples master cheesesquare obj debug generated resources layout list item xml error the type or namespace name android does not exist in the namespace cheesesquare are you missing an assembly reference cheesesquare build json sample link steps to reproduce download or clone the above sample and open it in vs for mac set build config as debug and build the project repeat same for release mode expected app should built successfully actual build failure is observed with error mentioned above build info reproducibility failure is observed in automation run as well screen recording attaching ide and build logs too
0
71,918
30,927,999,339
IssuesEvent
2023-08-06 18:24:29
ArjunaCluster/ArjunaUsers
https://api.github.com/repos/ArjunaCluster/ArjunaUsers
opened
Arjuna is Offline (Unplanned Outage)
service-outage notify
Arjuna is currently offline (likely storm related), and we have disabled ssh access to the cluster while we return the nodes to service. We will update this thread with more information as it becomes available. Best, The Arjuna Admin Team
1.0
Arjuna is Offline (Unplanned Outage) - Arjuna is currently offline (likely storm related), and we have disabled ssh access to the cluster while we return the nodes to service. We will update this thread with more information as it becomes available. Best, The Arjuna Admin Team
non_process
arjuna is offline unplanned outage arjuna is currently offline likely storm related and we have disabled ssh access to the cluster while we return the nodes to service we will update this thread with more information as it becomes available best the arjuna admin team
0
781,267
27,430,210,787
IssuesEvent
2023-03-02 00:19:07
coral-xyz/backpack
https://api.github.com/repos/coral-xyz/backpack
closed
trailing malformed characters for nft names
help wanted good first issue priority 1 polish
This is probably a result of metaplex's deserialization of null bytes of the metadata. ![image](https://user-images.githubusercontent.com/6990215/222263527-78a419e1-9d29-4577-82e1-6e6d2387cdb8.png)
1.0
trailing malformed characters for nft names - This is probably a result of metaplex's deserialization of null bytes of the metadata. ![image](https://user-images.githubusercontent.com/6990215/222263527-78a419e1-9d29-4577-82e1-6e6d2387cdb8.png)
non_process
trailing malformed characters for nft names this is probably a result of metaplex s deserialization of null bytes of the metadata
0
95,656
16,096,857,278
IssuesEvent
2021-04-27 01:59:09
AnhaaD/auth-server
https://api.github.com/repos/AnhaaD/auth-server
opened
CVE-2021-31597 (Medium) detected in xmlhttprequest-ssl-1.5.3.tgz
security vulnerability
## CVE-2021-31597 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.3.tgz</b></p></summary> <p>XMLHttpRequest for Node</p> <p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz</a></p> <p>Path to dependency file: /auth-server/package.json</p> <p>Path to vulnerable library: auth-server/node_modules/xmlhttprequest-ssl/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-client-1.7.3.tgz - engine.io-client-1.8.3.tgz - :x: **xmlhttprequest-ssl-1.5.3.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The xmlhttprequest-ssl package before 1.6.1 for Node.js disables SSL certificate validation by default, because rejectUnauthorized (when the property exists but is undefined) is considered to be false within the https.request function of Node.js. In other words, no certificate is ever rejected. <p>Publish Date: 2021-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31597>CVE-2021-31597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597</a></p> <p>Release Date: 2021-04-23</p> <p>Fix Resolution: xmlhttprequest-ssl - 1.6.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-31597 (Medium) detected in xmlhttprequest-ssl-1.5.3.tgz - ## CVE-2021-31597 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.3.tgz</b></p></summary> <p>XMLHttpRequest for Node</p> <p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz</a></p> <p>Path to dependency file: /auth-server/package.json</p> <p>Path to vulnerable library: auth-server/node_modules/xmlhttprequest-ssl/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-client-1.7.3.tgz - engine.io-client-1.8.3.tgz - :x: **xmlhttprequest-ssl-1.5.3.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The xmlhttprequest-ssl package before 1.6.1 for Node.js disables SSL certificate validation by default, because rejectUnauthorized (when the property exists but is undefined) is considered to be false within the https.request function of Node.js. In other words, no certificate is ever rejected. <p>Publish Date: 2021-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31597>CVE-2021-31597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597</a></p> <p>Release Date: 2021-04-23</p> <p>Fix Resolution: xmlhttprequest-ssl - 1.6.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in xmlhttprequest ssl tgz cve medium severity vulnerability vulnerable library xmlhttprequest ssl tgz xmlhttprequest for node library home page a href path to dependency file auth server package json path to vulnerable library auth server node modules xmlhttprequest ssl package json dependency hierarchy karma tgz root library socket io tgz socket io client tgz engine io client tgz x xmlhttprequest ssl tgz vulnerable library vulnerability details the xmlhttprequest ssl package before for node js disables ssl certificate validation by default because rejectunauthorized when the property exists but is undefined is considered to be false within the https request function of node js in other words no certificate is ever rejected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmlhttprequest ssl step up your open source security game with whitesource
0
20,378
27,031,457,435
IssuesEvent
2023-02-12 08:43:11
Open-Data-Product-Initiative/open-data-product-spec
https://api.github.com/repos/Open-Data-Product-Initiative/open-data-product-spec
closed
Product management: Data product portfolio, product line, product series etc.
enhancement unprocessed
Data product document level attribute contains information about the product portfolio, family, line, series etc. to which it belongs. This requires a discussion, especially from a product management perspective.
1.0
Product management: Data product portfolio, product line, product series etc. - Data product document level attribute contains information about the product portfolio, family, line, series etc. to which it belongs. This requires a discussion, especially from a product management perspective.
process
product management data product portfolio product line product series etc data product document level attribute contains information about the product portfolio family line series etc to which it belongs this requires a discussion especially from a product management perspective
1
20,932
27,778,609,847
IssuesEvent
2023-03-16 19:08:30
googleapis/python-bigquery-reservation
https://api.github.com/repos/googleapis/python-bigquery-reservation
opened
Warning: a recent release failed
type: process
The following release PRs may have failed: * #49 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #42 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #49 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #42 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
process
warning a recent release failed the following release prs may have failed the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published
1
9,565
12,519,484,773
IssuesEvent
2020-06-03 14:30:33
code4romania/expert-consultation-api
https://api.github.com/repos/code4romania/expert-consultation-api
closed
[Users] Import a list of users by copy/paste from Excel file
document processing documents java nice to have spring
As an admin user of the Legal Consultation platform I want to be able to import a list of users by copy pasting from an excel file. ![Adauga membri (2)](https://user-images.githubusercontent.com/15039873/58744598-c7946f80-83f9-11e9-86a1-7a5d4a2925ad.png)
1.0
[Users] Import a list of users by copy/paste from Excel file - As an admin user of the Legal Consultation platform I want to be able to import a list of users by copy pasting from an excel file. ![Adauga membri (2)](https://user-images.githubusercontent.com/15039873/58744598-c7946f80-83f9-11e9-86a1-7a5d4a2925ad.png)
process
import a list of users by copy paste from excel file as an admin user of the legal consultation platform i want to be able to import a list of users by copy pasting from an excel file
1
5,440
8,304,002,889
IssuesEvent
2018-09-21 19:33:16
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
Release google-cloud-bigquery 1.5.1
api: bigquery type: process
We've had a recent bug fix PR that needs to get released. I forget if we've had any feature PRs since the last one.
1.0
Release google-cloud-bigquery 1.5.1 - We've had a recent bug fix PR that needs to get released. I forget if we've had any feature PRs since the last one.
process
release google cloud bigquery we ve had a recent bug fix pr that needs to get released i forget if we ve had any feature prs since the last one
1
4,998
7,834,347,697
IssuesEvent
2018-06-16 12:41:41
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Add documentation for Python Scripts (Trac #768)
Migrated from Trac enhancement post_processing weberjk@uwm.edu
Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/768 ```json { "status": "closed", "changetime": "2015-07-15T02:14:44", "description": "Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. ", "reporter": "weberjk@uwm.edu", "cc": "vlarson@uwm.edu", "resolution": "worksforme", "_ts": "1436926484509531", "component": "post_processing", "summary": "Add documentation for Python Scripts", "priority": "minor", "keywords": "", "time": "2015-04-23T15:30:13", "milestone": "", "owner": "weberjk@uwm.edu", "type": "enhancement" } ```
1.0
Add documentation for Python Scripts (Trac #768) - Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/768 ```json { "status": "closed", "changetime": "2015-07-15T02:14:44", "description": "Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. ", "reporter": "weberjk@uwm.edu", "cc": "vlarson@uwm.edu", "resolution": "worksforme", "_ts": "1436926484509531", "component": "post_processing", "summary": "Add documentation for Python Scripts", "priority": "minor", "keywords": "", "time": "2015-04-23T15:30:13", "milestone": "", "owner": "weberjk@uwm.edu", "type": "enhancement" } ```
process
add documentation for python scripts trac justin added some python scripts that he used for his thesis we may want to use them in the future let s add to their documentation use this ticket when making svn commits attachments migrated from json status closed changetime description justin added some python scripts that he used for his thesis we may want to use them in the future let s add to their documentation use this ticket when making svn commits reporter weberjk uwm edu cc vlarson uwm edu resolution worksforme ts component post processing summary add documentation for python scripts priority minor keywords time milestone owner weberjk uwm edu type enhancement
1
12,043
14,738,753,031
IssuesEvent
2021-01-07 05:38:07
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Receiving error message in SAB
anc-ops anc-process anp-1 ant-bug ant-support has attachment
In GitLab by @kdjstudios on Jul 20, 2018, 08:23 **Submitted by:** "Sarah Baptist" <sarah.baptist@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-19-29832/conversation **Server:** Internal **Client/Site:** Stockton **Account:** NA **Issue:** When trying to update/edit client account information in SAB I keep receiving this error. The client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time. ![sab+error](/uploads/2b7f4d2bc2825c5e1f631e9ca982b923/sab+error.PNG)
1.0
Receiving error message in SAB - In GitLab by @kdjstudios on Jul 20, 2018, 08:23 **Submitted by:** "Sarah Baptist" <sarah.baptist@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-19-29832/conversation **Server:** Internal **Client/Site:** Stockton **Account:** NA **Issue:** When trying to update/edit client account information in SAB I keep receiving this error. The client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time. ![sab+error](/uploads/2b7f4d2bc2825c5e1f631e9ca982b923/sab+error.PNG)
process
receiving error message in sab in gitlab by kdjstudios on jul submitted by sarah baptist helpdesk server internal client site stockton account na issue when trying to update edit client account information in sab i keep receiving this error the client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time uploads sab error png
1
11,397
14,234,427,052
IssuesEvent
2020-11-18 13:35:05
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
fatal error: 'sys/inotify.h' file not found
P3 team-Rules-CPP type: support / not a bug (process)
### Description of the problem / feature request: when I run `bazel build //test:test`, sometimes I got the error info , ``` INFO: Analyzed target //test:test (4 packages loaded, 73 targets configured). INFO: Found 1 target... ERROR: /private/var/tmp/_bazel_core/c1311ca8ca6f559bd475572aa40dd5b5/external/glfw/BUILD.bazel:120:11: C++ compilation of rule '@glfw//:linux_joystick' failed (Exit 1) wrapped_clang failed: error executing command external/local_config_cc/wrapped_clang '-D_FORTIFY_SOURCE=1' -fstack-protector -fcolor-diagnostics -Wall -Wthread-safety -Wself-assign -fno-omit-frame-pointer -O0 -DDEBUG -iquote external/glfw -iquote ... (remaining 29 argument(s) skipped) Use --sandbox_debug to see verbose messages from the sandbox external/glfw/src/linux_joystick.c:34:10: fatal error: 'sys/inotify.h' file not found #include <sys/inotify.h> ^~~~~~~~~~~~~~~ 1 error generated. Target //test:test failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 8.039s, Critical Path: 4.03s INFO: 14 processes: 5 internal, 9 darwin-sandbox. FAILED: Build did NOT complete successfully ``` So my problem is , **how to import `<sys/inotify.h>`** ### Feature requests: what underlying problem are you trying to solve with this feature? import glfw library by bazel.First of all , I solve the problem `fatal error : X11/Xlib.h not found`.But this problem occurs randomly ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. in my `third_party/glfw`, the docs structure is : ``` . ├── BUILD ├── glfw.BUILD └── workspace.bzl ``` in my `workspace.bzl`: ``` def clean_dep(dep): return str(Label(dep)) def repo(): //download glfw to local native.new_local_repository( name = "glfw", path = "/Users/core/Documents/Github/glfw", build_file = clean_dep("//third_party/glfw:glfw.BUILD"), ) ``` in my glfw.BUILD : ``` package( default_visibility = ["//visibility:public"], ) DEFINES = [ "_GLFW_X11", "_GLFW_HAS_XF86VM", ] cc_library( name = "includes", hdrs = glob([ "include/GLFW/*.h", "src/*.h", ]), ) cc_library( name = "context", srcs = ["src/context.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "init", srcs = ["src/init.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "input", srcs = ["src/input.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "monitor", srcs = ["src/monitor.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "vulkan", srcs = ["src/vulkan.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "window", srcs = ["src/window.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_init", srcs = ["src/x11_init.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_monitor", srcs = ["src/x11_monitor.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_window", srcs = ["src/x11_window.c"], defines = DEFINES, deps = [ ":egl_context", ":glx_context", ":includes", "@x11", ], ) cc_library( name = "xkb_unicode", srcs = ["src/xkb_unicode.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "linux_joystick", srcs = ["src/linux_joystick.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "posix_time", srcs = ["src/posix_time.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "glx_context", srcs = ["src/glx_context.c"], defines = DEFINES, deps = [ ":context", ":includes", # ":posix_tls", "@x11", ], ) cc_library( name = "egl_context", srcs = ["src/egl_context.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "glfw", hdrs = glob(["include/GLFW/*.h"]), defines = DEFINES, strip_include_prefix = "include", visibility = ["//visibility:public"], deps = [ ":context", ":egl_context", ":glx_context", ":init", ":input", ":linux_joystick", ":monitor", ":posix_time", #":posix_tls", ":vulkan", ":window", ":x11_init", ":x11_monitor", ":x11_window", ":xkb_unicode", ], ) ``` ### What operating system are you running Bazel on? Macos10.14.6 ### What's the output of `bazel info release`? release 3.6.0-homebrew
1.0
fatal error: 'sys/inotify.h' file not found - ### Description of the problem / feature request: when I run `bazel build //test:test`, sometimes I got the error info , ``` INFO: Analyzed target //test:test (4 packages loaded, 73 targets configured). INFO: Found 1 target... ERROR: /private/var/tmp/_bazel_core/c1311ca8ca6f559bd475572aa40dd5b5/external/glfw/BUILD.bazel:120:11: C++ compilation of rule '@glfw//:linux_joystick' failed (Exit 1) wrapped_clang failed: error executing command external/local_config_cc/wrapped_clang '-D_FORTIFY_SOURCE=1' -fstack-protector -fcolor-diagnostics -Wall -Wthread-safety -Wself-assign -fno-omit-frame-pointer -O0 -DDEBUG -iquote external/glfw -iquote ... (remaining 29 argument(s) skipped) Use --sandbox_debug to see verbose messages from the sandbox external/glfw/src/linux_joystick.c:34:10: fatal error: 'sys/inotify.h' file not found #include <sys/inotify.h> ^~~~~~~~~~~~~~~ 1 error generated. Target //test:test failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 8.039s, Critical Path: 4.03s INFO: 14 processes: 5 internal, 9 darwin-sandbox. FAILED: Build did NOT complete successfully ``` So my problem is , **how to import `<sys/inotify.h>`** ### Feature requests: what underlying problem are you trying to solve with this feature? import glfw library by bazel.First of all , I solve the problem `fatal error : X11/Xlib.h not found`.But this problem occurs randomly ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. in my `third_party/glfw`, the docs structure is : ``` . ├── BUILD ├── glfw.BUILD └── workspace.bzl ``` in my `workspace.bzl`: ``` def clean_dep(dep): return str(Label(dep)) def repo(): //download glfw to local native.new_local_repository( name = "glfw", path = "/Users/core/Documents/Github/glfw", build_file = clean_dep("//third_party/glfw:glfw.BUILD"), ) ``` in my glfw.BUILD : ``` package( default_visibility = ["//visibility:public"], ) DEFINES = [ "_GLFW_X11", "_GLFW_HAS_XF86VM", ] cc_library( name = "includes", hdrs = glob([ "include/GLFW/*.h", "src/*.h", ]), ) cc_library( name = "context", srcs = ["src/context.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "init", srcs = ["src/init.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "input", srcs = ["src/input.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "monitor", srcs = ["src/monitor.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "vulkan", srcs = ["src/vulkan.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "window", srcs = ["src/window.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_init", srcs = ["src/x11_init.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_monitor", srcs = ["src/x11_monitor.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "x11_window", srcs = ["src/x11_window.c"], defines = DEFINES, deps = [ ":egl_context", ":glx_context", ":includes", "@x11", ], ) cc_library( name = "xkb_unicode", srcs = ["src/xkb_unicode.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "linux_joystick", srcs = ["src/linux_joystick.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "posix_time", srcs = ["src/posix_time.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "glx_context", srcs = ["src/glx_context.c"], defines = DEFINES, deps = [ ":context", ":includes", # ":posix_tls", "@x11", ], ) cc_library( name = "egl_context", srcs = ["src/egl_context.c"], defines = DEFINES, deps = [ ":includes", "@x11", ], ) cc_library( name = "glfw", hdrs = glob(["include/GLFW/*.h"]), defines = DEFINES, strip_include_prefix = "include", visibility = ["//visibility:public"], deps = [ ":context", ":egl_context", ":glx_context", ":init", ":input", ":linux_joystick", ":monitor", ":posix_time", #":posix_tls", ":vulkan", ":window", ":x11_init", ":x11_monitor", ":x11_window", ":xkb_unicode", ], ) ``` ### What operating system are you running Bazel on? Macos10.14.6 ### What's the output of `bazel info release`? release 3.6.0-homebrew
process
fatal error sys inotify h file not found description of the problem feature request when i run bazel build test test sometimes i got the error info info analyzed target test test packages loaded targets configured info found target error private var tmp bazel core external glfw build bazel c compilation of rule glfw linux joystick failed exit wrapped clang failed error executing command external local config cc wrapped clang d fortify source fstack protector fcolor diagnostics wall wthread safety wself assign fno omit frame pointer ddebug iquote external glfw iquote remaining argument s skipped use sandbox debug to see verbose messages from the sandbox external glfw src linux joystick c fatal error sys inotify h file not found include error generated target test test failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal darwin sandbox failed build did not complete successfully so my problem is how to import feature requests what underlying problem are you trying to solve with this feature import glfw library by bazel first of all i solve the problem fatal error xlib h not found but this problem occurs randomly bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible in my third party glfw the docs structure is ├── build ├── glfw build └── workspace bzl in my workspace bzl def clean dep dep return str label dep def repo download glfw to local native new local repository name glfw path users core documents github glfw build file clean dep third party glfw glfw build in my glfw build package default visibility defines glfw glfw has cc library name includes hdrs glob include glfw h src h cc library name context srcs defines defines deps includes cc library name init srcs defines defines deps includes cc library name input srcs defines defines deps includes cc library name monitor srcs defines defines deps includes cc library name vulkan srcs defines defines deps includes cc library name window srcs defines defines deps includes cc library name init srcs defines defines deps includes cc library name monitor srcs defines defines deps includes cc library name window srcs defines defines deps egl context glx context includes cc library name xkb unicode srcs defines defines deps includes cc library name linux joystick srcs defines defines deps includes cc library name posix time srcs defines defines deps includes cc library name glx context srcs defines defines deps context includes posix tls cc library name egl context srcs defines defines deps includes cc library name glfw hdrs glob defines defines strip include prefix include visibility deps context egl context glx context init input linux joystick monitor posix time posix tls vulkan window init monitor window xkb unicode what operating system are you running bazel on what s the output of bazel info release release homebrew
1
7,318
10,452,928,460
IssuesEvent
2019-09-19 15:33:15
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
opened
Bug: Experience dates display incorrectly on application review page
Apply Process Bug State Dept.
Environment: Production Issue: Work experience start and end dates are displaying incorrectly on the application review page Steps to reproduce. Related ticket:
1.0
Bug: Experience dates display incorrectly on application review page - Environment: Production Issue: Work experience start and end dates are displaying incorrectly on the application review page Steps to reproduce. Related ticket:
process
bug experience dates display incorrectly on application review page environment production issue work experience start and end dates are displaying incorrectly on the application review page steps to reproduce related ticket
1
17,276
23,065,817,808
IssuesEvent
2022-07-25 13:50:59
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
Need help writing scrip 9s12
Type: Bug Feature: Processor/HCS12 Status: Internal
Ok -- getting to know Ghidra, (old IDA user). I have an issue with the CPU that I am working with (9s12) where allot of my the references to ram addresses were disassembled pointing to the wrong region of memory (CPU has a paged window from 0x8000->0xB000 that code can run in, and that code maps to physical addresses anywhere from FF:8000 -> FFB000 to E1:8000 -> E1:B000) The mapped in page is set by the ppgae register - but all ram accesses (regardless of the ppage) are always at 0x2000 -> 0x4000. Ghidra made the assumption (probably something I did wrong) that all ram accesses have the page pre-pended to the address. So, for example, if a branch is made to 0x FE:8000 and at that location there is a LDAA 0x2000 it creates that load reference as FE2000 -- which is wrong. for example: 784470 7d 29 86 STY offset DAT_0fe986 should really be: 784470 7d 29 86 STY DAT_002986 I can do this manually, by hitting the R key, and replacing the FE with the value 2, but i really want to write a script for that. I want to write a script to examine all of the references, and fix them up. can someone get me started in the right direction?
1.0
Need help writing scrip 9s12 - Ok -- getting to know Ghidra, (old IDA user). I have an issue with the CPU that I am working with (9s12) where allot of my the references to ram addresses were disassembled pointing to the wrong region of memory (CPU has a paged window from 0x8000->0xB000 that code can run in, and that code maps to physical addresses anywhere from FF:8000 -> FFB000 to E1:8000 -> E1:B000) The mapped in page is set by the ppgae register - but all ram accesses (regardless of the ppage) are always at 0x2000 -> 0x4000. Ghidra made the assumption (probably something I did wrong) that all ram accesses have the page pre-pended to the address. So, for example, if a branch is made to 0x FE:8000 and at that location there is a LDAA 0x2000 it creates that load reference as FE2000 -- which is wrong. for example: 784470 7d 29 86 STY offset DAT_0fe986 should really be: 784470 7d 29 86 STY DAT_002986 I can do this manually, by hitting the R key, and replacing the FE with the value 2, but i really want to write a script for that. I want to write a script to examine all of the references, and fix them up. can someone get me started in the right direction?
process
need help writing scrip ok getting to know ghidra old ida user i have an issue with the cpu that i am working with where allot of my the references to ram addresses were disassembled pointing to the wrong region of memory cpu has a paged window from that code can run in and that code maps to physical addresses anywhere from ff to the mapped in page is set by the ppgae register but all ram accesses regardless of the ppage are always at ghidra made the assumption probably something i did wrong that all ram accesses have the page pre pended to the address so for example if a branch is made to fe and at that location there is a ldaa it creates that load reference as which is wrong for example sty offset dat should really be sty dat i can do this manually by hitting the r key and replacing the fe with the value but i really want to write a script for that i want to write a script to examine all of the references and fix them up can someone get me started in the right direction
1