Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
12,000
| 14,738,129,531
|
IssuesEvent
|
2021-01-07 03:50:37
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
068- portland - Payment issue
|
anc-ops anc-process anp-0.5 ant-bug ant-support
|
In GitLab by @kdjstudios on May 8, 2018, 12:13
**Submitted by:** "Lettice Ross" <lettice.ross@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-23481/conversation
**Server:** Internal
**Client/Site:** 068
**Account:** Lucas Emergy B00671
**Issue:**
I’m sending this email, because I have received an error message when trying to process credit card payment. The account I’m trying to process payment for is Lucas Emergy B00671.
The error message read: We’re sorry, but something went wrong.
If you can please help, and let me know once finish.
|
1.0
|
068- portland - Payment issue - In GitLab by @kdjstudios on May 8, 2018, 12:13
**Submitted by:** "Lettice Ross" <lettice.ross@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-23481/conversation
**Server:** Internal
**Client/Site:** 068
**Account:** Lucas Emergy B00671
**Issue:**
I’m sending this email, because I have received an error message when trying to process credit card payment. The account I’m trying to process payment for is Lucas Emergy B00671.
The error message read: We’re sorry, but something went wrong.
If you can please help, and let me know once finish.
|
process
|
portland payment issue in gitlab by kdjstudios on may submitted by lettice ross helpdesk server internal client site account lucas emergy issue i’m sending this email because i have received an error message when trying to process credit card payment the account i’m trying to process payment for is lucas emergy the error message read we’re sorry but something went wrong if you can please help and let me know once finish
| 1
|
54,869
| 7,928,548,666
|
IssuesEvent
|
2018-07-06 12:09:02
|
txtdirect/txtdirect
|
https://api.github.com/repos/txtdirect/txtdirect
|
opened
|
Document that arbitrary key/value pairs are allowed
|
documentation
|
<!--
This form is for bug reports and feature requests ONLY!
If you're looking for help check out [our support guidelines](/SUPPORT.md).
-->
**Is this a BUG REPORT or FEATURE REQUEST?**:
feature
Arbitrary key/value pairs should just be discarded (this allows to upgrade TXT records in case a newer version of TXTDirect uses new keys, without breaking older versions)
Arbitrary data, however, shouldn't be allowed.
Summary:
```
v=txtv0;to=https://txtdirect.org;code=302 -> works
v=txtv0;https://txtdirect.org;code=302 -> works
v=txtv0;to=https://txtdirect.org;code=302;tracking=true -> works
v=txtv0;https://txtdirect.org;302 -> fails
v=txtv0;to=https://txtdirect.org;302 -> fails
```
|
1.0
|
Document that arbitrary key/value pairs are allowed - <!--
This form is for bug reports and feature requests ONLY!
If you're looking for help check out [our support guidelines](/SUPPORT.md).
-->
**Is this a BUG REPORT or FEATURE REQUEST?**:
feature
Arbitrary key/value pairs should just be discarded (this allows to upgrade TXT records in case a newer version of TXTDirect uses new keys, without breaking older versions)
Arbitrary data, however, shouldn't be allowed.
Summary:
```
v=txtv0;to=https://txtdirect.org;code=302 -> works
v=txtv0;https://txtdirect.org;code=302 -> works
v=txtv0;to=https://txtdirect.org;code=302;tracking=true -> works
v=txtv0;https://txtdirect.org;302 -> fails
v=txtv0;to=https://txtdirect.org;302 -> fails
```
|
non_process
|
document that arbitrary key value pairs are allowed this form is for bug reports and feature requests only if you re looking for help check out support md is this a bug report or feature request feature arbitrary key value pairs should just be discarded this allows to upgrade txt records in case a newer version of txtdirect uses new keys without breaking older versions arbitrary data however shouldn t be allowed summary v to works v works v to works v fails v to fails
| 0
|
677,122
| 23,151,777,742
|
IssuesEvent
|
2022-07-29 09:01:55
|
wasmerio/wasmer
|
https://api.github.com/repos/wasmerio/wasmer
|
closed
|
Re-add create exe to Wasmer
|
priority-high create-exe
|
This is a follow up from #2916 (it's blocked by it).
We need to re-add `create-exe` to Wasmer.
Steps:
1. Re-use the previous existing code (from C) to create the executable
2. Create a new `create-exe` that uses Zig under the hood.
3. Make zig on-create exe used with wax
5. Include the virtual filesystem into the generated executable (`python` should work) - not required for Wasmer 3.0
## Details for (2)
We need to Zig version of the C generated file, if zig is found in the system, we can use it by default, if not, we need to call a subprocess with `wax zig`.
Unknowns:
1. linker might not work the same way
## Details for (3)
We will include the virtual filesystem structure into Wasmer, so Python can be actually converted to a native executable
|
1.0
|
Re-add create exe to Wasmer - This is a follow up from #2916 (it's blocked by it).
We need to re-add `create-exe` to Wasmer.
Steps:
1. Re-use the previous existing code (from C) to create the executable
2. Create a new `create-exe` that uses Zig under the hood.
3. Make zig on-create exe used with wax
5. Include the virtual filesystem into the generated executable (`python` should work) - not required for Wasmer 3.0
## Details for (2)
We need to Zig version of the C generated file, if zig is found in the system, we can use it by default, if not, we need to call a subprocess with `wax zig`.
Unknowns:
1. linker might not work the same way
## Details for (3)
We will include the virtual filesystem structure into Wasmer, so Python can be actually converted to a native executable
|
non_process
|
re add create exe to wasmer this is a follow up from it s blocked by it we need to re add create exe to wasmer steps re use the previous existing code from c to create the executable create a new create exe that uses zig under the hood make zig on create exe used with wax include the virtual filesystem into the generated executable python should work not required for wasmer details for we need to zig version of the c generated file if zig is found in the system we can use it by default if not we need to call a subprocess with wax zig unknowns linker might not work the same way details for we will include the virtual filesystem structure into wasmer so python can be actually converted to a native executable
| 0
|
22,276
| 30,826,946,846
|
IssuesEvent
|
2023-08-01 20:54:15
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
closed
|
storage: determine what to do on chunksize=0 [GRPC]
|
api: storage priority: p2 type: process
|
Address TODOs around L1472 in grpc_client.go
Should we actually use the minimum of 256 KB here [chunksize=0] when the user indicates they want minimal memory usage? We cannot do a zero-copy, bufferless upload like HTTP/JSON can.
Must make a note to update documentation accordingly.
|
1.0
|
storage: determine what to do on chunksize=0 [GRPC] - Address TODOs around L1472 in grpc_client.go
Should we actually use the minimum of 256 KB here [chunksize=0] when the user indicates they want minimal memory usage? We cannot do a zero-copy, bufferless upload like HTTP/JSON can.
Must make a note to update documentation accordingly.
|
process
|
storage determine what to do on chunksize address todos around in grpc client go should we actually use the minimum of kb here when the user indicates they want minimal memory usage we cannot do a zero copy bufferless upload like http json can must make a note to update documentation accordingly
| 1
|
3,302
| 6,399,092,377
|
IssuesEvent
|
2017-08-04 22:42:53
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
System.ServiceProcess.ServiceController & ServiceBase
|
area-System.ServiceProcess enhancement port-to-core
|
We've had requests to port https://github.com/topshelf/topshelf to .NET Core. To do this, we would love to be able to query if we had access to the SCM and act as needed so we don't need to ship a version targeting both core & full profile.
|
1.0
|
System.ServiceProcess.ServiceController & ServiceBase - We've had requests to port https://github.com/topshelf/topshelf to .NET Core. To do this, we would love to be able to query if we had access to the SCM and act as needed so we don't need to ship a version targeting both core & full profile.
|
process
|
system serviceprocess servicecontroller servicebase we ve had requests to port to net core to do this we would love to be able to query if we had access to the scm and act as needed so we don t need to ship a version targeting both core full profile
| 1
|
751,657
| 26,252,677,466
|
IssuesEvent
|
2023-01-05 20:52:28
|
awesomemotive/affiliatewp-affiliate-area-shortcodes
|
https://api.github.com/repos/awesomemotive/affiliatewp-affiliate-area-shortcodes
|
closed
|
If AffiliateWP is not active, there is no longer a requirement admin notice
|
type-bug priority-high workflow-has-pr
|
While this addon exhibits the issue, it's likely also affecting any other addon using the same activation class. I'm opening it here because I need to do some work on this addon for the WordPress.org repo and this is going to hold us back. Let's fix it in this addon first and then push it out to other addons.
## The issue:
In the following screenshot I have just activated the Affiliate Area Shortcodes addon. Note how AffiliateWP is not activated and there is no notice that tells the user that AffiliateWP must be activated.

We used to show a message like this:

In the case of this addon, the notice worked for all versions up until `1.2` when this was added: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/issues/37
Aside from the notice not displaying there are other issues:
1. There are two `AffiliateWP_Activation` classes included. See https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/lib/affwp/class-affiliatewp-activation.php and https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/class-activation.php
2. The activation code is super convoluted. No idea why it needs to have that much code.
3. There are incorrect textdomains, likely a copy/paste error when implementing on another addon. See `affiliatewp-afgf` in the codebase
This is probably only a small part of this but this conditional is preventing the banner from ever loading: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/affiliatewp-affiliate-area-shortcodes.php#L108-L112
The first conditional with `version_compare( $affwp_version, '2.7', '>=' )` is `true` for me since I have a version higher than `2.7`. Then it tries to load the `bootstrap` method but can't because it runs on the `affwp_plugins_loaded` hook. That hook is only available within AffiliateWP, which of course is not active.
To test and at least see the banner in the above screenshot, change `affwp_plugins_loaded` to `plugins_loaded`.
So what we need to do here is:
1. Get the notice working again. It should show when AffiliateWP is not active
2. Preserve any changes added which cater to the minimum requirements (you'll see these browsing through the code)
3. Simplify the activation code, remove the duplicate class, fix the text domains
After we have thoroughly tested this we need to roll this out to the other addons. It's a super broken experience when a user installs our free addons and there's zero indication of what to do next.
|
1.0
|
If AffiliateWP is not active, there is no longer a requirement admin notice - While this addon exhibits the issue, it's likely also affecting any other addon using the same activation class. I'm opening it here because I need to do some work on this addon for the WordPress.org repo and this is going to hold us back. Let's fix it in this addon first and then push it out to other addons.
## The issue:
In the following screenshot I have just activated the Affiliate Area Shortcodes addon. Note how AffiliateWP is not activated and there is no notice that tells the user that AffiliateWP must be activated.

We used to show a message like this:

In the case of this addon, the notice worked for all versions up until `1.2` when this was added: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/issues/37
Aside from the notice not displaying there are other issues:
1. There are two `AffiliateWP_Activation` classes included. See https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/lib/affwp/class-affiliatewp-activation.php and https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/includes/class-activation.php
2. The activation code is super convoluted. No idea why it needs to have that much code.
3. There are incorrect textdomains, likely a copy/paste error when implementing on another addon. See `affiliatewp-afgf` in the codebase
This is probably only a small part of this but this conditional is preventing the banner from ever loading: https://github.com/awesomemotive/affiliatewp-affiliate-area-shortcodes/blob/master/affiliatewp-affiliate-area-shortcodes.php#L108-L112
The first conditional with `version_compare( $affwp_version, '2.7', '>=' )` is `true` for me since I have a version higher than `2.7`. Then it tries to load the `bootstrap` method but can't because it runs on the `affwp_plugins_loaded` hook. That hook is only available within AffiliateWP, which of course is not active.
To test and at least see the banner in the above screenshot, change `affwp_plugins_loaded` to `plugins_loaded`.
So what we need to do here is:
1. Get the notice working again. It should show when AffiliateWP is not active
2. Preserve any changes added which cater to the minimum requirements (you'll see these browsing through the code)
3. Simplify the activation code, remove the duplicate class, fix the text domains
After we have thoroughly tested this we need to roll this out to the other addons. It's a super broken experience when a user installs our free addons and there's zero indication of what to do next.
|
non_process
|
if affiliatewp is not active there is no longer a requirement admin notice while this addon exhibits the issue it s likely also affecting any other addon using the same activation class i m opening it here because i need to do some work on this addon for the wordpress org repo and this is going to hold us back let s fix it in this addon first and then push it out to other addons the issue in the following screenshot i have just activated the affiliate area shortcodes addon note how affiliatewp is not activated and there is no notice that tells the user that affiliatewp must be activated we used to show a message like this in the case of this addon the notice worked for all versions up until when this was added aside from the notice not displaying there are other issues there are two affiliatewp activation classes included see and the activation code is super convoluted no idea why it needs to have that much code there are incorrect textdomains likely a copy paste error when implementing on another addon see affiliatewp afgf in the codebase this is probably only a small part of this but this conditional is preventing the banner from ever loading the first conditional with version compare affwp version is true for me since i have a version higher than then it tries to load the bootstrap method but can t because it runs on the affwp plugins loaded hook that hook is only available within affiliatewp which of course is not active to test and at least see the banner in the above screenshot change affwp plugins loaded to plugins loaded so what we need to do here is get the notice working again it should show when affiliatewp is not active preserve any changes added which cater to the minimum requirements you ll see these browsing through the code simplify the activation code remove the duplicate class fix the text domains after we have thoroughly tested this we need to roll this out to the other addons it s a super broken experience when a user installs our free addons and there s zero indication of what to do next
| 0
|
1,209
| 3,711,323,307
|
IssuesEvent
|
2016-03-02 09:51:32
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
NPE in [keyref] processing (DITA-OT 2.2.2)
|
bug P1 preprocess/keyref
|
Following is the result log file.
```
[echo] com.antennahouse.i18n_index.2.3 is integrated.
init:
[echo] com.antennahouse.i18n_index.2.3 is integrated.
check-arg:
[mkdir] Created dir: D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
log-arg:
[echo] *****************************************************************
[echo] * basedir = D:\DITA-OT\dita-ot-2.2.2
[echo] * dita.dir = D:\DITA-OT\dita-ot-2.2.2
[echo] * transtype = pdf5.ml
[echo] * tempdir = D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
[echo] * outputdir = D:\DITA-OT\dita-ot-2.2.2\out
[echo] * clean.temp = true
[echo] * DITA-OT version = 2.2.2
[echo] * XML parser = Xerces
[echo] * XSLT processor = Saxon
[echo] * collator = ICU
[echo] *****************************************************************
[echo] #Ant properties
[echo] #Wed Mar 02 16:53:01 GMT+09:00 2016
[echo] args.grammar.cache=yes
[echo] args.input=samples/sample_en/sample_en.ditamap
[echo] args.logdir=D\:\\DITA-OT\\dita-ot-2.2.2\\out
[echo] args.xml.systemid.set=yes
[echo] dita.dir=D\:\\DITA-OT\\dita-ot-2.2.2
[echo] dita.plugin.com.antennahouse.dita.dita13.doctypes.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.dita.dita13.doctypes
[echo] dita.plugin.com.antennahouse.i18n_index.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.i18n_index.2.3
[echo] dita.plugin.com.antennahouse.pdf5.ml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.pdf5.ml
[echo] dita.plugin.com.antennahouse.samples.form.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.samples.form
[echo] dita.plugin.com.sophos.tocjs.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.sophos.tocjs
[echo] dita.plugin.org.dita.base.dir=D\:\\DITA-OT\\dita-ot-2.2.2
[echo] dita.plugin.org.dita.docbook.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.docbook
[echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsecontent
[echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsehelp
[echo] dita.plugin.org.dita.html5.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.html5
[echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.htmlhelp
[echo] dita.plugin.org.dita.javahelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.javahelp
[echo] dita.plugin.org.dita.odt.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.odt
[echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.axf
[echo] dita.plugin.org.dita.pdf2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2
[echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.fop
[echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.xep
[echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.dita11
[echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.eclipsemap
[echo] dita.plugin.org.dita.troff.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.troff
[echo] dita.plugin.org.dita.wordrtf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.wordrtf
[echo] dita.plugin.org.dita.xhtml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.xhtml
[echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_2
[echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_3
[echo] dita.temp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\temp\\temp20160302165301227
[echo] preprocess.copy-html.skip=true
[echo] preprocess.copy-image.skip=true
[echo] *****************************************************************
build-init:
preprocess.init:
[echo] *****************************************************************
[echo] * input = samples/sample_en/sample_en.ditamap
[echo] *****************************************************************
gen-list:
[gen-list] Using Xerces grammar pool for DTD and schema caching.
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml
[gen-list] Serializing job specification
debug-filter:
[filter] Using Xerces grammar pool for DTD and schema caching.
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\outditafiles.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamap.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\image.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\flagimage.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\html.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subjectscheme.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\copytosource.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subtargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\user.input.file.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
mapref-check:
mapref:
[mapref] Transforming into D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
[mapref] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\preprocess\mapref.xsl
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\m_keydef.ditamap
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\glossary_en.ditamap
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\sample_en.ditamap
branch-filter:
[branch-filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap
copy-image:
copy-html:
copy-flag-check:
copy-flag:
copy-files:
keyref:
[keyref] Reading file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap
[keyref] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/topics-en/p_longquoteref.xml
BUILD FAILED
D:\DITA-OT\dita-ot-2.2.2\build.xml:41: The following error occurred while executing this line:
D:\DITA-OT\dita-ot-2.2.2\plugins\org.dita.base\build_preprocess.xml:270: java.lang.NullPointerException
at org.dita.dost.writer.KeyrefPaser.processElement(KeyrefPaser.java:374)
at org.dita.dost.writer.KeyrefPaser.startElement(KeyrefPaser.java:364)
at org.xml.sax.helpers.XMLFilterImpl.startElement(Unknown Source)
at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:62)
at org.dita.dost.writer.ConkeyrefFilter.startElement(ConkeyrefFilter.java:89)
at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source)
at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source)
at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(Unknown Source)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:404)
at net.sf.saxon.event.Sender.send(Sender.java:193)
at net.sf.saxon.IdentityTransformer.transform(IdentityTransformer.java:30)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:259)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:219)
at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:234)
at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:93)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.Project.executeTarget(Project.java:1364)
at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.Main.runBuild(Main.java:851)
at org.apache.tools.ant.Main.startAnt(Main.java:235)
at org.apache.tools.ant.launch.Launcher.run(Launcher.java:280)
at org.apache.tools.ant.launch.Launcher.main(Launcher.java:109)
Total time: 4 seconds
```
It seems that p_longquoteref.xml has no validation errors.
I attached used DITA instance.
[sample_en.zip](https://github.com/dita-ot/dita-ot/files/154305/sample_en.zip)
|
1.0
|
NPE in [keyref] processing (DITA-OT 2.2.2) - Following is the result log file.
```
[echo] com.antennahouse.i18n_index.2.3 is integrated.
init:
[echo] com.antennahouse.i18n_index.2.3 is integrated.
check-arg:
[mkdir] Created dir: D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
log-arg:
[echo] *****************************************************************
[echo] * basedir = D:\DITA-OT\dita-ot-2.2.2
[echo] * dita.dir = D:\DITA-OT\dita-ot-2.2.2
[echo] * transtype = pdf5.ml
[echo] * tempdir = D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
[echo] * outputdir = D:\DITA-OT\dita-ot-2.2.2\out
[echo] * clean.temp = true
[echo] * DITA-OT version = 2.2.2
[echo] * XML parser = Xerces
[echo] * XSLT processor = Saxon
[echo] * collator = ICU
[echo] *****************************************************************
[echo] #Ant properties
[echo] #Wed Mar 02 16:53:01 GMT+09:00 2016
[echo] args.grammar.cache=yes
[echo] args.input=samples/sample_en/sample_en.ditamap
[echo] args.logdir=D\:\\DITA-OT\\dita-ot-2.2.2\\out
[echo] args.xml.systemid.set=yes
[echo] dita.dir=D\:\\DITA-OT\\dita-ot-2.2.2
[echo] dita.plugin.com.antennahouse.dita.dita13.doctypes.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.dita.dita13.doctypes
[echo] dita.plugin.com.antennahouse.i18n_index.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.i18n_index.2.3
[echo] dita.plugin.com.antennahouse.pdf5.ml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.pdf5.ml
[echo] dita.plugin.com.antennahouse.samples.form.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.antennahouse.samples.form
[echo] dita.plugin.com.sophos.tocjs.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\com.sophos.tocjs
[echo] dita.plugin.org.dita.base.dir=D\:\\DITA-OT\\dita-ot-2.2.2
[echo] dita.plugin.org.dita.docbook.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.docbook
[echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsecontent
[echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.eclipsehelp
[echo] dita.plugin.org.dita.html5.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.html5
[echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.htmlhelp
[echo] dita.plugin.org.dita.javahelp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.javahelp
[echo] dita.plugin.org.dita.odt.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.odt
[echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.axf
[echo] dita.plugin.org.dita.pdf2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2
[echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.fop
[echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.pdf2.xep
[echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.dita11
[echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.specialization.eclipsemap
[echo] dita.plugin.org.dita.troff.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.troff
[echo] dita.plugin.org.dita.wordrtf.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.wordrtf
[echo] dita.plugin.org.dita.xhtml.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.dita.xhtml
[echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_2
[echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\plugins\\org.oasis-open.dita.v1_3
[echo] dita.temp.dir=D\:\\DITA-OT\\dita-ot-2.2.2\\temp\\temp20160302165301227
[echo] preprocess.copy-html.skip=true
[echo] preprocess.copy-image.skip=true
[echo] *****************************************************************
build-init:
preprocess.init:
[echo] *****************************************************************
[echo] * input = samples/sample_en/sample_en.ditamap
[echo] *****************************************************************
gen-list:
[gen-list] Using Xerces grammar pool for DTD and schema caching.
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml
[gen-list] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml
[gen-list] Serializing job specification
debug-filter:
[filter] Using Xerces grammar pool for DTD and schema caching.
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_programmingelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longdescref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/glossary_en.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_preface.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_typographic.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_table.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/dita_sample.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_reference_to_no_print.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_tys125f.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_miscellaneouselements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/sample_en.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle1.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_fig.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_specializationelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLFO.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_xref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_note.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_properties.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLandHTML.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_utilityelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DTD.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_DOM.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_complecated_index_example.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XMLSchema.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dir_attribute.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/m_keydef.ditamap
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_SAX.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/r_xslt.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_abstract.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_dita12.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_no_print.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_bodyelements.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_sectiondiv.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_weirdtitle2.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/gloss_XSLT.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_longquoteref.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/c_test_introduction.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/t_logging_in_to_client.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_backmatter.xml
[filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/samples/sample_en/topics-en/p_softwareelements.xml
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\outditafiles.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamap.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\image.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\flagimage.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\html.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subjectscheme.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\copytosource.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\subtargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\user.input.file.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conref.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\hrefditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditatopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\fullditamapandtopic.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\conreftargets.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\canditopics.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
[job-helper] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\.job.xml to D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\resourceonly.list
[job-helper] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\job-helper.xsl
mapref-check:
mapref:
[mapref] Transforming into D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227
[mapref] Loading stylesheet D:\DITA-OT\dita-ot-2.2.2\xsl\preprocess\mapref.xsl
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\m_keydef.ditamap
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\topics-en\glossary_en.ditamap
[mapref] Processing D:\DITA-OT\dita-ot-2.2.2\temp\temp20160302165301227\sample_en.ditamap
branch-filter:
[branch-filter] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap
copy-image:
copy-html:
copy-flag-check:
copy-flag:
copy-files:
keyref:
[keyref] Reading file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/sample_en.ditamap
[keyref] Processing file:/D:/DITA-OT/dita-ot-2.2.2/temp/temp20160302165301227/topics-en/p_longquoteref.xml
BUILD FAILED
D:\DITA-OT\dita-ot-2.2.2\build.xml:41: The following error occurred while executing this line:
D:\DITA-OT\dita-ot-2.2.2\plugins\org.dita.base\build_preprocess.xml:270: java.lang.NullPointerException
at org.dita.dost.writer.KeyrefPaser.processElement(KeyrefPaser.java:374)
at org.dita.dost.writer.KeyrefPaser.startElement(KeyrefPaser.java:364)
at org.xml.sax.helpers.XMLFilterImpl.startElement(Unknown Source)
at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:62)
at org.dita.dost.writer.ConkeyrefFilter.startElement(ConkeyrefFilter.java:89)
at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source)
at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source)
at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(Unknown Source)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:404)
at net.sf.saxon.event.Sender.send(Sender.java:193)
at net.sf.saxon.IdentityTransformer.transform(IdentityTransformer.java:30)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:259)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:219)
at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:234)
at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:93)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.Project.executeTarget(Project.java:1364)
at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.Main.runBuild(Main.java:851)
at org.apache.tools.ant.Main.startAnt(Main.java:235)
at org.apache.tools.ant.launch.Launcher.run(Launcher.java:280)
at org.apache.tools.ant.launch.Launcher.main(Launcher.java:109)
Total time: 4 seconds
```
It seems that p_longquoteref.xml has no validation errors.
I attached used DITA instance.
[sample_en.zip](https://github.com/dita-ot/dita-ot/files/154305/sample_en.zip)
|
process
|
npe in processing dita ot following is the result log file com antennahouse index is integrated init com antennahouse index is integrated check arg created dir d dita ot dita ot temp log arg basedir d dita ot dita ot dita dir d dita ot dita ot transtype ml tempdir d dita ot dita ot temp outputdir d dita ot dita ot out clean temp true dita ot version xml parser xerces xslt processor saxon collator icu ant properties wed mar gmt args grammar cache yes args input samples sample en sample en ditamap args logdir d dita ot dita ot out args xml systemid set yes dita dir d dita ot dita ot dita plugin com antennahouse dita doctypes dir d dita ot dita ot plugins com antennahouse dita doctypes dita plugin com antennahouse index dir d dita ot dita ot plugins com antennahouse index dita plugin com antennahouse ml dir d dita ot dita ot plugins com antennahouse ml dita plugin com antennahouse samples form dir d dita ot dita ot plugins com antennahouse samples form dita plugin com sophos tocjs dir d dita ot dita ot plugins com sophos tocjs dita plugin org dita base dir d dita ot dita ot dita plugin org dita docbook dir d dita ot dita ot plugins org dita docbook dita plugin org dita eclipsecontent dir d dita ot dita ot plugins org dita eclipsecontent dita plugin org dita eclipsehelp dir d dita ot dita ot plugins org dita eclipsehelp dita plugin org dita dir d dita ot dita ot plugins org dita dita plugin org dita htmlhelp dir d dita ot dita ot plugins org dita htmlhelp dita plugin org dita javahelp dir d dita ot dita ot plugins org dita javahelp dita plugin org dita odt dir d dita ot dita ot plugins org dita odt dita plugin org dita axf dir d dita ot dita ot plugins org dita axf dita plugin org dita dir d dita ot dita ot plugins org dita dita plugin org dita fop dir d dita ot dita ot plugins org dita fop dita plugin org dita xep dir d dita ot dita ot plugins org dita xep dita plugin org dita specialization dir d dita ot dita ot plugins org dita specialization dita plugin org dita specialization eclipsemap dir d dita ot dita ot plugins org dita specialization eclipsemap dita plugin org dita troff dir d dita ot dita ot plugins org dita troff dita plugin org dita wordrtf dir d dita ot dita ot plugins org dita wordrtf dita plugin org dita xhtml dir d dita ot dita ot plugins org dita xhtml dita plugin org oasis open dita dir d dita ot dita ot plugins org oasis open dita dita plugin org oasis open dita dir d dita ot dita ot plugins org oasis open dita dita temp dir d dita ot dita ot temp preprocess copy html skip true preprocess copy image skip true build init preprocess init input samples sample en sample en ditamap gen list using xerces grammar pool for dtd and schema caching processing file d dita ot dita ot samples sample en sample en ditamap processing file d dita ot dita ot samples sample en topics en dita sample xml processing file d dita ot dita ot samples sample en topics en c preface xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en c test introduction xml processing file d dita ot dita ot samples sample en topics en p abstract xml processing file d dita ot dita ot samples sample en topics en p xref xml processing file d dita ot dita ot samples sample en topics en p note xml processing file d dita ot dita ot samples sample en topics en p bodyelements xml processing file d dita ot dita ot samples sample en topics en p miscellaneouselements xml processing file d dita ot dita ot samples sample en topics en p specializationelements xml processing file d dita ot dita ot samples sample en topics en p typographic xml processing file d dita ot dita ot samples sample en topics en p programmingelements xml processing file d dita ot dita ot samples sample en topics en p softwareelements xml processing file d dita ot dita ot samples sample en topics en p utilityelements xml processing file d dita ot dita ot samples sample en topics en p fig xml processing file d dita ot dita ot samples sample en topics en p table xml processing file d dita ot dita ot samples sample en topics en r properties xml processing file d dita ot dita ot samples sample en topics en t logging in to client xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p longdescref xml processing file d dita ot dita ot samples sample en topics en p longquoteref xml processing file d dita ot dita ot samples sample en topics en m keydef ditamap processing file d dita ot dita ot samples sample en topics en p sectiondiv xml processing file d dita ot dita ot samples sample en topics en p reference to no print xml processing file d dita ot dita ot samples sample en topics en p complecated index example xml processing file d dita ot dita ot samples sample en topics en r xslt xml processing file d dita ot dita ot samples sample en topics en p dir attribute xml processing file d dita ot dita ot samples sample en topics en p backmatter xml processing file d dita ot dita ot samples sample en topics en glossary en ditamap processing file d dita ot dita ot samples sample en topics en p no print xml processing file d dita ot dita ot samples sample en topics en r xml processing file d dita ot dita ot samples sample en topics en gloss xslt xml processing file d dita ot dita ot samples sample en topics en gloss xslfo xml processing file d dita ot dita ot samples sample en topics en gloss xmlandhtml xml processing file d dita ot dita ot samples sample en topics en gloss xmlschema xml processing file d dita ot dita ot samples sample en topics en gloss dtd xml processing file d dita ot dita ot samples sample en topics en gloss dom xml processing file d dita ot dita ot samples sample en topics en gloss sax xml serializing job specification debug filter using xerces grammar pool for dtd and schema caching processing file d dita ot dita ot samples sample en topics en p programmingelements xml processing file d dita ot dita ot samples sample en topics en p longdescref xml processing file d dita ot dita ot samples sample en topics en glossary en ditamap processing file d dita ot dita ot samples sample en topics en c preface xml processing file d dita ot dita ot samples sample en topics en p typographic xml processing file d dita ot dita ot samples sample en topics en p table xml processing file d dita ot dita ot samples sample en topics en dita sample xml processing file d dita ot dita ot samples sample en topics en p reference to no print xml processing file d dita ot dita ot samples sample en topics en r xml processing file d dita ot dita ot samples sample en topics en p miscellaneouselements xml processing file d dita ot dita ot samples sample en sample en ditamap processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p fig xml processing file d dita ot dita ot samples sample en topics en p specializationelements xml processing file d dita ot dita ot samples sample en topics en gloss xslfo xml processing file d dita ot dita ot samples sample en topics en p xref xml processing file d dita ot dita ot samples sample en topics en p note xml processing file d dita ot dita ot samples sample en topics en r properties xml processing file d dita ot dita ot samples sample en topics en gloss xmlandhtml xml processing file d dita ot dita ot samples sample en topics en p utilityelements xml processing file d dita ot dita ot samples sample en topics en gloss dtd xml processing file d dita ot dita ot samples sample en topics en gloss dom xml processing file d dita ot dita ot samples sample en topics en p complecated index example xml processing file d dita ot dita ot samples sample en topics en gloss xmlschema xml processing file d dita ot dita ot samples sample en topics en p dir attribute xml processing file d dita ot dita ot samples sample en topics en m keydef ditamap processing file d dita ot dita ot samples sample en topics en gloss sax xml processing file d dita ot dita ot samples sample en topics en r xslt xml processing file d dita ot dita ot samples sample en topics en p abstract xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en p no print xml processing file d dita ot dita ot samples sample en topics en p bodyelements xml processing file d dita ot dita ot samples sample en topics en p sectiondiv xml processing file d dita ot dita ot samples sample en topics en p xml processing file d dita ot dita ot samples sample en topics en gloss xslt xml processing file d dita ot dita ot samples sample en topics en p longquoteref xml processing file d dita ot dita ot samples sample en topics en c test introduction xml processing file d dita ot dita ot samples sample en topics en t logging in to client xml processing file d dita ot dita ot samples sample en topics en p backmatter xml processing file d dita ot dita ot samples sample en topics en p softwareelements xml processing d dita ot dita ot temp job xml to d dita ot dita ot temp outditafiles list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamapandtopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamap list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hrefditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conref list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp image list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp flagimage list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp html list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp canditopics list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp subjectscheme list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp copytosource list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp subtargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp resourceonly list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp user input file list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conref list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp hrefditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditatopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp fullditamapandtopic list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp conreftargets list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp canditopics list loading stylesheet d dita ot dita ot xsl job helper xsl processing d dita ot dita ot temp job xml to d dita ot dita ot temp resourceonly list loading stylesheet d dita ot dita ot xsl job helper xsl mapref check mapref transforming into d dita ot dita ot temp loading stylesheet d dita ot dita ot xsl preprocess mapref xsl processing d dita ot dita ot temp topics en m keydef ditamap processing d dita ot dita ot temp topics en glossary en ditamap processing d dita ot dita ot temp sample en ditamap branch filter processing file d dita ot dita ot temp sample en ditamap copy image copy html copy flag check copy flag copy files keyref reading file d dita ot dita ot temp sample en ditamap processing file d dita ot dita ot temp topics en p longquoteref xml build failed d dita ot dita ot build xml the following error occurred while executing this line d dita ot dita ot plugins org dita base build preprocess xml java lang nullpointerexception at org dita dost writer keyrefpaser processelement keyrefpaser java at org dita dost writer keyrefpaser startelement keyrefpaser java at org xml sax helpers xmlfilterimpl startelement unknown source at org dita dost writer topicfragmentfilter startelement topicfragmentfilter java at org dita dost writer conkeyreffilter startelement conkeyreffilter java at org apache xerces parsers abstractsaxparser startelement unknown source at org apache xerces parsers abstractxmldocumentparser emptyelement unknown source at org apache xerces impl xmlnsdocumentscannerimpl scanstartelement unknown source at org apache xerces impl xmldocumentfragmentscannerimpl fragmentcontentdispatcher dispatch unknown source at org apache xerces impl xmldocumentfragmentscannerimpl scandocument unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers abstractsaxparser parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at net sf saxon event sender sendsaxsource sender java at net sf saxon event sender send sender java at net sf saxon identitytransformer transform identitytransformer java at org dita dost util xmlutils transform xmlutils java at org dita dost util xmlutils transform xmlutils java at org dita dost module keyrefmodule processfile keyrefmodule java at org dita dost module keyrefmodule execute keyrefmodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant project executetarget project java at org apache tools ant helper defaultexecutor executetargets defaultexecutor java at org apache tools ant project executetargets project java at org apache tools ant main runbuild main java at org apache tools ant main startant main java at org apache tools ant launch launcher run launcher java at org apache tools ant launch launcher main launcher java total time seconds it seems that p longquoteref xml has no validation errors i attached used dita instance
| 1
|
2,608
| 5,367,284,967
|
IssuesEvent
|
2017-02-22 03:27:04
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[subtitles] [eng] Mélenchon - FAQ 30/11/16
|
Language: English Process: [6] Approved
|
# Video title
MÉLENCHON - FAQ - 6E RÉPUBLIQUE, CANNABIS, BIO, YOUTUBE, SNOWDEN, ASSANGE…
# URL
https://www.youtube.com/watch?v=GkPZ3sIhMPg
# Youtube subtitles language
Anglais
# Duration
48:34
# Subtitles URL
https://www.youtube.com/timedtext_editor?bl=vmp&action_mde_edit_form=1&ui=hd&v=GkPZ3sIhMPg&lang=en&ref=player&tab=captions
|
1.0
|
[subtitles] [eng] Mélenchon - FAQ 30/11/16 - # Video title
MÉLENCHON - FAQ - 6E RÉPUBLIQUE, CANNABIS, BIO, YOUTUBE, SNOWDEN, ASSANGE…
# URL
https://www.youtube.com/watch?v=GkPZ3sIhMPg
# Youtube subtitles language
Anglais
# Duration
48:34
# Subtitles URL
https://www.youtube.com/timedtext_editor?bl=vmp&action_mde_edit_form=1&ui=hd&v=GkPZ3sIhMPg&lang=en&ref=player&tab=captions
|
process
|
mélenchon faq video title mélenchon faq république cannabis bio youtube snowden assange… url youtube subtitles language anglais duration subtitles url
| 1
|
19,080
| 25,126,461,506
|
IssuesEvent
|
2022-11-09 12:07:52
|
prisma/prisma-engines
|
https://api.github.com/repos/prisma/prisma-engines
|
opened
|
Only use buildjet action runner where needed
|
process/candidate kind/tech team/schema team/client
|
We use buildjet runner in multiple actions https://github.com/prisma/prisma-engines/search?q=buildjet
Buildjet is great but we should only use it where it adds value, when it makes tests a lot faster.
Overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue.
Example situation
https://prisma-company.slack.com/archives/C1FPU5FPT/p1667986934989559
Where a ~3 min long action was waiting for ~1h20
<img width="725" alt="Screenshot 2022-11-09 at 12 18 20" src="https://user-images.githubusercontent.com/1328733/200825862-af63c693-150a-4f33-9e32-43529a07bf3b.png">
https://github.com/prisma/prisma-engines/actions/runs/3426620655
- [ ] We should switch to GitHub linux runner for Formatting https://github.com/prisma/prisma-engines/blob/main/.github/workflows/formatting.yml
- [ ] Check https://github.com/prisma/prisma-engines/search?q=buildjet and decide if we want to do the switch in other places as well
|
1.0
|
Only use buildjet action runner where needed - We use buildjet runner in multiple actions https://github.com/prisma/prisma-engines/search?q=buildjet
Buildjet is great but we should only use it where it adds value, when it makes tests a lot faster.
Overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue.
Example situation
https://prisma-company.slack.com/archives/C1FPU5FPT/p1667986934989559
Where a ~3 min long action was waiting for ~1h20
<img width="725" alt="Screenshot 2022-11-09 at 12 18 20" src="https://user-images.githubusercontent.com/1328733/200825862-af63c693-150a-4f33-9e32-43529a07bf3b.png">
https://github.com/prisma/prisma-engines/actions/runs/3426620655
- [ ] We should switch to GitHub linux runner for Formatting https://github.com/prisma/prisma-engines/blob/main/.github/workflows/formatting.yml
- [ ] Check https://github.com/prisma/prisma-engines/search?q=buildjet and decide if we want to do the switch in other places as well
|
process
|
only use buildjet action runner where needed we use buildjet runner in multiple actions buildjet is great but we should only use it where it adds value when it makes tests a lot faster overusing it means that we are more likely to run into concurrency limits which defeats the purpose of making the test faster because instead they will wait for a long time in a queue example situation where a min long action was waiting for img width alt screenshot at src we should switch to github linux runner for formatting check and decide if we want to do the switch in other places as well
| 1
|
13,804
| 16,564,730,393
|
IssuesEvent
|
2021-05-29 06:47:18
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Details Threshold has lost functionality
|
bug: pending scope: image processing
|
**Description** The Details Threshold appears inoperable with DT 3.5+2360. The Details Threshold option was working with my copy of DT 3.5+2226
**To Reproduce**
1. Activate a module such as local contrast and enable the parametric mask option and activate the mask. The display will become entirely yellow
2. Details Threshold slider has no effect on the mask when moved right or left
3. I also turned off the mask verified that the slider has no effect on the image, so it's not just an issue with the mask
**Screenshots** See attached.
1. Original image with DT 3.5+2360
2. Parametric mask activated with mask displayed
3. Mask display with Details Threshold moved left
4. Mask display with Details Threshold moved right
5. Correctly operating Details Threshold using DT 3.5+2226
**Which commit introduced the error** As mentioned, this seems to be a new error introduced since DT 3.5+2226
**Platform**
darktable version : 3.5.0+2360~ge734491a9
OS : Windows 10
Memory : 8 GB
Graphics card : AMD A8-7410 APU with AMD Radeon R5 Graphics, 2.20 Ghz
Graphics driver : Unk
OpenCL installed : Yes
OpenCL activated : No (Open CL always causes my system to lock up - different issue)
Xorg : Unk
Desktop : HP Pavalion
GTK+ : Unk
gcc : Unk
cflags : Unk
CMAKE_BUILD_TYPE : Unk





|
1.0
|
Details Threshold has lost functionality - **Description** The Details Threshold appears inoperable with DT 3.5+2360. The Details Threshold option was working with my copy of DT 3.5+2226
**To Reproduce**
1. Activate a module such as local contrast and enable the parametric mask option and activate the mask. The display will become entirely yellow
2. Details Threshold slider has no effect on the mask when moved right or left
3. I also turned off the mask verified that the slider has no effect on the image, so it's not just an issue with the mask
**Screenshots** See attached.
1. Original image with DT 3.5+2360
2. Parametric mask activated with mask displayed
3. Mask display with Details Threshold moved left
4. Mask display with Details Threshold moved right
5. Correctly operating Details Threshold using DT 3.5+2226
**Which commit introduced the error** As mentioned, this seems to be a new error introduced since DT 3.5+2226
**Platform**
darktable version : 3.5.0+2360~ge734491a9
OS : Windows 10
Memory : 8 GB
Graphics card : AMD A8-7410 APU with AMD Radeon R5 Graphics, 2.20 Ghz
Graphics driver : Unk
OpenCL installed : Yes
OpenCL activated : No (Open CL always causes my system to lock up - different issue)
Xorg : Unk
Desktop : HP Pavalion
GTK+ : Unk
gcc : Unk
cflags : Unk
CMAKE_BUILD_TYPE : Unk





|
process
|
details threshold has lost functionality description the details threshold appears inoperable with dt the details threshold option was working with my copy of dt to reproduce activate a module such as local contrast and enable the parametric mask option and activate the mask the display will become entirely yellow details threshold slider has no effect on the mask when moved right or left i also turned off the mask verified that the slider has no effect on the image so it s not just an issue with the mask screenshots see attached original image with dt parametric mask activated with mask displayed mask display with details threshold moved left mask display with details threshold moved right correctly operating details threshold using dt which commit introduced the error as mentioned this seems to be a new error introduced since dt platform darktable version os windows memory gb graphics card amd apu with amd radeon graphics ghz graphics driver unk opencl installed yes opencl activated no open cl always causes my system to lock up different issue xorg unk desktop hp pavalion gtk unk gcc unk cflags unk cmake build type unk
| 1
|
172,619
| 13,325,342,247
|
IssuesEvent
|
2020-08-27 09:49:08
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: tpcc/w=100/nodes=3/chaos=true failed
|
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
|
[(roachtest).tpcc/w=100/nodes=3/chaos=true failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=buildLog) on [release-19.2@603c352969903579f78c7a580afd729bb198de22](https://github.com/cockroachdb/cockroach/commits/603c352969903579f78c7a580afd729bb198de22):
```
| 40.0s 0 15.0 10.6 71.3 113.2 192.9 192.9 orderStatus
| 40.0s 0 124.0 106.0 302.0 1140.9 2147.5 2818.6 payment
| 40.0s 0 10.0 10.5 56.6 1610.6 1610.6 1610.6 stockLevel
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 41.0s 0 23.0 10.8 302.0 503.3 570.4 570.4 delivery
| 41.0s 0 145.0 106.9 486.5 939.5 1409.3 1409.3 newOrder
| 41.0s 0 17.0 10.8 48.2 113.2 134.2 134.2 orderStatus
| 41.0s 0 142.0 106.9 302.0 1342.2 3221.2 3758.1 payment
| 41.0s 0 16.0 10.7 54.5 302.0 318.8 318.8 stockLevel
| 42.0s 0 16.0 10.9 453.0 838.9 838.9 838.9 delivery
| 42.0s 0 130.0 107.5 520.1 939.5 1476.4 1610.6 newOrder
| 42.0s 0 7.0 10.7 75.5 151.0 151.0 151.0 orderStatus
| 42.0s 0 112.0 107.0 369.1 906.0 1275.1 1677.7 payment
| 42.0s 0 11.0 10.7 88.1 209.7 260.0 260.0 stockLevel
| 43.0s 0 14.0 11.0 402.7 838.9 1275.1 1275.1 delivery
| 43.0s 0 129.0 108.0 570.4 1073.7 1275.1 1610.6 newOrder
| 43.0s 0 11.0 10.7 54.5 130.0 167.8 167.8 orderStatus
| 43.0s 0 123.0 107.4 369.1 1476.4 1811.9 2818.6 payment
| 43.0s 0 11.0 10.7 134.2 469.8 1677.7 1677.7 stockLevel
Wraps: (5) exit status 30
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2456,tpcc.go:168,tpcc.go:312,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452
| main.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:168
| main.registerTPCC.func5
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:312
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5222
| runtime.main
| /usr/local/go/src/runtime/proc.go:190
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (6) Goexit() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/tpcc/w=100/nodes=3/chaos=true](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true)
Related:
- #48199 roachtest: tpcc/w=100/nodes=3/chaos=true failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpcc%2Fw%3D100%2Fnodes%3D3%2Fchaos%3Dtrue.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: tpcc/w=100/nodes=3/chaos=true failed - [(roachtest).tpcc/w=100/nodes=3/chaos=true failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=buildLog) on [release-19.2@603c352969903579f78c7a580afd729bb198de22](https://github.com/cockroachdb/cockroach/commits/603c352969903579f78c7a580afd729bb198de22):
```
| 40.0s 0 15.0 10.6 71.3 113.2 192.9 192.9 orderStatus
| 40.0s 0 124.0 106.0 302.0 1140.9 2147.5 2818.6 payment
| 40.0s 0 10.0 10.5 56.6 1610.6 1610.6 1610.6 stockLevel
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 41.0s 0 23.0 10.8 302.0 503.3 570.4 570.4 delivery
| 41.0s 0 145.0 106.9 486.5 939.5 1409.3 1409.3 newOrder
| 41.0s 0 17.0 10.8 48.2 113.2 134.2 134.2 orderStatus
| 41.0s 0 142.0 106.9 302.0 1342.2 3221.2 3758.1 payment
| 41.0s 0 16.0 10.7 54.5 302.0 318.8 318.8 stockLevel
| 42.0s 0 16.0 10.9 453.0 838.9 838.9 838.9 delivery
| 42.0s 0 130.0 107.5 520.1 939.5 1476.4 1610.6 newOrder
| 42.0s 0 7.0 10.7 75.5 151.0 151.0 151.0 orderStatus
| 42.0s 0 112.0 107.0 369.1 906.0 1275.1 1677.7 payment
| 42.0s 0 11.0 10.7 88.1 209.7 260.0 260.0 stockLevel
| 43.0s 0 14.0 11.0 402.7 838.9 1275.1 1275.1 delivery
| 43.0s 0 129.0 108.0 570.4 1073.7 1275.1 1610.6 newOrder
| 43.0s 0 11.0 10.7 54.5 130.0 167.8 167.8 orderStatus
| 43.0s 0 123.0 107.4 369.1 1476.4 1811.9 2818.6 payment
| 43.0s 0 11.0 10.7 134.2 469.8 1677.7 1677.7 stockLevel
Wraps: (5) exit status 30
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2456,tpcc.go:168,tpcc.go:312,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452
| main.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:168
| main.registerTPCC.func5
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:312
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5222
| runtime.main
| /usr/local/go/src/runtime/proc.go:190
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (6) Goexit() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/tpcc/w=100/nodes=3/chaos=true](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928490&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true)
Related:
- #48199 roachtest: tpcc/w=100/nodes=3/chaos=true failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpcc%2Fw%3D100%2Fnodes%3D3%2Fchaos%3Dtrue.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_process
|
roachtest tpcc w nodes chaos true failed on orderstatus payment stocklevel elapsed errors ops sec inst ops sec cum ms ms ms pmax ms delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go tpcc go tpcc go test runner go monitor failure monitor task failed goexit was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runtpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main registertpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps goexit was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring more artifacts related roachtest tpcc w nodes chaos true failed powered by
| 0
|
558,933
| 16,544,703,664
|
IssuesEvent
|
2021-05-27 21:54:06
|
lalitpagaria/obsei
|
https://api.github.com/repos/lalitpagaria/obsei
|
closed
|
Add Google news and website crawler as Source
|
high priority observer
|
Idea to add Google News as Source.
Google News provide RSS feed and query support hence it is easy to crawl it.
RSS link -
```
https://news.google.com/rss/search?q=[INPUT]
```
For now just add GoogleNews as source later we can add few other news sources.
Google RSS feed give title, headlight, date and url. So inorder to fetch full article we need to use another library like https://github.com/adbar/trafilatura
|
1.0
|
Add Google news and website crawler as Source - Idea to add Google News as Source.
Google News provide RSS feed and query support hence it is easy to crawl it.
RSS link -
```
https://news.google.com/rss/search?q=[INPUT]
```
For now just add GoogleNews as source later we can add few other news sources.
Google RSS feed give title, headlight, date and url. So inorder to fetch full article we need to use another library like https://github.com/adbar/trafilatura
|
non_process
|
add google news and website crawler as source idea to add google news as source google news provide rss feed and query support hence it is easy to crawl it rss link for now just add googlenews as source later we can add few other news sources google rss feed give title headlight date and url so inorder to fetch full article we need to use another library like
| 0
|
22,279
| 30,831,085,346
|
IssuesEvent
|
2023-08-02 01:48:50
|
mmattDonk/AI-TTS-Donations
|
https://api.github.com/repos/mmattDonk/AI-TTS-Donations
|
closed
|
[SOLP-14] "Crossing Paths"
|
@solrock/processor processor Bug Low priority
|
basically the \`backend\` server got that adam-driver message, and maybe a Dog_Pirate TTS was processing on the \`processor\` server. so adam-driver got processed first on the same \`processor\` server, so Dog_Pirate's Overlay ID was still on the \`processor\` server, however adam-driver got saved first, so it pushed the adam-driver message to Dog_Pirate's overlay, because their overlay ID was still there, and somehow xBunniBee's overlay ID just got lost

<sub>From [SyncLinear.com](https://synclinear.com) | [SOLP-14](https://linear.app/donk/issue/SOLP-14/crossing-paths)</sub>
|
2.0
|
[SOLP-14] "Crossing Paths" - basically the \`backend\` server got that adam-driver message, and maybe a Dog_Pirate TTS was processing on the \`processor\` server. so adam-driver got processed first on the same \`processor\` server, so Dog_Pirate's Overlay ID was still on the \`processor\` server, however adam-driver got saved first, so it pushed the adam-driver message to Dog_Pirate's overlay, because their overlay ID was still there, and somehow xBunniBee's overlay ID just got lost

<sub>From [SyncLinear.com](https://synclinear.com) | [SOLP-14](https://linear.app/donk/issue/SOLP-14/crossing-paths)</sub>
|
process
|
crossing paths basically the backend server got that adam driver message and maybe a dog pirate tts was processing on the processor server so adam driver got processed first on the same processor server so dog pirate s overlay id was still on the processor server however adam driver got saved first so it pushed the adam driver message to dog pirate s overlay because their overlay id was still there and somehow xbunnibee s overlay id just got lost from
| 1
|
22,559
| 31,777,927,522
|
IssuesEvent
|
2023-09-12 15:27:18
|
IMAP-Science-Operations-Center/imap_processing
|
https://api.github.com/repos/IMAP-Science-Operations-Center/imap_processing
|
opened
|
[L5] Produce L1B CDF formatted data products for Ultra
|
IMAP-Ultra L1 Requirement: Level 5 Untested Parent Req:SDC Execution of L1 Processing Software
|
### Summary of the L5 requirement
The SDC needs to produce L1B CDF formatted data products for the Ultra instrument. The data products for L1B can be found in the Ultra algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams
### Parent requirement
<!DOCTYPE html>
SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3
-- | -- | -- | -- | -- | -- | -- | -- | --
### Tasks
- [ ]
|
1.0
|
[L5] Produce L1B CDF formatted data products for Ultra - ### Summary of the L5 requirement
The SDC needs to produce L1B CDF formatted data products for the Ultra instrument. The data products for L1B can be found in the Ultra algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams
### Parent requirement
<!DOCTYPE html>
SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3
-- | -- | -- | -- | -- | -- | -- | -- | --
### Tasks
- [ ]
|
process
|
produce cdf formatted data products for ultra summary of the requirement the sdc needs to produce cdf formatted data products for the ultra instrument the data products for can be found in the ultra algorithm document here parent requirement soc sdc sdc execution of processing software the sdc shall be capable of executing processing software to produce data products within seven days of receipt of all requisite data inputs requirement the sdc produces data products via instrument team supplied algorithms and sdc produced processing code soc test sit processing sit tasks
| 1
|
23,983
| 5,008,429,050
|
IssuesEvent
|
2016-12-12 19:29:18
|
DjalmaPessoa/convey
|
https://api.github.com/repos/DjalmaPessoa/convey
|
opened
|
create a replication vignette for svygei
|
documentation
|
@guilhermejacob how do we hit the numbers `GE(-1)` ? flip back assignee to me once you finish or if we shouldn't bother? thanks
library(convey)
library(survey)
library(foreign)
tf <- tempfile()
download.file( "http://econpapers.repec.org/scripts/redir.pf?u=http%3A%2F%2Frepec.org%2Fnasug2006%2Fnasug2006_jenkins.zip;h=repec:boc:asug06:16" , tf , mode = 'wb' )
z <- unzip( tf , exdir = tempdir() )
x81 <- read.dta( grep( "ifs81" , z , value = TRUE ) )
x85 <- read.dta( grep( "ifs85" , z , value = TRUE ) )
x91 <- read.dta( grep( "ifs91" , z , value = TRUE ) )
x <- rbind( x81 , x85 , x91 )
y <- svydesign( ~ hrn , data = x , weights = ~ wgt )
z <- convey_prep( y )
z81 <- subset( z , year == 1981 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 0 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 3 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 0 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 2 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 3 )
# . svygei x if year == 1981
# Warning: x has 20 values = 0. Not used in calculations
# Complex survey estimates of Generalized Entropy inequality indices
# pweight: wgt Number of obs = 9752
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7459
# Population size = 54766261
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# GE(-1) | .1902062 .02474921 7.69 0.000 .1416987 .2387138
# MLD | .1142851 .00275138 41.54 0.000 .1088925 .1196777
# Theil | .1116923 .00226489 49.31 0.000 .1072532 .1161314
# GE(2) | .128793 .00330774 38.94 0.000 .1223099 .135276
# GE(3) | .1739994 .00662015 26.28 0.000 .1610242 .1869747
# ---------------------------------------------------------------------------
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 1.5 )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2.5 )
# . * account for clustering within HHs
# . version 8: svyset [pweight = wgt], psu(hrn)
# pweight is wgt
# psu is hrn
# . svyatk x if year == 1981
# Warning: x has 20 values = 0. Not used in calculations
# Complex survey estimates of Atkinson inequality indices
# pweight: wgt Number of obs = 9752
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7459
# Population size = 54766261
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# A(0.5) | .0543239 .00107583 50.49 0.000 .0522153 .0564324
# A(1) | .1079964 .00245424 44.00 0.000 .1031862 .1128066
# A(1.5) | .1701794 .0066943 25.42 0.000 .1570588 .1833
# A(2) | .2755788 .02597608 10.61 0.000 .2246666 .326491
# A(2.5) | .4992701 .06754311 7.39 0.000 .366888 .6316522
# ---------------------------------------------------------------------------
z81_two <- subset( z , year == 1981 & eybhc0 > 1 )
svyatk( ~ eybhc0 , z81_two , epsilon = 0.5 )
svyatk( ~ eybhc0 , z81_two )
svyatk( ~ eybhc0 , z81_two , epsilon = 1.5 )
svyatk( ~ eybhc0 , z81_two , epsilon = 2 )
svyatk( ~ eybhc0 , z81_two , epsilon = 2.5 )
# . svyatk x if year == 1981 & x >= 1
# Complex survey estimates of Atkinson inequality indices
# pweight: wgt Number of obs = 9748
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7457
# Population size = 54744234
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# A(0.5) | .0540059 .00105011 51.43 0.000 .0519477 .0560641
# A(1) | .1066082 .00223318 47.74 0.000 .1022313 .1109852
# A(1.5) | .1638299 .00483069 33.91 0.000 .154362 .1732979
# A(2) | .2443206 .01425258 17.14 0.000 .2163861 .2722552
# A(2.5) | .394787 .04155221 9.50 0.000 .3133461 .4762278
# ---------------------------------------------------------------------------
# . svygei x if year == 1985 & x >= 1
# Complex survey estimates of Generalized Entropy inequality indices
# pweight: wgt Number of obs = 8969
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 6950
# Population size = 55042871
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# GE(-1) | .1602358 .00936931 17.10 0.000 .1418723 .1785993
# MLD | .127616 .00332187 38.42 0.000 .1211052 .1341267
# Theil | .1337177 .00406302 32.91 0.000 .1257543 .141681
# GE(2) | .1676393 .00730057 22.96 0.000 .1533304 .1819481
# GE(3) | .2609507 .01850689 14.10 0.000 .2246779 .2972235
# ---------------------------------------------------------------------------
|
1.0
|
create a replication vignette for svygei - @guilhermejacob how do we hit the numbers `GE(-1)` ? flip back assignee to me once you finish or if we shouldn't bother? thanks
library(convey)
library(survey)
library(foreign)
tf <- tempfile()
download.file( "http://econpapers.repec.org/scripts/redir.pf?u=http%3A%2F%2Frepec.org%2Fnasug2006%2Fnasug2006_jenkins.zip;h=repec:boc:asug06:16" , tf , mode = 'wb' )
z <- unzip( tf , exdir = tempdir() )
x81 <- read.dta( grep( "ifs81" , z , value = TRUE ) )
x85 <- read.dta( grep( "ifs85" , z , value = TRUE ) )
x91 <- read.dta( grep( "ifs91" , z , value = TRUE ) )
x <- rbind( x81 , x85 , x91 )
y <- svydesign( ~ hrn , data = x , weights = ~ wgt )
z <- convey_prep( y )
z81 <- subset( z , year == 1981 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 0 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 3 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 0 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 2 )
svygei( ~ eybhc0 , subset( z81 , eybhc0 > 1 ) , epsilon = 3 )
# . svygei x if year == 1981
# Warning: x has 20 values = 0. Not used in calculations
# Complex survey estimates of Generalized Entropy inequality indices
# pweight: wgt Number of obs = 9752
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7459
# Population size = 54766261
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# GE(-1) | .1902062 .02474921 7.69 0.000 .1416987 .2387138
# MLD | .1142851 .00275138 41.54 0.000 .1088925 .1196777
# Theil | .1116923 .00226489 49.31 0.000 .1072532 .1161314
# GE(2) | .128793 .00330774 38.94 0.000 .1223099 .135276
# GE(3) | .1739994 .00662015 26.28 0.000 .1610242 .1869747
# ---------------------------------------------------------------------------
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 1.5 )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2 )
svyatk( ~ eybhc0 , subset( z81 , eybhc0 > 0 ) , epsilon = 2.5 )
# . * account for clustering within HHs
# . version 8: svyset [pweight = wgt], psu(hrn)
# pweight is wgt
# psu is hrn
# . svyatk x if year == 1981
# Warning: x has 20 values = 0. Not used in calculations
# Complex survey estimates of Atkinson inequality indices
# pweight: wgt Number of obs = 9752
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7459
# Population size = 54766261
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# A(0.5) | .0543239 .00107583 50.49 0.000 .0522153 .0564324
# A(1) | .1079964 .00245424 44.00 0.000 .1031862 .1128066
# A(1.5) | .1701794 .0066943 25.42 0.000 .1570588 .1833
# A(2) | .2755788 .02597608 10.61 0.000 .2246666 .326491
# A(2.5) | .4992701 .06754311 7.39 0.000 .366888 .6316522
# ---------------------------------------------------------------------------
z81_two <- subset( z , year == 1981 & eybhc0 > 1 )
svyatk( ~ eybhc0 , z81_two , epsilon = 0.5 )
svyatk( ~ eybhc0 , z81_two )
svyatk( ~ eybhc0 , z81_two , epsilon = 1.5 )
svyatk( ~ eybhc0 , z81_two , epsilon = 2 )
svyatk( ~ eybhc0 , z81_two , epsilon = 2.5 )
# . svyatk x if year == 1981 & x >= 1
# Complex survey estimates of Atkinson inequality indices
# pweight: wgt Number of obs = 9748
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 7457
# Population size = 54744234
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# A(0.5) | .0540059 .00105011 51.43 0.000 .0519477 .0560641
# A(1) | .1066082 .00223318 47.74 0.000 .1022313 .1109852
# A(1.5) | .1638299 .00483069 33.91 0.000 .154362 .1732979
# A(2) | .2443206 .01425258 17.14 0.000 .2163861 .2722552
# A(2.5) | .394787 .04155221 9.50 0.000 .3133461 .4762278
# ---------------------------------------------------------------------------
# . svygei x if year == 1985 & x >= 1
# Complex survey estimates of Generalized Entropy inequality indices
# pweight: wgt Number of obs = 8969
# Strata: <one> Number of strata = 1
# PSU: hrn Number of PSUs = 6950
# Population size = 55042871
# ---------------------------------------------------------------------------
# Index | Estimate Std. Err. z P>|z| [95% Conf. Interval]
# ---------+-----------------------------------------------------------------
# GE(-1) | .1602358 .00936931 17.10 0.000 .1418723 .1785993
# MLD | .127616 .00332187 38.42 0.000 .1211052 .1341267
# Theil | .1337177 .00406302 32.91 0.000 .1257543 .141681
# GE(2) | .1676393 .00730057 22.96 0.000 .1533304 .1819481
# GE(3) | .2609507 .01850689 14.10 0.000 .2246779 .2972235
# ---------------------------------------------------------------------------
|
non_process
|
create a replication vignette for svygei guilhermejacob how do we hit the numbers ge flip back assignee to me once you finish or if we shouldn t bother thanks library convey library survey library foreign tf tempfile download file tf mode wb z unzip tf exdir tempdir read dta grep z value true read dta grep z value true read dta grep z value true x rbind y svydesign hrn data x weights wgt z convey prep y subset z year svygei subset epsilon svygei subset svygei subset epsilon svygei subset epsilon svygei subset epsilon svygei subset svygei subset epsilon svygei subset epsilon svygei x if year warning x has values not used in calculations complex survey estimates of generalized entropy inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z ge mld theil ge ge svyatk subset svyatk subset epsilon svyatk subset epsilon svyatk subset epsilon account for clustering within hhs version svyset psu hrn pweight is wgt psu is hrn svyatk x if year warning x has values not used in calculations complex survey estimates of atkinson inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z a a a a a two svyatk two epsilon svyatk two svyatk two epsilon svyatk two epsilon svyatk two epsilon svyatk x if year x complex survey estimates of atkinson inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z a a a a a svygei x if year x complex survey estimates of generalized entropy inequality indices pweight wgt number of obs strata number of strata psu hrn number of psus population size index estimate std err z p z ge mld theil ge ge
| 0
|
3,005
| 6,006,945,885
|
IssuesEvent
|
2017-06-06 00:56:15
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Add ability to hide referrers from report (e.g., local sites).
|
add change log-processing
|
I've got 80...90% local referrers within a website.
With
goaccess my.log -a -o report.html
I get a nice report including all requests. But in the referrer urls and referrer sites reports, I only find referrers within the same website, as there are so many. Works as intended.
The report including the local referrers is correct, but not nice - I'd like to see only the interesting referrers directing from external websites to the local site.
So I try a
goaccess my.log -a --ignore-referer=www.example.com --ignore-referer=example.com -o report.html
and get nice referrer urls and referrer sites reports without the local site referrers.
Only problem: the ignore-referer option makes goaccess completely ignore the log lines containing these referrers (correctly behaviour as described in the .conf file: "Ignore referers from being counted"), resulting in a report that is missing many hits.
An additional or changed option would be nice: counting the loglines as hits, but ignoring their referrer information
|
1.0
|
Add ability to hide referrers from report (e.g., local sites). - I've got 80...90% local referrers within a website.
With
goaccess my.log -a -o report.html
I get a nice report including all requests. But in the referrer urls and referrer sites reports, I only find referrers within the same website, as there are so many. Works as intended.
The report including the local referrers is correct, but not nice - I'd like to see only the interesting referrers directing from external websites to the local site.
So I try a
goaccess my.log -a --ignore-referer=www.example.com --ignore-referer=example.com -o report.html
and get nice referrer urls and referrer sites reports without the local site referrers.
Only problem: the ignore-referer option makes goaccess completely ignore the log lines containing these referrers (correctly behaviour as described in the .conf file: "Ignore referers from being counted"), resulting in a report that is missing many hits.
An additional or changed option would be nice: counting the loglines as hits, but ignoring their referrer information
|
process
|
add ability to hide referrers from report e g local sites i ve got local referrers within a website with goaccess my log a o report html i get a nice report including all requests but in the referrer urls and referrer sites reports i only find referrers within the same website as there are so many works as intended the report including the local referrers is correct but not nice i d like to see only the interesting referrers directing from external websites to the local site so i try a goaccess my log a ignore referer ignore referer example com o report html and get nice referrer urls and referrer sites reports without the local site referrers only problem the ignore referer option makes goaccess completely ignore the log lines containing these referrers correctly behaviour as described in the conf file ignore referers from being counted resulting in a report that is missing many hits an additional or changed option would be nice counting the loglines as hits but ignoring their referrer information
| 1
|
2,084
| 4,912,510,406
|
IssuesEvent
|
2016-11-23 09:21:22
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
opened
|
Tabs do not display in completed tasks
|
browser: all bug comp: activiti-processList
|
1. Start a new process
2. Ensure user task within process has form attached that includes more than one tab
3. Go to tasklist
4. Complete task
5. Go to process list
6. Open task completed
**Expected results**
All tabs are displayed
**Actual results**
Only one tab is displayed
|
1.0
|
Tabs do not display in completed tasks - 1. Start a new process
2. Ensure user task within process has form attached that includes more than one tab
3. Go to tasklist
4. Complete task
5. Go to process list
6. Open task completed
**Expected results**
All tabs are displayed
**Actual results**
Only one tab is displayed
|
process
|
tabs do not display in completed tasks start a new process ensure user task within process has form attached that includes more than one tab go to tasklist complete task go to process list open task completed expected results all tabs are displayed actual results only one tab is displayed
| 1
|
83,493
| 10,369,211,431
|
IssuesEvent
|
2019-09-08 00:21:30
|
phetsims/vector-addition
|
https://api.github.com/repos/phetsims/vector-addition
|
closed
|
incorrect labeling of magnitudes in Base Vectors accordion box
|
design:polish
|
In 1.0.0-dev.11 (baseline), and still in master, the Base Vectors accordion box looks like this for the polar snap mode:
<img width="176" alt="screenshot_1444" src="https://user-images.githubusercontent.com/3046552/64452639-c9cf2f80-d0b4-11e9-9a28-74caa25460a6.png">
@arouinfar shouldn't the magnitudes be labeled as |d⃗| and |e⃗|, with a vector arrow over each symbol?
|
1.0
|
incorrect labeling of magnitudes in Base Vectors accordion box - In 1.0.0-dev.11 (baseline), and still in master, the Base Vectors accordion box looks like this for the polar snap mode:
<img width="176" alt="screenshot_1444" src="https://user-images.githubusercontent.com/3046552/64452639-c9cf2f80-d0b4-11e9-9a28-74caa25460a6.png">
@arouinfar shouldn't the magnitudes be labeled as |d⃗| and |e⃗|, with a vector arrow over each symbol?
|
non_process
|
incorrect labeling of magnitudes in base vectors accordion box in dev baseline and still in master the base vectors accordion box looks like this for the polar snap mode img width alt screenshot src arouinfar shouldn t the magnitudes be labeled as d and e with a vector arrow over each symbol
| 0
|
12,029
| 14,738,585,558
|
IssuesEvent
|
2021-01-07 05:10:59
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Session expiry functionality to be implemented
|
Bug P1 Participant manager Process: Dev Process: Release 2 Process: Tested QA
|
Once session get Expired (30 minutes Inactive), User should be logged out of the application and sign-in page should be displayed
|
3.0
|
[PM] Session expiry functionality to be implemented - Once session get Expired (30 minutes Inactive), User should be logged out of the application and sign-in page should be displayed
|
process
|
session expiry functionality to be implemented once session get expired minutes inactive user should be logged out of the application and sign in page should be displayed
| 1
|
16,216
| 20,742,611,722
|
IssuesEvent
|
2022-03-14 19:13:54
|
ankidroid/Anki-Android
|
https://api.github.com/repos/ankidroid/Anki-Android
|
closed
|
Regression test: no crash if sound is not found
|
Help Wanted Good First Issue! Stale Test process
|
See stack trace for the cause: https://github.com/ankidroid/Anki-Android/issues/9401
The deliverable will be a unit test which fails before PR https://github.com/ankidroid/Anki-Android/pull/9407 goes in, and passes afterwards
----
Thanks! Could you add a regression test here for when the sound is not in the collection?
_Originally posted by @david-allison-1 in https://github.com/ankidroid/Anki-Android/issues/9407#issuecomment-898621819_
|
1.0
|
Regression test: no crash if sound is not found - See stack trace for the cause: https://github.com/ankidroid/Anki-Android/issues/9401
The deliverable will be a unit test which fails before PR https://github.com/ankidroid/Anki-Android/pull/9407 goes in, and passes afterwards
----
Thanks! Could you add a regression test here for when the sound is not in the collection?
_Originally posted by @david-allison-1 in https://github.com/ankidroid/Anki-Android/issues/9407#issuecomment-898621819_
|
process
|
regression test no crash if sound is not found see stack trace for the cause the deliverable will be a unit test which fails before pr goes in and passes afterwards thanks could you add a regression test here for when the sound is not in the collection originally posted by david allison in
| 1
|
53,317
| 11,035,667,136
|
IssuesEvent
|
2019-12-07 15:19:46
|
paulevsGitch/BetterNether
|
https://api.github.com/repos/paulevsGitch/BetterNether
|
closed
|
Extreme load times
|
Fixed in dev code
|
Hello I added this to my modpack but I noticed a considerable load time for the Nether, after some tests these are the results:
**Nether Performance Test**
BOP + BetterNether
_37,73s_
BetterNether
_35,23s_
BOP
_6,08s_
None
_5,08s_
And this isn't even on a low-end machine.
|
1.0
|
Extreme load times - Hello I added this to my modpack but I noticed a considerable load time for the Nether, after some tests these are the results:
**Nether Performance Test**
BOP + BetterNether
_37,73s_
BetterNether
_35,23s_
BOP
_6,08s_
None
_5,08s_
And this isn't even on a low-end machine.
|
non_process
|
extreme load times hello i added this to my modpack but i noticed a considerable load time for the nether after some tests these are the results nether performance test bop betternether betternether bop none and this isn t even on a low end machine
| 0
|
53,431
| 11,053,284,860
|
IssuesEvent
|
2019-12-10 11:02:52
|
wahlgenial/wahlgenial-webapp
|
https://api.github.com/repos/wahlgenial/wahlgenial-webapp
|
opened
|
Teacher_apps Seite: Verlinkung auf die Detailseiten überlappt mit Dropdown
|
Code Priority: High bug
|
Die Verlinkung der Detailseiten auf der Lehrer-apps Seite überlappt die aus dem Dropdownmenü. So sind die Verlinkungen im Dropdown an manchen Stellen überschrieben.
Das sollte man über ein Mouseover im Dropdownmenü erkennen.
|
1.0
|
Teacher_apps Seite: Verlinkung auf die Detailseiten überlappt mit Dropdown - Die Verlinkung der Detailseiten auf der Lehrer-apps Seite überlappt die aus dem Dropdownmenü. So sind die Verlinkungen im Dropdown an manchen Stellen überschrieben.
Das sollte man über ein Mouseover im Dropdownmenü erkennen.
|
non_process
|
teacher apps seite verlinkung auf die detailseiten überlappt mit dropdown die verlinkung der detailseiten auf der lehrer apps seite überlappt die aus dem dropdownmenü so sind die verlinkungen im dropdown an manchen stellen überschrieben das sollte man über ein mouseover im dropdownmenü erkennen
| 0
|
20,234
| 26,839,596,123
|
IssuesEvent
|
2023-02-02 22:45:23
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
Spotless Depends on NPM installed and available on $PATH
|
bug process
|
### Description
When importing the project, the Intellij sync process done to import a new gradle project fails.
Spotless attempts to resolve the NPM executable during the configuration phase which is prior to node gradle plugin installing node.
### Steps to reproduce
1. Create new intellij project via VCS or existing source
2. Go to the build tab to see gradle output
3. Notice spotless failures
### Additional context
```console
FAILURE: Build failed with an exception.
* What went wrong:
Could not create task ':common:spotlessJavascriptApply'.
Could not create task ':common:spotlessJavascript'.
Can't automatically determine npm executable and none was specifically supplied!
Spotless tries to find your npm executable automatically. It looks for npm in the following places:
- An executable referenced by the java system property 'npm.exec' - if such a system property exists.
- The environment variable 'NVM_BIN' - if such an environment variable exists.
- The environment variable 'NVM_SYMLINK' - if such an environment variable exists.
- The environment variable 'NODE_PATH' - if such an environment variable exists.
- In your 'PATH' environment variable
If autodiscovery fails for your system, try to set one of the environment variables correctly or
try setting the system property 'npm.exec' in the build process to override autodiscovery.
* Try:
> Run with --info or --debug option to get more log output.
* Exception is:
com.intellij.openapi.externalSystem.model.ExternalSystemException: Could not create task ':common:spotlessJavascriptApply'.
Could not create task ':common:spotlessJavascript'.
Can't automatically determine npm executable and none was specifically supplied!
Spotless tries to find your npm executable automatically. It looks for npm in the following places:
- An executable referenced by the java system property 'npm.exec' - if such a system property exists.
- The environment variable 'NVM_BIN' - if such an environment variable exists.
- The environment variable 'NVM_SYMLINK' - if such an environment variable exists.
- The environment variable 'NODE_PATH' - if such an environment variable exists.
- In your 'PATH' environment variable
If autodiscovery fails for your system, try to set one of the environment variables correctly or
try setting the system property 'npm.exec' in the build process to override autodiscovery.
at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:412)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42)
at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84)
at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113)
at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166)
at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247)
at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166)
at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44)
at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43)
at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53)
at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35)
at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49)
at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65)
at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136)
at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249)
at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110)
at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64)
at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88)
at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31)
at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75)
at org.gradle.util.internal.Swapper.swap(Swapper.java:38)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52)
at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49)
org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascriptApply'.
at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715)
at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929)
at org.gradle.api.internal.provider.AbstractMinimalProvider.getOrNull(AbstractMinimalProvider.java:97)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.findByName(DefaultNamedDomainObjectCollection.java:295)
at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:558)
at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:75)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.tasks(GradleProjectBuilder.java:105)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:80)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:58)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildAll(GradleProjectBuilder.java:52)
at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:78)
at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:60)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithNoParameter.build(DefaultToolingModelBuilderRegistry.java:264)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383)
at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334)
at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82)
at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106)
at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:634)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:655)
at org.jetbrains.plugins.gradle.model.ClassSetImportModelProvider.populateBuildModels(ClassSetImportModelProvider.java:27)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:400)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42)
at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84)
at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113)
at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166)
at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247)
at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166)
at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44)
at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43)
at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53)
at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35)
at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49)
at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65)
at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136)
at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249)
at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110)
at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64)
at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88)
at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31)
at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75)
at org.gradle.util.internal.Swapper.swap(Swapper.java:38)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52)
at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascript'.
at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715)
at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929)
at org.gradle.api.internal.provider.AbstractMinimalProvider.calculateValue(AbstractMinimalProvider.java:107)
at org.gradle.api.internal.provider.Collectors$ElementFromProvider.collectEntries(Collectors.java:100)
at org.gradle.api.internal.provider.Collectors$TypedCollector.collectEntries(Collectors.java:334)
at org.gradle.api.internal.provider.Collectors$TypedCollector.collectInto(Collectors.java:329)
at org.gradle.api.internal.collections.DefaultPendingSource.realize(DefaultPendingSource.java:62)
at org.gradle.api.internal.collections.DefaultPendingSource.realizePending(DefaultPendingSource.java:39)
at org.gradle.api.internal.collections.SortedSetElementSource.iterator(SortedSetElementSource.java:63)
at org.gradle.api.internal.DefaultDomainObjectCollection.iterator(DefaultDomainObjectCollection.java:128)
at org.gradle.api.internal.tasks.DefaultTaskContainer.iterator(DefaultTaskContainer.java:620)
at java.base/java.util.AbstractCollection.addAll(AbstractCollection.java:335)
at java.base/java.util.TreeSet.addAll(TreeSet.java:309)
at java.base/java.util.TreeSet.<init>(TreeSet.java:160)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getAllTasks(TasksFactory.java:56)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.collectTasks(TasksFactory.java:46)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getTasks(TasksFactory.java:78)
at org.jetbrains.gradle.profiler.builder.JavaTaskModelBuilderImpl.buildAll(JavaTaskModelBuilderImpl.java:34)
at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:107)
at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:76)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithParameter.build(DefaultToolingModelBuilderRegistry.java:287)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:374)
at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication.reapply(DefaultUserCodeApplicationContext.java:109)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.build(DefaultToolingModelBuilderRegistry.java:374)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383)
at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334)
at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82)
at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106)
at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:658)
at org.jetbrains.plugins.gradle.model.ClassSetProjectImportModelProvider.populateProjectModels(ClassSetProjectImportModelProvider.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.getProjectModels(ProjectImportAction.java:338)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.access$500(ProjectImportAction.java:42)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:249)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:246)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.fetchProjectBuildModels(ProjectImportAction.java:262)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:137)
... 87 more
Caused by: java.lang.IllegalStateException: Can't automatically determine npm executable and none was specifically supplied!
```
### Hedera network
other
### Version
0.71.0-SNAPSHOT
### Operating system
macOS
|
1.0
|
Spotless Depends on NPM installed and available on $PATH - ### Description
When importing the project, the Intellij sync process done to import a new gradle project fails.
Spotless attempts to resolve the NPM executable during the configuration phase which is prior to node gradle plugin installing node.
### Steps to reproduce
1. Create new intellij project via VCS or existing source
2. Go to the build tab to see gradle output
3. Notice spotless failures
### Additional context
```console
FAILURE: Build failed with an exception.
* What went wrong:
Could not create task ':common:spotlessJavascriptApply'.
Could not create task ':common:spotlessJavascript'.
Can't automatically determine npm executable and none was specifically supplied!
Spotless tries to find your npm executable automatically. It looks for npm in the following places:
- An executable referenced by the java system property 'npm.exec' - if such a system property exists.
- The environment variable 'NVM_BIN' - if such an environment variable exists.
- The environment variable 'NVM_SYMLINK' - if such an environment variable exists.
- The environment variable 'NODE_PATH' - if such an environment variable exists.
- In your 'PATH' environment variable
If autodiscovery fails for your system, try to set one of the environment variables correctly or
try setting the system property 'npm.exec' in the build process to override autodiscovery.
* Try:
> Run with --info or --debug option to get more log output.
* Exception is:
com.intellij.openapi.externalSystem.model.ExternalSystemException: Could not create task ':common:spotlessJavascriptApply'.
Could not create task ':common:spotlessJavascript'.
Can't automatically determine npm executable and none was specifically supplied!
Spotless tries to find your npm executable automatically. It looks for npm in the following places:
- An executable referenced by the java system property 'npm.exec' - if such a system property exists.
- The environment variable 'NVM_BIN' - if such an environment variable exists.
- The environment variable 'NVM_SYMLINK' - if such an environment variable exists.
- The environment variable 'NODE_PATH' - if such an environment variable exists.
- In your 'PATH' environment variable
If autodiscovery fails for your system, try to set one of the environment variables correctly or
try setting the system property 'npm.exec' in the build process to override autodiscovery.
at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:412)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42)
at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84)
at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113)
at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166)
at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247)
at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166)
at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44)
at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43)
at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53)
at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35)
at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49)
at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65)
at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136)
at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249)
at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110)
at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64)
at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88)
at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31)
at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75)
at org.gradle.util.internal.Swapper.swap(Swapper.java:38)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52)
at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49)
org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascriptApply'.
at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715)
at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929)
at org.gradle.api.internal.provider.AbstractMinimalProvider.getOrNull(AbstractMinimalProvider.java:97)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.findByName(DefaultNamedDomainObjectCollection.java:295)
at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:558)
at org.gradle.api.internal.tasks.DefaultTaskContainer.findByName(DefaultTaskContainer.java:75)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.tasks(GradleProjectBuilder.java:105)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:80)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildHierarchy(GradleProjectBuilder.java:58)
at org.gradle.plugins.ide.internal.tooling.GradleProjectBuilder.buildAll(GradleProjectBuilder.java:52)
at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:78)
at org.gradle.plugins.ide.internal.tooling.IdeaModelBuilder.buildAll(IdeaModelBuilder.java:60)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithNoParameter.build(DefaultToolingModelBuilderRegistry.java:264)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383)
at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334)
at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82)
at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106)
at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:634)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:655)
at org.jetbrains.plugins.gradle.model.ClassSetImportModelProvider.populateBuildModels(ClassSetImportModelProvider.java:27)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.addBuildModels(ProjectImportAction.java:400)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:138)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:42)
at org.gradle.tooling.internal.consumer.connection.InternalBuildActionAdapter.execute(InternalBuildActionAdapter.java:64)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.runAction(AbstractClientProvidedBuildActionRunner.java:131)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:104)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner$ActionAdapter.fromBuildModel(AbstractClientProvidedBuildActionRunner.java:84)
at org.gradle.internal.buildtree.DefaultBuildTreeModelCreator.fromBuildModel(DefaultBuildTreeModelCreator.java:57)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$fromBuildModel$2(DefaultBuildTreeLifecycleController.java:89)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$5(DefaultBuildTreeLifecycleController.java:113)
at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166)
at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247)
at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166)
at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44)
at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.runBuild(DefaultBuildTreeLifecycleController.java:110)
at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.fromBuildModel(DefaultBuildTreeLifecycleController.java:81)
at org.gradle.tooling.internal.provider.runner.AbstractClientProvidedBuildActionRunner.runClientAction(AbstractClientProvidedBuildActionRunner.java:43)
at org.gradle.tooling.internal.provider.runner.ClientProvidedPhasedActionRunner.run(ClientProvidedPhasedActionRunner.java:53)
at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35)
at org.gradle.internal.buildtree.ProblemReportingBuildActionRunner.run(ProblemReportingBuildActionRunner.java:49)
at org.gradle.launcher.exec.BuildOutcomeReportingBuildActionRunner.run(BuildOutcomeReportingBuildActionRunner.java:65)
at org.gradle.tooling.internal.provider.FileSystemWatchingBuildActionRunner.run(FileSystemWatchingBuildActionRunner.java:136)
at org.gradle.launcher.exec.BuildCompletionNotifyingBuildActionRunner.run(BuildCompletionNotifyingBuildActionRunner.java:41)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122)
at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40)
at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53)
at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249)
at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109)
at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36)
at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110)
at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64)
at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88)
at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62)
at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63)
at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52)
at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47)
at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31)
at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75)
at org.gradle.util.internal.Swapper.swap(Swapper.java:38)
at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84)
at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37)
at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104)
at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52)
at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:49)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreationException: Could not create task ':common:spotlessJavascript'.
at org.gradle.api.internal.tasks.DefaultTaskContainer.taskCreationException(DefaultTaskContainer.java:715)
at org.gradle.api.internal.tasks.DefaultTaskContainer.access$600(DefaultTaskContainer.java:76)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.domainObjectCreationException(DefaultTaskContainer.java:707)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:948)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929)
at org.gradle.api.internal.provider.AbstractMinimalProvider.calculateValue(AbstractMinimalProvider.java:107)
at org.gradle.api.internal.provider.Collectors$ElementFromProvider.collectEntries(Collectors.java:100)
at org.gradle.api.internal.provider.Collectors$TypedCollector.collectEntries(Collectors.java:334)
at org.gradle.api.internal.provider.Collectors$TypedCollector.collectInto(Collectors.java:329)
at org.gradle.api.internal.collections.DefaultPendingSource.realize(DefaultPendingSource.java:62)
at org.gradle.api.internal.collections.DefaultPendingSource.realizePending(DefaultPendingSource.java:39)
at org.gradle.api.internal.collections.SortedSetElementSource.iterator(SortedSetElementSource.java:63)
at org.gradle.api.internal.DefaultDomainObjectCollection.iterator(DefaultDomainObjectCollection.java:128)
at org.gradle.api.internal.tasks.DefaultTaskContainer.iterator(DefaultTaskContainer.java:620)
at java.base/java.util.AbstractCollection.addAll(AbstractCollection.java:335)
at java.base/java.util.TreeSet.addAll(TreeSet.java:309)
at java.base/java.util.TreeSet.<init>(TreeSet.java:160)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getAllTasks(TasksFactory.java:56)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.collectTasks(TasksFactory.java:46)
at org.jetbrains.plugins.gradle.tooling.builder.TasksFactory.getTasks(TasksFactory.java:78)
at org.jetbrains.gradle.profiler.builder.JavaTaskModelBuilderImpl.buildAll(JavaTaskModelBuilderImpl.java:34)
at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:107)
at org.jetbrains.plugins.gradle.tooling.internal.ExtraModelBuilder.buildAll(ExtraModelBuilder.java:76)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuilderWithParameter.build(DefaultToolingModelBuilderRegistry.java:287)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:374)
at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication.reapply(DefaultUserCodeApplicationContext.java:109)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$UserCodeAssigningBuilder.build(DefaultToolingModelBuilderRegistry.java:374)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.lambda$build$0(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.lambda$fromMutableState$1(DefaultProjectStateRegistry.java:383)
at org.gradle.internal.work.DefaultWorkerLeaseService.withReplacedLocks(DefaultWorkerLeaseService.java:345)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:383)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$LockSingleProjectBuilder.build(DefaultToolingModelBuilderRegistry.java:304)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder$1.call(DefaultToolingModelBuilderRegistry.java:337)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.tooling.provider.model.internal.DefaultToolingModelBuilderRegistry$BuildOperationWrappingBuilder.build(DefaultToolingModelBuilderRegistry.java:334)
at org.gradle.internal.build.DefaultBuildToolingModelController$AbstractToolingScope.getModel(DefaultBuildToolingModelController.java:82)
at org.gradle.tooling.internal.provider.runner.DefaultBuildController.getModel(DefaultBuildController.java:106)
at org.gradle.tooling.internal.consumer.connection.ParameterAwareBuildControllerAdapter.getModel(ParameterAwareBuildControllerAdapter.java:39)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.getModel(UnparameterizedBuildController.java:113)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.getModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:97)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.gradle.tooling.internal.consumer.connection.UnparameterizedBuildController.findModel(UnparameterizedBuildController.java:81)
at org.gradle.tooling.internal.consumer.connection.NestedActionAwareBuildControllerAdapter.findModel(NestedActionAwareBuildControllerAdapter.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$MyBuildController.findModel(ProjectImportAction.java:658)
at org.jetbrains.plugins.gradle.model.ClassSetProjectImportModelProvider.populateProjectModels(ClassSetProjectImportModelProvider.java:31)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.getProjectModels(ProjectImportAction.java:338)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.access$500(ProjectImportAction.java:42)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:249)
at org.jetbrains.plugins.gradle.model.ProjectImportAction$4.execute(ProjectImportAction.java:246)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.fetchProjectBuildModels(ProjectImportAction.java:262)
at org.jetbrains.plugins.gradle.model.ProjectImportAction.execute(ProjectImportAction.java:137)
... 87 more
Caused by: java.lang.IllegalStateException: Can't automatically determine npm executable and none was specifically supplied!
```
### Hedera network
other
### Version
0.71.0-SNAPSHOT
### Operating system
macOS
|
process
|
spotless depends on npm installed and available on path description when importing the project the intellij sync process done to import a new gradle project fails spotless attempts to resolve the npm executable during the configuration phase which is prior to node gradle plugin installing node steps to reproduce create new intellij project via vcs or existing source go to the build tab to see gradle output notice spotless failures additional context console failure build failed with an exception what went wrong could not create task common spotlessjavascriptapply could not create task common spotlessjavascript can t automatically determine npm executable and none was specifically supplied spotless tries to find your npm executable automatically it looks for npm in the following places an executable referenced by the java system property npm exec if such a system property exists the environment variable nvm bin if such an environment variable exists the environment variable nvm symlink if such an environment variable exists the environment variable node path if such an environment variable exists in your path environment variable if autodiscovery fails for your system try to set one of the environment variables correctly or try setting the system property npm exec in the build process to override autodiscovery try run with info or debug option to get more log output exception is com intellij openapi externalsystem model externalsystemexception could not create task common spotlessjavascriptapply could not create task common spotlessjavascript can t automatically determine npm executable and none was specifically supplied spotless tries to find your npm executable automatically it looks for npm in the following places an executable referenced by the java system property npm exec if such a system property exists the environment variable nvm bin if such an environment variable exists the environment variable nvm symlink if such an environment variable exists the environment variable node path if such an environment variable exists in your path environment variable if autodiscovery fails for your system try to set one of the environment variables correctly or try setting the system property npm exec in the build process to override autodiscovery at org jetbrains plugins gradle model projectimportaction addbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org gradle tooling internal consumer connection internalbuildactionadapter execute internalbuildactionadapter java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter runaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle internal buildtree defaultbuildtreemodelcreator frombuildmodel defaultbuildtreemodelcreator java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda runbuild defaultbuildtreelifecyclecontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transition statetransitioncontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller runbuild defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner runclientaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner clientprovidedphasedactionrunner run clientprovidedphasedactionrunner java at org gradle launcher exec chainingbuildactionrunner run chainingbuildactionrunner java at org gradle internal buildtree problemreportingbuildactionrunner run problemreportingbuildactionrunner java at org gradle launcher exec buildoutcomereportingbuildactionrunner run buildoutcomereportingbuildactionrunner java at org gradle tooling internal provider filesystemwatchingbuildactionrunner run filesystemwatchingbuildactionrunner java at org gradle launcher exec buildcompletionnotifyingbuildactionrunner run buildcompletionnotifyingbuildactionrunner java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor lambda execute rootbuildlifecyclebuildactionexecutor java at org gradle composite internal defaultrootbuildstate run defaultrootbuildstate java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor execute rootbuildlifecyclebuildactionexecutor java at org gradle internal buildtree defaultbuildtreecontext execute defaultbuildtreecontext java at org gradle launcher exec buildtreelifecyclebuildactionexecutor lambda execute buildtreelifecyclebuildactionexecutor java at org gradle internal buildtree buildtreestate run buildtreestate java at org gradle launcher exec buildtreelifecyclebuildactionexecutor execute buildtreelifecyclebuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor execute runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasworkerthreadbuildactionexecutor lambda execute runasworkerthreadbuildactionexecutor java at org gradle internal work defaultworkerleaseservice withlocks defaultworkerleaseservice java at org gradle internal work defaultworkerleaseservice runasworkerthread defaultworkerleaseservice java at org gradle launcher exec runasworkerthreadbuildactionexecutor execute runasworkerthreadbuildactionexecutor java at org gradle tooling internal provider continuous continuousbuildactionexecutor execute continuousbuildactionexecutor java at org gradle tooling internal provider subscribablebuildactionexecutor execute subscribablebuildactionexecutor java at org gradle internal session defaultbuildsessioncontext execute defaultbuildsessioncontext java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle internal session buildsessionstate run buildsessionstate java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle launcher daemon server exec executebuild dobuild executebuild java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec watchfordisconnection execute watchfordisconnection java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec resetdeprecationlogger execute resetdeprecationlogger java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec requeststopifsingleuseddaemon execute requeststopifsingleuseddaemon java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle util internal swapper swap swapper java at org gradle launcher daemon server exec forwardclientinput execute forwardclientinput java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logandcheckhealth execute logandcheckhealth java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logtoclient dobuild logtoclient java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec establishbuildenvironment dobuild establishbuildenvironment java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec startbuildorrespondwithbusy run startbuildorrespondwithbusy java at org gradle launcher daemon server daemonstatecoordinator run daemonstatecoordinator java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java org gradle api internal tasks defaulttaskcontainer taskcreationexception could not create task common spotlessjavascriptapply at org gradle api internal tasks defaulttaskcontainer taskcreationexception defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider domainobjectcreationexception defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider trycreate defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider run defaulttaskcontainer java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider trycreate defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider calculateownvalue defaultnameddomainobjectcollection java at org gradle api internal provider abstractminimalprovider getornull abstractminimalprovider java at org gradle api internal defaultnameddomainobjectcollection findbyname defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer findbyname defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer findbyname defaulttaskcontainer java at org gradle plugins ide internal tooling gradleprojectbuilder tasks gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildhierarchy gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildhierarchy gradleprojectbuilder java at org gradle plugins ide internal tooling gradleprojectbuilder buildall gradleprojectbuilder java at org gradle plugins ide internal tooling ideamodelbuilder buildall ideamodelbuilder java at org gradle plugins ide internal tooling ideamodelbuilder buildall ideamodelbuilder java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry builderwithnoparameter build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle api internal project defaultprojectstateregistry projectstateimpl lambda frommutablestate defaultprojectstateregistry java at org gradle internal work defaultworkerleaseservice withreplacedlocks defaultworkerleaseservice java at org gradle api internal project defaultprojectstateregistry projectstateimpl frommutablestate defaultprojectstateregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder call defaulttoolingmodelbuilderregistry java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder build defaulttoolingmodelbuilderregistry java at org gradle internal build defaultbuildtoolingmodelcontroller abstracttoolingscope getmodel defaultbuildtoolingmodelcontroller java at org gradle tooling internal provider runner defaultbuildcontroller getmodel defaultbuildcontroller java at org gradle tooling internal consumer connection parameterawarebuildcontrolleradapter getmodel parameterawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller getmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter getmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model classsetimportmodelprovider populatebuildmodels classsetimportmodelprovider java at org jetbrains plugins gradle model projectimportaction addbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org gradle tooling internal consumer connection internalbuildactionadapter execute internalbuildactionadapter java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter runaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner actionadapter frombuildmodel abstractclientprovidedbuildactionrunner java at org gradle internal buildtree defaultbuildtreemodelcreator frombuildmodel defaultbuildtreemodelcreator java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda runbuild defaultbuildtreelifecyclecontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transition statetransitioncontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller runbuild defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller frombuildmodel defaultbuildtreelifecyclecontroller java at org gradle tooling internal provider runner abstractclientprovidedbuildactionrunner runclientaction abstractclientprovidedbuildactionrunner java at org gradle tooling internal provider runner clientprovidedphasedactionrunner run clientprovidedphasedactionrunner java at org gradle launcher exec chainingbuildactionrunner run chainingbuildactionrunner java at org gradle internal buildtree problemreportingbuildactionrunner run problemreportingbuildactionrunner java at org gradle launcher exec buildoutcomereportingbuildactionrunner run buildoutcomereportingbuildactionrunner java at org gradle tooling internal provider filesystemwatchingbuildactionrunner run filesystemwatchingbuildactionrunner java at org gradle launcher exec buildcompletionnotifyingbuildactionrunner run buildcompletionnotifyingbuildactionrunner java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor lambda execute rootbuildlifecyclebuildactionexecutor java at org gradle composite internal defaultrootbuildstate run defaultrootbuildstate java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor execute rootbuildlifecyclebuildactionexecutor java at org gradle internal buildtree defaultbuildtreecontext execute defaultbuildtreecontext java at org gradle launcher exec buildtreelifecyclebuildactionexecutor lambda execute buildtreelifecyclebuildactionexecutor java at org gradle internal buildtree buildtreestate run buildtreestate java at org gradle launcher exec buildtreelifecyclebuildactionexecutor execute buildtreelifecyclebuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor execute runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasworkerthreadbuildactionexecutor lambda execute runasworkerthreadbuildactionexecutor java at org gradle internal work defaultworkerleaseservice withlocks defaultworkerleaseservice java at org gradle internal work defaultworkerleaseservice runasworkerthread defaultworkerleaseservice java at org gradle launcher exec runasworkerthreadbuildactionexecutor execute runasworkerthreadbuildactionexecutor java at org gradle tooling internal provider continuous continuousbuildactionexecutor execute continuousbuildactionexecutor java at org gradle tooling internal provider subscribablebuildactionexecutor execute subscribablebuildactionexecutor java at org gradle internal session defaultbuildsessioncontext execute defaultbuildsessioncontext java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle internal session buildsessionstate run buildsessionstate java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle launcher daemon server exec executebuild dobuild executebuild java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec watchfordisconnection execute watchfordisconnection java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec resetdeprecationlogger execute resetdeprecationlogger java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec requeststopifsingleuseddaemon execute requeststopifsingleuseddaemon java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle util internal swapper swap swapper java at org gradle launcher daemon server exec forwardclientinput execute forwardclientinput java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logandcheckhealth execute logandcheckhealth java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logtoclient dobuild logtoclient java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec establishbuildenvironment dobuild establishbuildenvironment java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec startbuildorrespondwithbusy run startbuildorrespondwithbusy java at org gradle launcher daemon server daemonstatecoordinator run daemonstatecoordinator java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by org gradle api internal tasks defaulttaskcontainer taskcreationexception could not create task common spotlessjavascript at org gradle api internal tasks defaulttaskcontainer taskcreationexception defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider domainobjectcreationexception defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider trycreate defaultnameddomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider access defaulttaskcontainer java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider run defaulttaskcontainer java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal tasks defaulttaskcontainer taskcreatingprovider trycreate defaulttaskcontainer java at org gradle api internal defaultnameddomainobjectcollection abstractdomainobjectcreatingprovider calculateownvalue defaultnameddomainobjectcollection java at org gradle api internal provider abstractminimalprovider calculatevalue abstractminimalprovider java at org gradle api internal provider collectors elementfromprovider collectentries collectors java at org gradle api internal provider collectors typedcollector collectentries collectors java at org gradle api internal provider collectors typedcollector collectinto collectors java at org gradle api internal collections defaultpendingsource realize defaultpendingsource java at org gradle api internal collections defaultpendingsource realizepending defaultpendingsource java at org gradle api internal collections sortedsetelementsource iterator sortedsetelementsource java at org gradle api internal defaultdomainobjectcollection iterator defaultdomainobjectcollection java at org gradle api internal tasks defaulttaskcontainer iterator defaulttaskcontainer java at java base java util abstractcollection addall abstractcollection java at java base java util treeset addall treeset java at java base java util treeset treeset java at org jetbrains plugins gradle tooling builder tasksfactory getalltasks tasksfactory java at org jetbrains plugins gradle tooling builder tasksfactory collecttasks tasksfactory java at org jetbrains plugins gradle tooling builder tasksfactory gettasks tasksfactory java at org jetbrains gradle profiler builder javataskmodelbuilderimpl buildall javataskmodelbuilderimpl java at org jetbrains plugins gradle tooling internal extramodelbuilder buildall extramodelbuilder java at org jetbrains plugins gradle tooling internal extramodelbuilder buildall extramodelbuilder java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry builderwithparameter build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry usercodeassigningbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle configuration internal defaultusercodeapplicationcontext currentapplication reapply defaultusercodeapplicationcontext java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry usercodeassigningbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder lambda build defaulttoolingmodelbuilderregistry java at org gradle api internal project defaultprojectstateregistry projectstateimpl lambda frommutablestate defaultprojectstateregistry java at org gradle internal work defaultworkerleaseservice withreplacedlocks defaultworkerleaseservice java at org gradle api internal project defaultprojectstateregistry projectstateimpl frommutablestate defaultprojectstateregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry locksingleprojectbuilder build defaulttoolingmodelbuilderregistry java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder call defaulttoolingmodelbuilderregistry java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle tooling provider model internal defaulttoolingmodelbuilderregistry buildoperationwrappingbuilder build defaulttoolingmodelbuilderregistry java at org gradle internal build defaultbuildtoolingmodelcontroller abstracttoolingscope getmodel defaultbuildtoolingmodelcontroller java at org gradle tooling internal provider runner defaultbuildcontroller getmodel defaultbuildcontroller java at org gradle tooling internal consumer connection parameterawarebuildcontrolleradapter getmodel parameterawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller getmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter getmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org gradle tooling internal consumer connection unparameterizedbuildcontroller findmodel unparameterizedbuildcontroller java at org gradle tooling internal consumer connection nestedactionawarebuildcontrolleradapter findmodel nestedactionawarebuildcontrolleradapter java at org jetbrains plugins gradle model projectimportaction mybuildcontroller findmodel projectimportaction java at org jetbrains plugins gradle model classsetprojectimportmodelprovider populateprojectmodels classsetprojectimportmodelprovider java at org jetbrains plugins gradle model projectimportaction getprojectmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction access projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java at org jetbrains plugins gradle model projectimportaction fetchprojectbuildmodels projectimportaction java at org jetbrains plugins gradle model projectimportaction execute projectimportaction java more caused by java lang illegalstateexception can t automatically determine npm executable and none was specifically supplied hedera network other version snapshot operating system macos
| 1
|
10,201
| 13,065,958,398
|
IssuesEvent
|
2020-07-30 20:44:00
|
keep-network/keep-ecdsa
|
https://api.github.com/repos/keep-network/keep-ecdsa
|
closed
|
Allow the application to configure expectations about the minimum bond
|
process & client team ⛓chain
|
Right now, bonded sortition pool updates the minimum bondable value for each successful signer set selection. This way, the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application's expectations.
Instead of doing it automatically, we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane, balancing signer selection cost, to the application. For tBTC, this value could be the minimum lot size.
|
1.0
|
Allow the application to configure expectations about the minimum bond - Right now, bonded sortition pool updates the minimum bondable value for each successful signer set selection. This way, the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application's expectations.
Instead of doing it automatically, we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane, balancing signer selection cost, to the application. For tBTC, this value could be the minimum lot size.
|
process
|
allow the application to configure expectations about the minimum bond right now bonded sortition pool updates the minimum bondable value for each successful signer set selection this way the gas cost of skipping ineligible signers can be minimized and the sortition pool automatically adjusts the requirements based on the current application s expectations instead of doing it automatically we could have the application set the minimum bondable value and leave the responsibility of keeping this value sane balancing signer selection cost to the application for tbtc this value could be the minimum lot size
| 1
|
416,437
| 12,146,458,777
|
IssuesEvent
|
2020-04-24 11:11:37
|
PoltyPoltou/CACAO2020
|
https://api.github.com/repos/PoltyPoltou/CACAO2020
|
opened
|
Implémenter IActeur
|
High Priority
|
Modification obligatoire :
getIndicateurs()
getParametres()
getJournaux()
getDescription()
next()
getNomsFilieresProposees()
getFiliere(string)
Modification possible :
initialiser()
|
1.0
|
Implémenter IActeur - Modification obligatoire :
getIndicateurs()
getParametres()
getJournaux()
getDescription()
next()
getNomsFilieresProposees()
getFiliere(string)
Modification possible :
initialiser()
|
non_process
|
implémenter iacteur modification obligatoire getindicateurs getparametres getjournaux getdescription next getnomsfilieresproposees getfiliere string modification possible initialiser
| 0
|
35,466
| 12,339,815,099
|
IssuesEvent
|
2020-05-14 18:47:38
|
wrbejar/JavaVulnerableLabIcatu
|
https://api.github.com/repos/wrbejar/JavaVulnerableLabIcatu
|
opened
|
CVE-2015-6420 (High) detected in commons-collections-3.2.1.jar
|
security vulnerability
|
## CVE-2015-6420 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to vulnerable library: /JavaVulnerableLabIcatu/target/JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar,/root/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/JavaVulnerableLabIcatu/commit/68aef85b943fe5994864ad4b1984daaa93ad6530">68aef85b943fe5994864ad4b1984daaa93ad6530</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.
<p>Publish Date: 2015-12-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420>CVE-2015-6420</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420</a></p>
<p>Release Date: 2015-12-15</p>
<p>Fix Resolution: org.apache.commons:commons-collections4:4.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","isTransitiveDependency":false,"dependencyTree":"commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-collections4:4.1"}],"vulnerabilityIdentifier":"CVE-2015-6420","vulnerabilityDetails":"Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2015-6420 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-6420 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to vulnerable library: /JavaVulnerableLabIcatu/target/JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar,/root/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/JavaVulnerableLabIcatu/commit/68aef85b943fe5994864ad4b1984daaa93ad6530">68aef85b943fe5994864ad4b1984daaa93ad6530</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.
<p>Publish Date: 2015-12-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420>CVE-2015-6420</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6420</a></p>
<p>Release Date: 2015-12-15</p>
<p>Fix Resolution: org.apache.commons:commons-collections4:4.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","isTransitiveDependency":false,"dependencyTree":"commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-collections4:4.1"}],"vulnerabilityIdentifier":"CVE-2015-6420","vulnerabilityDetails":"Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6420","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to vulnerable library javavulnerablelabicatu target javavulnerablelab web inf lib commons collections jar root repository commons collections commons collections commons collections jar dependency hierarchy x commons collections jar vulnerable library found in head commit a href vulnerability details serialized object interfaces in certain cisco collaboration and social media endpoint clients and client software network application service and acceleration network and content security devices network management and provisioning routing and switching enterprise and service provider unified computing voice and unified communications devices video streaming telepresence and transcoding devices wireless and cisco hosted services products allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution org apache commons commons rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails serialized object interfaces in certain cisco collaboration and social media endpoint clients and client software network application service and acceleration network and content security devices network management and provisioning routing and switching enterprise and service provider unified computing voice and unified communications devices video streaming telepresence and transcoding devices wireless and cisco hosted services products allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library vulnerabilityurl
| 0
|
16,250
| 20,799,284,015
|
IssuesEvent
|
2022-03-17 12:26:26
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
opened
|
[FLINK-26445] Fix When users specify only consume from 1 partition but the source consumes from all partitions.
|
compute/data-processing type/bug
|
https://issues.apache.org/jira/browse/FLINK-26645
|
1.0
|
[FLINK-26445] Fix When users specify only consume from 1 partition but the source consumes from all partitions. - https://issues.apache.org/jira/browse/FLINK-26645
|
process
|
fix when users specify only consume from partition but the source consumes from all partitions
| 1
|
34,783
| 6,376,761,091
|
IssuesEvent
|
2017-08-02 08:22:45
|
apache/bookkeeper
|
https://api.github.com/repos/apache/bookkeeper
|
opened
|
[documentation] weight-based placement policy
|
area/documentation area/release type/task
|
- general introduction of placement policies
- weight-based placement policy
|
1.0
|
[documentation] weight-based placement policy - - general introduction of placement policies
- weight-based placement policy
|
non_process
|
weight based placement policy general introduction of placement policies weight based placement policy
| 0
|
19,335
| 25,472,614,005
|
IssuesEvent
|
2022-11-25 11:32:06
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[IDP] [PM] Organizational admin is able to login with invalid password
|
Bug P1 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
**Pre-condition:** IDP and MFA should be enabled in the PM
**Steps:**
1. Add organizational user in the participant manager
2. Complete set up your account process
3. Try to sign in with an invalid password and Verify
**AR:** Organizational admin is able to sign in with invalid password
**ER:** Organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
|
3.0
|
[IDP] [PM] Organizational admin is able to login with invalid password - **Pre-condition:** IDP and MFA should be enabled in the PM
**Steps:**
1. Add organizational user in the participant manager
2. Complete set up your account process
3. Try to sign in with an invalid password and Verify
**AR:** Organizational admin is able to sign in with invalid password
**ER:** Organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
|
process
|
organizational admin is able to login with invalid password pre condition idp and mfa should be enabled in the pm steps add organizational user in the participant manager complete set up your account process try to sign in with an invalid password and verify ar organizational admin is able to sign in with invalid password er organizational admin should not be able to sign in with an invalid password and a proper validation message should get displayed
| 1
|
106,869
| 16,735,432,893
|
IssuesEvent
|
2021-06-11 00:08:39
|
AlexRogalskiy/qiitos
|
https://api.github.com/repos/AlexRogalskiy/qiitos
|
closed
|
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed
|
security vulnerability
|
## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: qiitos/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: qiitos/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/qiitos/commit/9ea0c7ba4083c92f216516728105e2b17415771e">9ea0c7ba4083c92f216516728105e2b17415771e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: qiitos/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: qiitos/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/qiitos/commit/9ea0c7ba4083c92f216516728105e2b17415771e">9ea0c7ba4083c92f216516728105e2b17415771e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file qiitos node modules redeyed examples browser index html path to vulnerable library qiitos node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
184,141
| 14,272,717,105
|
IssuesEvent
|
2020-11-21 18:14:48
|
cybersemics/em
|
https://api.github.com/repos/cybersemics/em
|
opened
|
Upgrade to react-scripts v4
|
test
|
The [pullQueue tests](https://github.com/cybersemics/em/blob/dev/src/redux-middleware/__tests__/pullQueue.ts) fail on `react-scripts` v4. It is related to the fake timer and mock debounce logic, but I can't figure out how to fix it. We should be able to use `useFakeTimers('modern')` now and not have to mock lodash's debounce, but that causes an error.
Here are the CHANGELOGS although I didn't find anything relevant:
- create-react-app CHANGELOG: https://github.com/facebook/create-react-app/blob/master/CHANGELOG.md
- jest CHANGELOG: https://github.com/facebook/jest/blob/master/CHANGELOG.md
|
1.0
|
Upgrade to react-scripts v4 - The [pullQueue tests](https://github.com/cybersemics/em/blob/dev/src/redux-middleware/__tests__/pullQueue.ts) fail on `react-scripts` v4. It is related to the fake timer and mock debounce logic, but I can't figure out how to fix it. We should be able to use `useFakeTimers('modern')` now and not have to mock lodash's debounce, but that causes an error.
Here are the CHANGELOGS although I didn't find anything relevant:
- create-react-app CHANGELOG: https://github.com/facebook/create-react-app/blob/master/CHANGELOG.md
- jest CHANGELOG: https://github.com/facebook/jest/blob/master/CHANGELOG.md
|
non_process
|
upgrade to react scripts the fail on react scripts it is related to the fake timer and mock debounce logic but i can t figure out how to fix it we should be able to use usefaketimers modern now and not have to mock lodash s debounce but that causes an error here are the changelogs although i didn t find anything relevant create react app changelog jest changelog
| 0
|
185,914
| 14,384,565,803
|
IssuesEvent
|
2020-12-02 10:37:57
|
hashicorp/nomad
|
https://api.github.com/repos/hashicorp/nomad
|
opened
|
e2e: flakey TestMetricsLinux test
|
theme/flaky-tests type/bug
|
The [Linux alloc metrics e2e test](https://github.com/hashicorp/nomad/blob/5fbcf82ceefe5facacaff31f45e04ccd89fed3ac/e2e/metrics/metrics.go#L89) occasinally fails with the following error:
```
prometheus query failed (nomad_client_allocs_memory_usage{exported_job="memstress"}): expected non-zero metrics, got: nomad_client_allocs_memory_usage{alloc_id="70ed194c-e5e5-099b-d000-17cc0ad6920e", exported_job="memstress", instance="172.31.13.159:4646", job="nomad_metrics", namespace="default", task="memstress", task_group="memstress"} => 0 @[1606903456.227]
```
|
1.0
|
e2e: flakey TestMetricsLinux test - The [Linux alloc metrics e2e test](https://github.com/hashicorp/nomad/blob/5fbcf82ceefe5facacaff31f45e04ccd89fed3ac/e2e/metrics/metrics.go#L89) occasinally fails with the following error:
```
prometheus query failed (nomad_client_allocs_memory_usage{exported_job="memstress"}): expected non-zero metrics, got: nomad_client_allocs_memory_usage{alloc_id="70ed194c-e5e5-099b-d000-17cc0ad6920e", exported_job="memstress", instance="172.31.13.159:4646", job="nomad_metrics", namespace="default", task="memstress", task_group="memstress"} => 0 @[1606903456.227]
```
|
non_process
|
flakey testmetricslinux test the occasinally fails with the following error prometheus query failed nomad client allocs memory usage exported job memstress expected non zero metrics got nomad client allocs memory usage alloc id exported job memstress instance job nomad metrics namespace default task memstress task group memstress
| 0
|
146,474
| 19,404,088,479
|
IssuesEvent
|
2021-12-19 17:49:57
|
vincenzodistasio97/BenchmarkJava
|
https://api.github.com/repos/vincenzodistasio97/BenchmarkJava
|
opened
|
CVE-2021-22096 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2021-22096 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-core-4.3.30.RELEASE.jar</b>, <b>spring-webmvc-4.3.30.RELEASE.jar</b>, <b>spring-web-4.3.30.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-core-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.30.RELEASE/spring-core-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-context-4.3.30.RELEASE.jar (Root Library)
- :x: **spring-core-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-webmvc-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Web MVC</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: ository/org/springframework/spring-webmvc/4.3.30.RELEASE/spring-webmvc-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-webmvc-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-web-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: ository/org/springframework/spring-web/4.3.30.RELEASE/spring-web-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-web-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/BenchmarkJava/commit/0f7ddc18e36494a62123563264c65e3344dae7af">0f7ddc18e36494a62123563264c65e3344dae7af</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-22096 (Medium) detected in multiple libraries - ## CVE-2021-22096 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-core-4.3.30.RELEASE.jar</b>, <b>spring-webmvc-4.3.30.RELEASE.jar</b>, <b>spring-web-4.3.30.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-core-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.30.RELEASE/spring-core-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-context-4.3.30.RELEASE.jar (Root Library)
- :x: **spring-core-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-webmvc-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Web MVC</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: ository/org/springframework/spring-webmvc/4.3.30.RELEASE/spring-webmvc-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-webmvc-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-web-4.3.30.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: BenchmarkJava/pom.xml</p>
<p>Path to vulnerable library: ository/org/springframework/spring-web/4.3.30.RELEASE/spring-web-4.3.30.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-web-4.3.30.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/BenchmarkJava/commit/0f7ddc18e36494a62123563264c65e3344dae7af">0f7ddc18e36494a62123563264c65e3344dae7af</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework:spring-core:5.2.18.RELEASE,5.3.12;org.springframework:spring-web:5.2.18.RELEASE,5.3.12;org.springframework:spring-webmvc:5.2.18.RELEASE,5.3.12;org.springframework:spring-webflux:5.2.18.RELEASE,5.3.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries spring core release jar spring webmvc release jar spring web release jar spring core release jar spring core library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library home wss scanner repository org springframework spring core release spring core release jar dependency hierarchy spring context release jar root library x spring core release jar vulnerable library spring webmvc release jar spring web mvc library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library ository org springframework spring webmvc release spring webmvc release jar dependency hierarchy x spring webmvc release jar vulnerable library spring web release jar spring web library home page a href path to dependency file benchmarkjava pom xml path to vulnerable library ository org springframework spring web release spring web release jar dependency hierarchy x spring web release jar vulnerable library found in head commit a href found in base branch master vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release org springframework spring web release org springframework spring webmvc release org springframework spring webflux release step up your open source security game with whitesource
| 0
|
23,645
| 22,441,322,602
|
IssuesEvent
|
2022-06-21 01:39:16
|
bevyengine/bevy
|
https://api.github.com/repos/bevyengine/bevy
|
opened
|
Add a method to get a `ComponentId` from a `Component`
|
A-ECS C-Usability
|
## What problem does this solve or what need does it fill?
Being able to extract the `ComponentId` for a given component of type `C` is surprisingly unintuitive.
## What solution would you like?
Create a public method on `Component` that mirrors [`Bundle::component_ids()`](https://dev-docs.bevyengine.org/bevy/ecs/bundle/trait.Bundle.html#tymethod.component_ids).
## What alternative(s) have you considered?
1. Use `world.init_component::<C>()`. This is unintuitive, but basically works.
2. Convert the component type into a bundle, and call `(C, )::component_ids().iter().next().unwrap()`. Which is more than a little cursed.
## Additional context
Discovered while working on #1481 with @plof27.
|
True
|
Add a method to get a `ComponentId` from a `Component` - ## What problem does this solve or what need does it fill?
Being able to extract the `ComponentId` for a given component of type `C` is surprisingly unintuitive.
## What solution would you like?
Create a public method on `Component` that mirrors [`Bundle::component_ids()`](https://dev-docs.bevyengine.org/bevy/ecs/bundle/trait.Bundle.html#tymethod.component_ids).
## What alternative(s) have you considered?
1. Use `world.init_component::<C>()`. This is unintuitive, but basically works.
2. Convert the component type into a bundle, and call `(C, )::component_ids().iter().next().unwrap()`. Which is more than a little cursed.
## Additional context
Discovered while working on #1481 with @plof27.
|
non_process
|
add a method to get a componentid from a component what problem does this solve or what need does it fill being able to extract the componentid for a given component of type c is surprisingly unintuitive what solution would you like create a public method on component that mirrors what alternative s have you considered use world init component this is unintuitive but basically works convert the component type into a bundle and call c component ids iter next unwrap which is more than a little cursed additional context discovered while working on with
| 0
|
37,298
| 8,343,517,833
|
IssuesEvent
|
2018-09-30 05:44:03
|
liuxuewei/bluebee-accounting-system
|
https://api.github.com/repos/liuxuewei/bluebee-accounting-system
|
closed
|
反馈
|
Priority-Medium Type-Defect auto-migrated
|
```
软件还是比较好的,试用了一下午,发现了不上问题,有时��
�会响应比较慢,添加的商品要重启软件之后才能出来,进货�
��保存无返回值,继续点保存会保存多条数据,总之还是有不
少bug的,希望开发者再接再励!!
```
Original issue reported on code.google.com by `wsjta...@gmail.com` on 18 Aug 2013 at 5:06
|
1.0
|
反馈 - ```
软件还是比较好的,试用了一下午,发现了不上问题,有时��
�会响应比较慢,添加的商品要重启软件之后才能出来,进货�
��保存无返回值,继续点保存会保存多条数据,总之还是有不
少bug的,希望开发者再接再励!!
```
Original issue reported on code.google.com by `wsjta...@gmail.com` on 18 Aug 2013 at 5:06
|
non_process
|
反馈 软件还是比较好的,试用了一下午,发现了不上问题,有时�� �会响应比较慢,添加的商品要重启软件之后才能出来,进货� ��保存无返回值,继续点保存会保存多条数据,总之还是有不 少bug的,希望开发者再接再励!! original issue reported on code google com by wsjta gmail com on aug at
| 0
|
779,638
| 27,360,959,923
|
IssuesEvent
|
2023-02-27 15:50:29
|
rangav/thunder-client-support
|
https://api.github.com/repos/rangav/thunder-client-support
|
closed
|
Ability to Format Graphql Body
|
feature request Priority
|
Could you add the feature to format a Graphql body code snippet, please. That featured is available for JSON and XML.
|
1.0
|
Ability to Format Graphql Body - Could you add the feature to format a Graphql body code snippet, please. That featured is available for JSON and XML.
|
non_process
|
ability to format graphql body could you add the feature to format a graphql body code snippet please that featured is available for json and xml
| 0
|
8,770
| 11,886,457,510
|
IssuesEvent
|
2020-03-27 21:59:20
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
child_process.exec/execFile docs have some inconsistencies and inaccuracies
|
child_process doc
|
I don't have time to fix this now, but while looking at sec issues related to these APIs, I found some oddities.
https://nodejs.org/api/child_process.html#child_process_child_process_execfile_file_args_options_callback says
> shell <boolean> | <string> If true, runs command inside of a shell.
But execFile() doesn't have a `command` argument... this was pasted from exec(), it seems. Probably what happens is that if there is a shell, then `file` and `args` are all concatenated together, `' '` seperated, and passed to the shell.
Since the shell option AFAICT ends up following the same path as from exec(), it suggests that the exec docs:
> shell <string> Shell to execute the command with. See Shell Requirements and Default Windows Shell. Default: '/bin/sh' on Unix, process.env.ComSpec on Windows.
are incomplete, probably `false` would work just fine as an arg there, making exec() behave exactly like execFile().
This seems to be a bit legacy as well:
> The child_process.execFile() function is similar to child_process.exec() except that it does not spawn a shell by default. Rather, the specified executable file is spawned directly as a new process making it slightly more efficient than child_process.exec().
Now that both exec and execFile() have a shell, differing only be the default value, its probably more accurate to say the difference is that one takes an array of strings as an argument `execFile(file, argv, ..` and the other takes a single string `exec(command, ...)`.
The text following is now wrong:
> The same options as child_process.exec() are supported. Since a shell is not spawned, behaviors such as I/O redirection and file globbing are not supported.
It can't both *support* the same options, and *not support* some of the options.
It should probably say "If a shell is ..." (only one word different, but its important).
exec should probably have docs saying the same thing, shell behaviours are not supported when shell is `false`.
And execFile() should probably include the warnings from exec about how shell special chars vary by platform.
Some of these issues are shared with the "sync" versions of the APIs.
|
1.0
|
child_process.exec/execFile docs have some inconsistencies and inaccuracies - I don't have time to fix this now, but while looking at sec issues related to these APIs, I found some oddities.
https://nodejs.org/api/child_process.html#child_process_child_process_execfile_file_args_options_callback says
> shell <boolean> | <string> If true, runs command inside of a shell.
But execFile() doesn't have a `command` argument... this was pasted from exec(), it seems. Probably what happens is that if there is a shell, then `file` and `args` are all concatenated together, `' '` seperated, and passed to the shell.
Since the shell option AFAICT ends up following the same path as from exec(), it suggests that the exec docs:
> shell <string> Shell to execute the command with. See Shell Requirements and Default Windows Shell. Default: '/bin/sh' on Unix, process.env.ComSpec on Windows.
are incomplete, probably `false` would work just fine as an arg there, making exec() behave exactly like execFile().
This seems to be a bit legacy as well:
> The child_process.execFile() function is similar to child_process.exec() except that it does not spawn a shell by default. Rather, the specified executable file is spawned directly as a new process making it slightly more efficient than child_process.exec().
Now that both exec and execFile() have a shell, differing only be the default value, its probably more accurate to say the difference is that one takes an array of strings as an argument `execFile(file, argv, ..` and the other takes a single string `exec(command, ...)`.
The text following is now wrong:
> The same options as child_process.exec() are supported. Since a shell is not spawned, behaviors such as I/O redirection and file globbing are not supported.
It can't both *support* the same options, and *not support* some of the options.
It should probably say "If a shell is ..." (only one word different, but its important).
exec should probably have docs saying the same thing, shell behaviours are not supported when shell is `false`.
And execFile() should probably include the warnings from exec about how shell special chars vary by platform.
Some of these issues are shared with the "sync" versions of the APIs.
|
process
|
child process exec execfile docs have some inconsistencies and inaccuracies i don t have time to fix this now but while looking at sec issues related to these apis i found some oddities says shell if true runs command inside of a shell but execfile doesn t have a command argument this was pasted from exec it seems probably what happens is that if there is a shell then file and args are all concatenated together seperated and passed to the shell since the shell option afaict ends up following the same path as from exec it suggests that the exec docs shell shell to execute the command with see shell requirements and default windows shell default bin sh on unix process env comspec on windows are incomplete probably false would work just fine as an arg there making exec behave exactly like execfile this seems to be a bit legacy as well the child process execfile function is similar to child process exec except that it does not spawn a shell by default rather the specified executable file is spawned directly as a new process making it slightly more efficient than child process exec now that both exec and execfile have a shell differing only be the default value its probably more accurate to say the difference is that one takes an array of strings as an argument execfile file argv and the other takes a single string exec command the text following is now wrong the same options as child process exec are supported since a shell is not spawned behaviors such as i o redirection and file globbing are not supported it can t both support the same options and not support some of the options it should probably say if a shell is only one word different but its important exec should probably have docs saying the same thing shell behaviours are not supported when shell is false and execfile should probably include the warnings from exec about how shell special chars vary by platform some of these issues are shared with the sync versions of the apis
| 1
|
15,437
| 19,651,448,236
|
IssuesEvent
|
2022-01-10 07:44:59
|
linuxdeepin/developer-center
|
https://api.github.com/repos/linuxdeepin/developer-center
|
closed
|
Terminal blur is displaced with deepin-kwin
|
bug | functional behavior other | delay processing
|
Bug: with deepin-kwin, the blur on the terminal seems to be displaced.

Update: issue is not present when the terminal is fullscreen.
|
1.0
|
Terminal blur is displaced with deepin-kwin - Bug: with deepin-kwin, the blur on the terminal seems to be displaced.

Update: issue is not present when the terminal is fullscreen.
|
process
|
terminal blur is displaced with deepin kwin bug with deepin kwin the blur on the terminal seems to be displaced update issue is not present when the terminal is fullscreen
| 1
|
472,778
| 13,631,234,447
|
IssuesEvent
|
2020-09-24 17:40:56
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
opened
|
[0.9.0.2 beta develop-62] Government Bank permissions further refinement
|
Priority: Medium Type: Exploit
|
When a player is missing a title with permissions "Has Government Banking privileges", it should disable government tagged account access regardless of the players manager/use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an "offshore tax free" account.
|
1.0
|
[0.9.0.2 beta develop-62] Government Bank permissions further refinement - When a player is missing a title with permissions "Has Government Banking privileges", it should disable government tagged account access regardless of the players manager/use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an "offshore tax free" account.
|
non_process
|
government bank permissions further refinement when a player is missing a title with permissions has government banking privileges it should disable government tagged account access regardless of the players manager use permissions and only allow to add managers but not actually use it for purchases or transferring money out of it this to avoid any issues of tax evasion using a government account as an offshore tax free account
| 0
|
197,776
| 6,963,758,753
|
IssuesEvent
|
2017-12-08 18:43:48
|
elementary/gala
|
https://api.github.com/repos/elementary/gala
|
opened
|
Send a persistent notification after taking a screenshot
|
Priority: Wishlist
|
This is a migration of https://github.com/elementary/screenshot-tool/issues/30
After taking a screenshot, Gala should send a persistent notification using the screenshot preview as the notification image. Clicking the notification should open the image in the default image viewer.
This would give:
1. A visual queue that the screenshot was completed, useful when sound is disabled or the user is not able to hear
2. A way to quickly view/edit/share the screenshot without having to know where screenshots are stored on the filesystem
|
1.0
|
Send a persistent notification after taking a screenshot - This is a migration of https://github.com/elementary/screenshot-tool/issues/30
After taking a screenshot, Gala should send a persistent notification using the screenshot preview as the notification image. Clicking the notification should open the image in the default image viewer.
This would give:
1. A visual queue that the screenshot was completed, useful when sound is disabled or the user is not able to hear
2. A way to quickly view/edit/share the screenshot without having to know where screenshots are stored on the filesystem
|
non_process
|
send a persistent notification after taking a screenshot this is a migration of after taking a screenshot gala should send a persistent notification using the screenshot preview as the notification image clicking the notification should open the image in the default image viewer this would give a visual queue that the screenshot was completed useful when sound is disabled or the user is not able to hear a way to quickly view edit share the screenshot without having to know where screenshots are stored on the filesystem
| 0
|
7,470
| 10,566,758,795
|
IssuesEvent
|
2019-10-05 21:13:21
|
bow-simulation/virtualbow
|
https://api.github.com/repos/bow-simulation/virtualbow
|
opened
|
Decide if macOS support is still feasible/worthwhile with Apples new notarization requirements
|
area: software process platform: macos type: discussion type: help wanted
|
[Reddit discussion](https://www.reddit.com/r/programming/comments/ddbxmw/sorry_macos_users_but_apple_has_gone_too_far_for/)
|
1.0
|
Decide if macOS support is still feasible/worthwhile with Apples new notarization requirements - [Reddit discussion](https://www.reddit.com/r/programming/comments/ddbxmw/sorry_macos_users_but_apple_has_gone_too_far_for/)
|
process
|
decide if macos support is still feasible worthwhile with apples new notarization requirements
| 1
|
450,822
| 13,019,845,030
|
IssuesEvent
|
2020-07-27 00:52:32
|
momentum-mod/game
|
https://api.github.com/repos/momentum-mod/game
|
closed
|
Stickybomb particles depend on TF2
|
Priority: High Size: Small Type: Enhancement
|
**What feature is your improvement idea related to? Please describe.**
Currently, the stickybomb particle depends on the TF2 material `effects/circle2`. Not having TF2 installed when using the sticky launcher results in this error: `C_OP_RenderSprites::RenderUnsorted: Attempting to use an unimplemented sprite renderer for system "mom_stickypulse_blue"!`
**Describe the solution you'd like**
Either add the material from TF2 or replace it with another that we do have.
**Additional context**
Our base Momentum assets shouldn't depend on any other game being installed. In the future when adding assets should make sure that all materials used are included in the game.
|
1.0
|
Stickybomb particles depend on TF2 - **What feature is your improvement idea related to? Please describe.**
Currently, the stickybomb particle depends on the TF2 material `effects/circle2`. Not having TF2 installed when using the sticky launcher results in this error: `C_OP_RenderSprites::RenderUnsorted: Attempting to use an unimplemented sprite renderer for system "mom_stickypulse_blue"!`
**Describe the solution you'd like**
Either add the material from TF2 or replace it with another that we do have.
**Additional context**
Our base Momentum assets shouldn't depend on any other game being installed. In the future when adding assets should make sure that all materials used are included in the game.
|
non_process
|
stickybomb particles depend on what feature is your improvement idea related to please describe currently the stickybomb particle depends on the material effects not having installed when using the sticky launcher results in this error c op rendersprites renderunsorted attempting to use an unimplemented sprite renderer for system mom stickypulse blue describe the solution you d like either add the material from or replace it with another that we do have additional context our base momentum assets shouldn t depend on any other game being installed in the future when adding assets should make sure that all materials used are included in the game
| 0
|
52,399
| 13,224,714,621
|
IssuesEvent
|
2020-08-17 19:41:52
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
segments in wimpsimreader.py (Trac #2161)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2161">https://code.icecube.wisc.edu/projects/icecube/ticket/2161</a>, reported by grenziand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"_ts": "1550067323910946",
"description": "Both segments at [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/python/wimpsimreader.py] are supposed to launch I3WimpSimReader module, but they are set with non acceptable values of variable InjectionRadius. The real default is NaN, and not 0 as visible here [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/private/wimpsim-reader/I3WimpSimReader.cxx]. The variable has this function (from docs):\n\n''InjectionRadius [Default=NAN] If >0, events will be injected in cylinder with specified radius and [zmin, zmax] height instead of rectangular box''\n\nIf the original intention was to avoid the cylinder, this value should be left to NaN, as default. (Or the value 0 should be accepted in I3 module at lines 180-181)\n\nFor WimpSimReaderEarth there is an issue with GCD, the code isn't able to read time in given format:\n\n{{{No registered converter was able to produce a C++ rvalue of type std::string from this Python object of type I3Time}}}\n\nTo avoid it, time should be taken as julian day (or another \"good\" format), so for example\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time}}}\n\ncould be:\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time.mod_julian_day_double}}}",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"time": "2018-06-12T10:09:52",
"component": "combo simulation",
"summary": "segments in wimpsimreader.py",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
segments in wimpsimreader.py (Trac #2161) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2161">https://code.icecube.wisc.edu/projects/icecube/ticket/2161</a>, reported by grenziand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"_ts": "1550067323910946",
"description": "Both segments at [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/python/wimpsimreader.py] are supposed to launch I3WimpSimReader module, but they are set with non acceptable values of variable InjectionRadius. The real default is NaN, and not 0 as visible here [http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/wimpsim-reader/trunk/private/wimpsim-reader/I3WimpSimReader.cxx]. The variable has this function (from docs):\n\n''InjectionRadius [Default=NAN] If >0, events will be injected in cylinder with specified radius and [zmin, zmax] height instead of rectangular box''\n\nIf the original intention was to avoid the cylinder, this value should be left to NaN, as default. (Or the value 0 should be accepted in I3 module at lines 180-181)\n\nFor WimpSimReaderEarth there is an issue with GCD, the code isn't able to read time in given format:\n\n{{{No registered converter was able to produce a C++ rvalue of type std::string from this Python object of type I3Time}}}\n\nTo avoid it, time should be taken as julian day (or another \"good\" format), so for example\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time}}}\n\ncould be:\n\n{{{line 55: return frame.Get(\"I3DetectorStatus\").start_time.mod_julian_day_double}}}",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"time": "2018-06-12T10:09:52",
"component": "combo simulation",
"summary": "segments in wimpsimreader.py",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
segments in wimpsimreader py trac migrated from json status closed changetime ts description both segments at are supposed to launch module but they are set with non acceptable values of variable injectionradius the real default is nan and not as visible here the variable has this function from docs n n injectionradius if events will be injected in cylinder with specified radius and height instead of rectangular box n nif the original intention was to avoid the cylinder this value should be left to nan as default or the value should be accepted in module at lines n nfor wimpsimreaderearth there is an issue with gcd the code isn t able to read time in given format n n no registered converter was able to produce a c rvalue of type std string from this python object of type n nto avoid it time should be taken as julian day or another good format so for example n n line return frame get start time n ncould be n n line return frame get start time mod julian day double reporter grenzi cc resolution fixed time component combo simulation summary segments in wimpsimreader py priority normal keywords milestone owner nega type defect
| 0
|
142,757
| 5,477,062,516
|
IssuesEvent
|
2017-03-12 03:27:36
|
NCEAS/eml
|
https://api.github.com/repos/NCEAS/eml
|
closed
|
DataManager cuts last row off data_table
|
Category: datamanager Component: Bugzilla-Id Priority: Normal Status: Resolved Tracker: Bug
|
---
Author Name: **Chad Burt** (Chad Burt)
Original Redmine Issue: 2775, https://projects.ecoinformatics.org/ecoinfo/issues/2775
Original Date: 2007-02-13
Original Assignee: ben leinfelder
---
Had a problem putting this data into a database:
http://data.piscoweb.org/catalog/metacat?action=read&docid=cburt.5.2&qformat=pisco
It looks like the last row,
(2006-02-01,3,10,"test fixtures are fun")
is getting cut off.
I haven't tested a dataset without a header file. I suspect either the header is being counted against it in a loop or there is a < where a <= would be more appropriate.
|
1.0
|
DataManager cuts last row off data_table - ---
Author Name: **Chad Burt** (Chad Burt)
Original Redmine Issue: 2775, https://projects.ecoinformatics.org/ecoinfo/issues/2775
Original Date: 2007-02-13
Original Assignee: ben leinfelder
---
Had a problem putting this data into a database:
http://data.piscoweb.org/catalog/metacat?action=read&docid=cburt.5.2&qformat=pisco
It looks like the last row,
(2006-02-01,3,10,"test fixtures are fun")
is getting cut off.
I haven't tested a dataset without a header file. I suspect either the header is being counted against it in a loop or there is a < where a <= would be more appropriate.
|
non_process
|
datamanager cuts last row off data table author name chad burt chad burt original redmine issue original date original assignee ben leinfelder had a problem putting this data into a database it looks like the last row test fixtures are fun is getting cut off i haven t tested a dataset without a header file i suspect either the header is being counted against it in a loop or there is a where a would be more appropriate
| 0
|
174,897
| 6,544,245,362
|
IssuesEvent
|
2017-09-03 13:45:10
|
nim-lang/Nim
|
https://api.github.com/repos/nim-lang/Nim
|
closed
|
Minor issue : compiling with doc2 throws Error: ambiguous call
|
Easy Low Priority
|
Simple example:
```nim
# test.nim
import os
## Documentation here
echo("User home dir " , getHomeDir())
```
Compile with nim doc2 test
```nim
test.nim(3, 36) Error: ambiguous call; both ospaths.getHomeDir() and os.getHomeDir() match for: ()
```
The error seems to have no effect and the file test.html is created fine.
|
1.0
|
Minor issue : compiling with doc2 throws Error: ambiguous call - Simple example:
```nim
# test.nim
import os
## Documentation here
echo("User home dir " , getHomeDir())
```
Compile with nim doc2 test
```nim
test.nim(3, 36) Error: ambiguous call; both ospaths.getHomeDir() and os.getHomeDir() match for: ()
```
The error seems to have no effect and the file test.html is created fine.
|
non_process
|
minor issue compiling with throws error ambiguous call simple example nim test nim import os documentation here echo user home dir gethomedir compile with nim test nim test nim error ambiguous call both ospaths gethomedir and os gethomedir match for the error seems to have no effect and the file test html is created fine
| 0
|
50,316
| 6,354,218,497
|
IssuesEvent
|
2017-07-29 07:13:20
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
C# Pure Function Keyword to Mark No Side Effects Or External Dependencies
|
Area-Language Design Discussion
|
Be able to mark a method as having no side effects or external dependencies, ie: it does not change any state outside the inputs or outputs. Any code that attempts to do this would throw an exception. My thought was that the keyword could be "functional", "pure" (as in a "pure functions" mentioned in some Msdn documentation ), "purefunction", or even "nosideffects".
See https://msdn.microsoft.com/en-us/library/bb669139.aspx for some current naming conventions and reasons for this feature.
|
1.0
|
C# Pure Function Keyword to Mark No Side Effects Or External Dependencies - Be able to mark a method as having no side effects or external dependencies, ie: it does not change any state outside the inputs or outputs. Any code that attempts to do this would throw an exception. My thought was that the keyword could be "functional", "pure" (as in a "pure functions" mentioned in some Msdn documentation ), "purefunction", or even "nosideffects".
See https://msdn.microsoft.com/en-us/library/bb669139.aspx for some current naming conventions and reasons for this feature.
|
non_process
|
c pure function keyword to mark no side effects or external dependencies be able to mark a method as having no side effects or external dependencies ie it does not change any state outside the inputs or outputs any code that attempts to do this would throw an exception my thought was that the keyword could be functional pure as in a pure functions mentioned in some msdn documentation purefunction or even nosideffects see for some current naming conventions and reasons for this feature
| 0
|
61,698
| 7,495,240,600
|
IssuesEvent
|
2018-04-07 18:41:36
|
NYU-Shopcarts/shopcarts
|
https://api.github.com/repos/NYU-Shopcarts/shopcarts
|
closed
|
Create the design for all CRUD routes
|
design
|
**As a** developler
**I need** to know the what to name the API routes
**So that** I can create a standardized API
**Assumptions:**
* We can create and publish a documentation for the Shopcarts API
**Acceptance Criteria:**
```
Given that I'm a user of the Shopcarts API
When my client calls the CRUD routes on the Shopcarts API
Then my products gets created, updated, deleted, or fetched from the Shopcarts DB
```
|
1.0
|
Create the design for all CRUD routes - **As a** developler
**I need** to know the what to name the API routes
**So that** I can create a standardized API
**Assumptions:**
* We can create and publish a documentation for the Shopcarts API
**Acceptance Criteria:**
```
Given that I'm a user of the Shopcarts API
When my client calls the CRUD routes on the Shopcarts API
Then my products gets created, updated, deleted, or fetched from the Shopcarts DB
```
|
non_process
|
create the design for all crud routes as a developler i need to know the what to name the api routes so that i can create a standardized api assumptions we can create and publish a documentation for the shopcarts api acceptance criteria given that i m a user of the shopcarts api when my client calls the crud routes on the shopcarts api then my products gets created updated deleted or fetched from the shopcarts db
| 0
|
12,553
| 14,976,935,706
|
IssuesEvent
|
2021-01-28 08:48:52
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
failed to deploy taiga
|
process_wontfix
|
Failed to initialize peertube, please contact support with this information: Domain: sashaastiadi-peertube-sasha11.webg2test.grid.tf VDC Name: vdc_vdctest4_sashaastiadi Farm name: freefarm Reason: Couldn't find resources in the cluster for the solution . Use the refresh button on the upper right to restart Peertube creation
|
1.0
|
failed to deploy taiga - Failed to initialize peertube, please contact support with this information: Domain: sashaastiadi-peertube-sasha11.webg2test.grid.tf VDC Name: vdc_vdctest4_sashaastiadi Farm name: freefarm Reason: Couldn't find resources in the cluster for the solution . Use the refresh button on the upper right to restart Peertube creation
|
process
|
failed to deploy taiga failed to initialize peertube please contact support with this information domain sashaastiadi peertube grid tf vdc name vdc sashaastiadi farm name freefarm reason couldn t find resources in the cluster for the solution use the refresh button on the upper right to restart peertube creation
| 1
|
17,347
| 23,171,755,230
|
IssuesEvent
|
2022-07-30 20:40:47
|
MartinBruun/P6
|
https://api.github.com/repos/MartinBruun/P6
|
closed
|
(CI) Setup self-hosted github runner, migrating to private repo
|
Need PR 2: Must have Need Tests Process
|
**Who wants it?**
Product Owner: Martin BM
**What is the proposed change?**
Setup a self-hosted runner (maybe on a Raspberry Pi) to run the github actions, so the project can be made possible to migrate to a private repo, in case this will be a business idea.
Follow these guidelines:
https://docs.github.com/en/actions/hosting-your-own-runners/about-self-hosted-runners
|
1.0
|
(CI) Setup self-hosted github runner, migrating to private repo - **Who wants it?**
Product Owner: Martin BM
**What is the proposed change?**
Setup a self-hosted runner (maybe on a Raspberry Pi) to run the github actions, so the project can be made possible to migrate to a private repo, in case this will be a business idea.
Follow these guidelines:
https://docs.github.com/en/actions/hosting-your-own-runners/about-self-hosted-runners
|
process
|
ci setup self hosted github runner migrating to private repo who wants it product owner martin bm what is the proposed change setup a self hosted runner maybe on a raspberry pi to run the github actions so the project can be made possible to migrate to a private repo in case this will be a business idea follow these guidelines
| 1
|
489,623
| 14,109,019,520
|
IssuesEvent
|
2020-11-06 18:50:03
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Python wheel missing for grpcio version > 1.27.2
|
kind/bug lang/Python priority/P2
|
<!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
Python 3.6.8
grpcio==1.33.2
### What operating system (Linux, Windows,...) and version?
Centos 7
Ubuntu 18..04
### What runtime / compiler are you using (e.g. python version or version of gcc)
gcc (GCC) 4.8.5 20150623 (Red Hat 4.8.5-39)
gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
### What did you do?
pip3 install grpcio==1.33.2 --only-binary ":all:"
### What did you expect to see?
Should simply install the package using an existing python wheel
### What did you see instead?
Based on the output below, which happens on both Centos 7 and Ubuntu 18.04, there isn't a python wheel for anything above grpcio 1.27.2.
```
sudo pip3 install grpcio==1.33.2 --only-binary ":all:"
WARNING: Running pip install with root privileges is generally not a good idea. Try `pip3 install --user` instead.
Collecting grpcio==1.33.2
Could not find a version that satisfies the requirement grpcio==1.33.2 (from versions: 1.1.3, 1.2.0rc1, 1.2.0rc2, 1.2.0, 1.2.1, 1.3.0, 1.3.3, 1.3.5, 1.4.0rc1, 1.4.0, 1.6.0rc1, 1.6.0, 1.6.3, 1.7.0rc1, 1.7.0, 1.7.3, 1.8.0rc1, 1.8.0rc2, 1.8.0rc3, 1.8.1, 1.8.2, 1.8.3, 1.8.4, 1.8.6, 1.9.0rc3, 1.9.0, 1.9.1, 1.10.0rc1, 1.10.0rc2, 1.10.0, 1.10.1rc1, 1.10.1rc2, 1.10.1, 1.11.0rc1, 1.11.0rc2, 1.11.0, 1.11.1rc1, 1.11.1, 1.12.0rc1, 1.12.0, 1.12.1, 1.13.0rc1, 1.13.0rc2, 1.13.0rc3, 1.13.0, 1.14.0rc1, 1.14.0rc2, 1.14.0, 1.14.1, 1.14.2rc1, 1.14.2, 1.15.0rc1, 1.15.0, 1.16.0rc1, 1.16.0, 1.16.1rc1, 1.16.1, 1.17.0rc1, 1.17.0, 1.17.1rc1, 1.17.1, 1.18.0rc1, 1.18.0, 1.19.0rc1, 1.19.0, 1.20.0rc1, 1.20.0rc2, 1.20.0rc3, 1.20.0, 1.20.1, 1.21.0rc1, 1.21.1rc1, 1.21.1, 1.22.0rc1, 1.22.0, 1.22.1, 1.23.0rc1, 1.23.0, 1.23.1, 1.24.0rc1, 1.24.0, 1.24.1, 1.24.3, 1.25.0rc1, 1.25.0, 1.26.0rc1, 1.26.0, 1.27.0rc1, 1.27.1, 1.27.2)
No matching distribution found for grpcio==1.33.2
```
### Anything else we should know about your project / environment?
We have to deploy this to multiple systems. If we don't use `--only-binary`, our deployments have to wait for grpci to compile on every system and it causes deployments to take forever.
|
1.0
|
Python wheel missing for grpcio version > 1.27.2 - <!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
Python 3.6.8
grpcio==1.33.2
### What operating system (Linux, Windows,...) and version?
Centos 7
Ubuntu 18..04
### What runtime / compiler are you using (e.g. python version or version of gcc)
gcc (GCC) 4.8.5 20150623 (Red Hat 4.8.5-39)
gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
### What did you do?
pip3 install grpcio==1.33.2 --only-binary ":all:"
### What did you expect to see?
Should simply install the package using an existing python wheel
### What did you see instead?
Based on the output below, which happens on both Centos 7 and Ubuntu 18.04, there isn't a python wheel for anything above grpcio 1.27.2.
```
sudo pip3 install grpcio==1.33.2 --only-binary ":all:"
WARNING: Running pip install with root privileges is generally not a good idea. Try `pip3 install --user` instead.
Collecting grpcio==1.33.2
Could not find a version that satisfies the requirement grpcio==1.33.2 (from versions: 1.1.3, 1.2.0rc1, 1.2.0rc2, 1.2.0, 1.2.1, 1.3.0, 1.3.3, 1.3.5, 1.4.0rc1, 1.4.0, 1.6.0rc1, 1.6.0, 1.6.3, 1.7.0rc1, 1.7.0, 1.7.3, 1.8.0rc1, 1.8.0rc2, 1.8.0rc3, 1.8.1, 1.8.2, 1.8.3, 1.8.4, 1.8.6, 1.9.0rc3, 1.9.0, 1.9.1, 1.10.0rc1, 1.10.0rc2, 1.10.0, 1.10.1rc1, 1.10.1rc2, 1.10.1, 1.11.0rc1, 1.11.0rc2, 1.11.0, 1.11.1rc1, 1.11.1, 1.12.0rc1, 1.12.0, 1.12.1, 1.13.0rc1, 1.13.0rc2, 1.13.0rc3, 1.13.0, 1.14.0rc1, 1.14.0rc2, 1.14.0, 1.14.1, 1.14.2rc1, 1.14.2, 1.15.0rc1, 1.15.0, 1.16.0rc1, 1.16.0, 1.16.1rc1, 1.16.1, 1.17.0rc1, 1.17.0, 1.17.1rc1, 1.17.1, 1.18.0rc1, 1.18.0, 1.19.0rc1, 1.19.0, 1.20.0rc1, 1.20.0rc2, 1.20.0rc3, 1.20.0, 1.20.1, 1.21.0rc1, 1.21.1rc1, 1.21.1, 1.22.0rc1, 1.22.0, 1.22.1, 1.23.0rc1, 1.23.0, 1.23.1, 1.24.0rc1, 1.24.0, 1.24.1, 1.24.3, 1.25.0rc1, 1.25.0, 1.26.0rc1, 1.26.0, 1.27.0rc1, 1.27.1, 1.27.2)
No matching distribution found for grpcio==1.33.2
```
### Anything else we should know about your project / environment?
We have to deploy this to multiple systems. If we don't use `--only-binary`, our deployments have to wait for grpci to compile on every system and it causes deployments to take forever.
|
non_process
|
python wheel missing for grpcio version please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using python grpcio what operating system linux windows and version centos ubuntu what runtime compiler are you using e g python version or version of gcc gcc gcc red hat gcc ubuntu what did you do install grpcio only binary all what did you expect to see should simply install the package using an existing python wheel what did you see instead based on the output below which happens on both centos and ubuntu there isn t a python wheel for anything above grpcio sudo install grpcio only binary all warning running pip install with root privileges is generally not a good idea try install user instead collecting grpcio could not find a version that satisfies the requirement grpcio from versions no matching distribution found for grpcio anything else we should know about your project environment we have to deploy this to multiple systems if we don t use only binary our deployments have to wait for grpci to compile on every system and it causes deployments to take forever
| 0
|
10,793
| 13,609,065,976
|
IssuesEvent
|
2020-09-23 04:10:29
|
googleapis/java-compute
|
https://api.github.com/repos/googleapis/java-compute
|
closed
|
Dependency Dashboard
|
api: compute type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10.1.0
- [ ] <!-- rebase-branch=renovate/com.google.apis-google-api-services-compute-1.x -->deps: update dependency com.google.apis:google-api-services-compute to v1-rev20200811-1.30.10
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-branch=renovate/org.easymock-easymock-4.x -->deps: update dependency org.easymock:easymock to v4
- [ ] <!-- rebase-branch=renovate/org.objenesis-objenesis-3.x -->deps: update dependency org.objenesis:objenesis to v3
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10.1.0
- [ ] <!-- rebase-branch=renovate/com.google.apis-google-api-services-compute-1.x -->deps: update dependency com.google.apis:google-api-services-compute to v1-rev20200811-1.30.10
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-branch=renovate/org.easymock-easymock-4.x -->deps: update dependency org.easymock:easymock to v4
- [ ] <!-- rebase-branch=renovate/org.objenesis-objenesis-3.x -->deps: update dependency org.objenesis:objenesis to v3
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud libraries bom to deps update dependency com google apis google api services compute to deps update dependency com google cloud google cloud storage to deps update dependency org easymock easymock to deps update dependency org objenesis objenesis to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
| 1
|
8,192
| 11,392,260,727
|
IssuesEvent
|
2020-01-30 01:51:39
|
aodn/imos-toolbox
|
https://api.github.com/repos/aodn/imos-toolbox
|
opened
|
Validate DB entries
|
Type:enhancement Unit:Processing
|
A common issue with users is unrecognizable errors in the toolbox when some database entry is wrong.
The pain point here is that the toolbox debugging msg is, most of the time, unclear for those cases.
Moreover, it's daunting to detect this kind of problems in large/lot of csv files. As the DB field size increases, so does the chance of errors and failure because of a single entry.
We should try to validate the entry(ies) at read time, and provide a more useful error message.
|
1.0
|
Validate DB entries - A common issue with users is unrecognizable errors in the toolbox when some database entry is wrong.
The pain point here is that the toolbox debugging msg is, most of the time, unclear for those cases.
Moreover, it's daunting to detect this kind of problems in large/lot of csv files. As the DB field size increases, so does the chance of errors and failure because of a single entry.
We should try to validate the entry(ies) at read time, and provide a more useful error message.
|
process
|
validate db entries a common issue with users is unrecognizable errors in the toolbox when some database entry is wrong the pain point here is that the toolbox debugging msg is most of the time unclear for those cases moreover it s daunting to detect this kind of problems in large lot of csv files as the db field size increases so does the chance of errors and failure because of a single entry we should try to validate the entry ies at read time and provide a more useful error message
| 1
|
45,454
| 5,716,261,609
|
IssuesEvent
|
2017-04-19 14:48:45
|
khartec/waltz
|
https://api.github.com/repos/khartec/waltz
|
closed
|
Fix batch attestation calculation for lineage
|
fixed (test & close)
|
this will now be for flow_diagram and not for lineage
|
1.0
|
Fix batch attestation calculation for lineage - this will now be for flow_diagram and not for lineage
|
non_process
|
fix batch attestation calculation for lineage this will now be for flow diagram and not for lineage
| 0
|
15,650
| 19,846,704,406
|
IssuesEvent
|
2022-01-21 07:30:48
|
ooi-data/RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered
|
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:30:47.907585.
## Details
Flow name: `RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:30:47.907585.
## Details
Flow name: `RS01SBPD-DP01A-01-CTDPFL104-recovered_inst-dpc_ctd_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst dpc ctd instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
194,505
| 22,262,009,611
|
IssuesEvent
|
2022-06-10 01:58:50
|
nanopathi/linux-4.19.72_CVE-2021-3347
|
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-3347
|
reopened
|
WS-2021-0524 (Medium) detected in linuxlinux-4.19.239
|
security vulnerability
|
## WS-2021-0524 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-3347/commit/8b7c061438f230c475fd8cd97a0917f6ebb9fbe0">8b7c061438f230c475fd8cd97a0917f6ebb9fbe0</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel is vulnerable to crash with CONFIG_FORTIRY_SOURCE in arch/arm/mach-socfpga/platsmp.c
<p>Publish Date: 2021-12-01
<p>URL: <a href=https://github.com/gregkh/linux/commit/20a09c8b25a21506cf8b60bdef78f31a6549ebfc>WS-2021-0524</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002532">https://osv.dev/vulnerability/GSD-2021-1002532</a></p>
<p>Release Date: 2021-12-01</p>
<p>Fix Resolution: Linux/Kernel -v4.14.257, v4.19.219, v5.4.163, v5.10.83, v5.15.6, v5.16-rc3
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0524 (Medium) detected in linuxlinux-4.19.239 - ## WS-2021-0524 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-3347/commit/8b7c061438f230c475fd8cd97a0917f6ebb9fbe0">8b7c061438f230c475fd8cd97a0917f6ebb9fbe0</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel is vulnerable to crash with CONFIG_FORTIRY_SOURCE in arch/arm/mach-socfpga/platsmp.c
<p>Publish Date: 2021-12-01
<p>URL: <a href=https://github.com/gregkh/linux/commit/20a09c8b25a21506cf8b60bdef78f31a6549ebfc>WS-2021-0524</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002532">https://osv.dev/vulnerability/GSD-2021-1002532</a></p>
<p>Release Date: 2021-12-01</p>
<p>Fix Resolution: Linux/Kernel -v4.14.257, v4.19.219, v5.4.163, v5.10.83, v5.15.6, v5.16-rc3
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in linuxlinux ws medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details in linux kernel is vulnerable to crash with config fortiry source in arch arm mach socfpga platsmp c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux kernel step up your open source security game with whitesource
| 0
|
600,302
| 18,293,040,110
|
IssuesEvent
|
2021-10-05 17:18:04
|
OpenNebula/one
|
https://api.github.com/repos/OpenNebula/one
|
opened
|
Dependency error for onezone command
|
Category: CLI Type: Bug Status: Accepted Priority: Normal
|
**Description**
When installing the CLI in a node different than the Frontend, `onezone` command will throw an error because of an unmet dependency:
```
$ onezone list
Traceback (most recent call last):
5: from /usr/bin/onezone:39:in `<main>'
4: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
3: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
2: from /usr/lib/one/ruby/cli/one_helper/onezone_helper.rb:19:in `<top (required)>'
1: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
/usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require': cannot load such file -- CommandManager (LoadError)
```
**Expected behavior**
Every command dependency should be correctly installed.
**Details**
- Affected Component: CLI
- Version:5.12 & master
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
1.0
|
Dependency error for onezone command - **Description**
When installing the CLI in a node different than the Frontend, `onezone` command will throw an error because of an unmet dependency:
```
$ onezone list
Traceback (most recent call last):
5: from /usr/bin/onezone:39:in `<main>'
4: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
3: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
2: from /usr/lib/one/ruby/cli/one_helper/onezone_helper.rb:19:in `<top (required)>'
1: from /usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require'
/usr/lib/ruby/2.7.0/rubygems/core_ext/kernel_require.rb:92:in `require': cannot load such file -- CommandManager (LoadError)
```
**Expected behavior**
Every command dependency should be correctly installed.
**Details**
- Affected Component: CLI
- Version:5.12 & master
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
non_process
|
dependency error for onezone command description when installing the cli in a node different than the frontend onezone command will throw an error because of an unmet dependency onezone list traceback most recent call last from usr bin onezone in from usr lib ruby rubygems core ext kernel require rb in require from usr lib ruby rubygems core ext kernel require rb in require from usr lib one ruby cli one helper onezone helper rb in from usr lib ruby rubygems core ext kernel require rb in require usr lib ruby rubygems core ext kernel require rb in require cannot load such file commandmanager loaderror expected behavior every command dependency should be correctly installed details affected component cli version master progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
| 0
|
10,115
| 13,044,162,218
|
IssuesEvent
|
2020-07-29 03:47:30
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `WeekWithoutMode` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `WeekWithoutMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `WeekWithoutMode` from TiDB -
## Description
Port the scalar function `WeekWithoutMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function weekwithoutmode from tidb description port the scalar function weekwithoutmode from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
181,326
| 30,668,353,371
|
IssuesEvent
|
2023-07-25 20:09:14
|
readthedocs/readthedocs.org
|
https://api.github.com/repos/readthedocs/readthedocs.org
|
closed
|
Support searching with facets.
|
Feature Needed: design decision
|
Search facets will be useful in many ways.
They can be used to target the search inside a subproject.
For example: Celery has a subproject Kombu, so we can search in Celery docs, we can search inside Kombu docs, like: `subproject: kombu serializers`
|
1.0
|
Support searching with facets. - Search facets will be useful in many ways.
They can be used to target the search inside a subproject.
For example: Celery has a subproject Kombu, so we can search in Celery docs, we can search inside Kombu docs, like: `subproject: kombu serializers`
|
non_process
|
support searching with facets search facets will be useful in many ways they can be used to target the search inside a subproject for example celery has a subproject kombu so we can search in celery docs we can search inside kombu docs like subproject kombu serializers
| 0
|
9,562
| 12,518,751,418
|
IssuesEvent
|
2020-06-03 13:26:48
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Reevaluate ASP.NET Core analyzer value
|
type: process
|
The analyzer complicates project generation - we should check that it still adds value. (It may well be that as the installation approach has changed, many of the things it guards against are irrelevant.)
Even if we keep it, we could potentially not generate the project file - or just generate certain parts of it (e.g. version and copyright), making the rest manual.
|
1.0
|
Reevaluate ASP.NET Core analyzer value - The analyzer complicates project generation - we should check that it still adds value. (It may well be that as the installation approach has changed, many of the things it guards against are irrelevant.)
Even if we keep it, we could potentially not generate the project file - or just generate certain parts of it (e.g. version and copyright), making the rest manual.
|
process
|
reevaluate asp net core analyzer value the analyzer complicates project generation we should check that it still adds value it may well be that as the installation approach has changed many of the things it guards against are irrelevant even if we keep it we could potentially not generate the project file or just generate certain parts of it e g version and copyright making the rest manual
| 1
|
8,299
| 8,222,993,130
|
IssuesEvent
|
2018-09-06 09:14:28
|
a2develop/bugTracker
|
https://api.github.com/repos/a2develop/bugTracker
|
closed
|
Свойства ФОПа. Отчество -- неточная подсказка.
|
c_beta c_service done! misspelling
|
Для поля "По батькові" подсказка "Введить прізвище та ініціали"
Наверное, нужна типа "Введить по батькові"

|
1.0
|
Свойства ФОПа. Отчество -- неточная подсказка. - Для поля "По батькові" подсказка "Введить прізвище та ініціали"
Наверное, нужна типа "Введить по батькові"

|
non_process
|
свойства фопа отчество неточная подсказка для поля по батькові подсказка введить прізвище та ініціали наверное нужна типа введить по батькові
| 0
|
10,326
| 13,162,135,300
|
IssuesEvent
|
2020-08-10 20:57:21
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
vsphere post processor lacks roles documentation
|
docs help wanted post-processor/vsphere
|
The only way I can get the vSphere post-processor to work, is when I add the configured user to the administrators group.
This will cause lots of red flags in a security audit. Please include in the documentation which vSphere roles by Packer are required so we can fine-tune access.
I tried the recommendations here: https://github.com/TotalJobsGroup/Packer/blob/master/README.md
But it seems this is not complete, as I still get permissions in Packer when I only enable those roles.
|
1.0
|
vsphere post processor lacks roles documentation - The only way I can get the vSphere post-processor to work, is when I add the configured user to the administrators group.
This will cause lots of red flags in a security audit. Please include in the documentation which vSphere roles by Packer are required so we can fine-tune access.
I tried the recommendations here: https://github.com/TotalJobsGroup/Packer/blob/master/README.md
But it seems this is not complete, as I still get permissions in Packer when I only enable those roles.
|
process
|
vsphere post processor lacks roles documentation the only way i can get the vsphere post processor to work is when i add the configured user to the administrators group this will cause lots of red flags in a security audit please include in the documentation which vsphere roles by packer are required so we can fine tune access i tried the recommendations here but it seems this is not complete as i still get permissions in packer when i only enable those roles
| 1
|
121,881
| 26,046,645,130
|
IssuesEvent
|
2022-12-22 14:56:08
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Language String Problem with Menu Heading Menu Type
|
No Code Attached Yet
|
### Steps to reproduce the issue
Joomla 4.2.6 create a new menu item select System Links >> Menu Heading type the language string of Options Tab does not display properly it displays constant as "COM_MENUS_PAGE-OPTIONS_FIELDSET_LABEL" see attached image. I have searched for language string whole package with text-search tool and found nothing.
image
https://prnt.sc/snrVNXAs7bCG
### Expected result
It should be Options
### Actual result
### System information (as much as possible)
### Additional comments
|
1.0
|
Language String Problem with Menu Heading Menu Type - ### Steps to reproduce the issue
Joomla 4.2.6 create a new menu item select System Links >> Menu Heading type the language string of Options Tab does not display properly it displays constant as "COM_MENUS_PAGE-OPTIONS_FIELDSET_LABEL" see attached image. I have searched for language string whole package with text-search tool and found nothing.
image
https://prnt.sc/snrVNXAs7bCG
### Expected result
It should be Options
### Actual result
### System information (as much as possible)
### Additional comments
|
non_process
|
language string problem with menu heading menu type steps to reproduce the issue joomla create a new menu item select system links menu heading type the language string of options tab does not display properly it displays constant as com menus page options fieldset label see attached image i have searched for language string whole package with text search tool and found nothing image expected result it should be options actual result system information as much as possible additional comments
| 0
|
2,524
| 5,288,207,339
|
IssuesEvent
|
2017-02-08 14:35:52
|
openvstorage/volumedriver
|
https://api.github.com/repos/openvstorage/volumedriver
|
closed
|
Catch error in case there is no RDMA device and show meaningful error
|
process_wontfix type_enhancement
|
```
2016-12-17 08:13:18 931793 -0800 - ubuntu-16 - 24199/0x00007fcd2ef50740 - dtl_server//usr/bin/failovercache - 0000000000000002 - info - run: starting, server path: "/mnt/hdd4/pstorage_write_dtl_1", address to bind to: 125.212.218.15, port: 26217, transport type: RSocket, busy-loop usecs: 0
[...]
dtl_server//usr/bin/failovercache - 0000000000000005 - fatal - operator(): caught exception: socket No such device (19)
```
DTL is configured to use RSocket, but it looks like there's no RDMA device / the wrong one is specified.
|
1.0
|
Catch error in case there is no RDMA device and show meaningful error - ```
2016-12-17 08:13:18 931793 -0800 - ubuntu-16 - 24199/0x00007fcd2ef50740 - dtl_server//usr/bin/failovercache - 0000000000000002 - info - run: starting, server path: "/mnt/hdd4/pstorage_write_dtl_1", address to bind to: 125.212.218.15, port: 26217, transport type: RSocket, busy-loop usecs: 0
[...]
dtl_server//usr/bin/failovercache - 0000000000000005 - fatal - operator(): caught exception: socket No such device (19)
```
DTL is configured to use RSocket, but it looks like there's no RDMA device / the wrong one is specified.
|
process
|
catch error in case there is no rdma device and show meaningful error ubuntu dtl server usr bin failovercache info run starting server path mnt pstorage write dtl address to bind to port transport type rsocket busy loop usecs dtl server usr bin failovercache fatal operator caught exception socket no such device dtl is configured to use rsocket but it looks like there s no rdma device the wrong one is specified
| 1
|
356,944
| 25,176,294,191
|
IssuesEvent
|
2022-11-11 09:33:25
|
ongwenpin/pe
|
https://api.github.com/repos/ongwenpin/pe
|
opened
|
Extra "G" in sequence diagram of DG
|
severity.VeryLow type.DocumentationBug
|

On page 30 of the DG, the sequence diagram has extra letter "G" which may lead to confusion for readers.
<!--session: 1668153030050-2e911811-a668-4ac5-b895-28e31f18355d-->
<!--Version: Web v3.4.4-->
|
1.0
|
Extra "G" in sequence diagram of DG - 
On page 30 of the DG, the sequence diagram has extra letter "G" which may lead to confusion for readers.
<!--session: 1668153030050-2e911811-a668-4ac5-b895-28e31f18355d-->
<!--Version: Web v3.4.4-->
|
non_process
|
extra g in sequence diagram of dg on page of the dg the sequence diagram has extra letter g which may lead to confusion for readers
| 0
|
15,807
| 20,008,700,617
|
IssuesEvent
|
2022-02-01 01:48:26
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Resources should be processed in Web Workers
|
TYPE: bug AREA: client SYSTEM: resource processing health-monitor SYSTEM: workers STATE: Stale
|
The Service Workers feature is usually used for resource-caching. Currently, when the user's service worker uses [Caches](https://w3c.github.io/ServiceWorker/#cache-objects) objects, all resources are cached unprocessed. So, in the Web Workers we should process resources before they will be placed into the cache.
Example in which the service worker puts a script into the cache after the installation is finished:
```js
require('http')
.createServer((req, res) => {
if (req.url === '/') {
res.writeHead(200, { 'content-type': 'text/html' });
res.end(`
<!DOCTYPE html>
<head>
<link rel="shortcut icon" href="#" />
</head>
<body>
<script>
navigator.serviceWorker.register('sw.js');
</script>
</body>
`);
} else if (req.url === '/sw.js') {
res.writeHead(200, { 'content-type': 'application/javascript' });
res.end(`
self.addEventListener('install', function (event) {
const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js';
event.waitUntil(
caches.open('v1').then(function (cache) {
return cache.add(url);
})
);
});
`);
} else
res.destroy();
})
.listen(2020, () => console.log('http://localhost:2020'));
```
also, `fetch` could be used:
```js
self.addEventListener('install', function (event) {
const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js';
event.waitUntil(
caches.open('v1').then(function (cache) {
return fetch(url).then(function(response) {
if (!response.ok)
throw new TypeError('bad response status');
return cache.put(url, response);
})
})
);
});
```
or the fetch event handler could respond with the Response object, containing a script:
```js
self.addEventListener('fetch', function(event) {
if(event.request.url.match(/some\-script\.js/)) {
event.respondWith(
new Response(`console.log('mocked some-script.js');`, {
headers: {'Content-Type': 'text/javascript'}
})
);
}
});
```
This bug was found during analysis of the Health-Monitor run issues on the twitter.com website.
|
1.0
|
Resources should be processed in Web Workers - The Service Workers feature is usually used for resource-caching. Currently, when the user's service worker uses [Caches](https://w3c.github.io/ServiceWorker/#cache-objects) objects, all resources are cached unprocessed. So, in the Web Workers we should process resources before they will be placed into the cache.
Example in which the service worker puts a script into the cache after the installation is finished:
```js
require('http')
.createServer((req, res) => {
if (req.url === '/') {
res.writeHead(200, { 'content-type': 'text/html' });
res.end(`
<!DOCTYPE html>
<head>
<link rel="shortcut icon" href="#" />
</head>
<body>
<script>
navigator.serviceWorker.register('sw.js');
</script>
</body>
`);
} else if (req.url === '/sw.js') {
res.writeHead(200, { 'content-type': 'application/javascript' });
res.end(`
self.addEventListener('install', function (event) {
const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js';
event.waitUntil(
caches.open('v1').then(function (cache) {
return cache.add(url);
})
);
});
`);
} else
res.destroy();
})
.listen(2020, () => console.log('http://localhost:2020'));
```
also, `fetch` could be used:
```js
self.addEventListener('install', function (event) {
const url = 'https://wentwrong.github.io/service-workers-test/twitter-issue/some-script.js';
event.waitUntil(
caches.open('v1').then(function (cache) {
return fetch(url).then(function(response) {
if (!response.ok)
throw new TypeError('bad response status');
return cache.put(url, response);
})
})
);
});
```
or the fetch event handler could respond with the Response object, containing a script:
```js
self.addEventListener('fetch', function(event) {
if(event.request.url.match(/some\-script\.js/)) {
event.respondWith(
new Response(`console.log('mocked some-script.js');`, {
headers: {'Content-Type': 'text/javascript'}
})
);
}
});
```
This bug was found during analysis of the Health-Monitor run issues on the twitter.com website.
|
process
|
resources should be processed in web workers the service workers feature is usually used for resource caching currently when the user s service worker uses objects all resources are cached unprocessed so in the web workers we should process resources before they will be placed into the cache example in which the service worker puts a script into the cache after the installation is finished js require http createserver req res if req url res writehead content type text html res end navigator serviceworker register sw js else if req url sw js res writehead content type application javascript res end self addeventlistener install function event const url event waituntil caches open then function cache return cache add url else res destroy listen console log also fetch could be used js self addeventlistener install function event const url event waituntil caches open then function cache return fetch url then function response if response ok throw new typeerror bad response status return cache put url response or the fetch event handler could respond with the response object containing a script js self addeventlistener fetch function event if event request url match some script js event respondwith new response console log mocked some script js headers content type text javascript this bug was found during analysis of the health monitor run issues on the twitter com website
| 1
|
675
| 3,145,705,071
|
IssuesEvent
|
2015-09-14 19:17:03
|
K0zka/kerub
|
https://api.github.com/repos/K0zka/kerub
|
opened
|
Add exception mapping for Jackson object mapping exceptions
|
component:data processing enhancement priority: normal
|
A partial vm json for example should not create an uggly 500, but some meaningful error.
|
1.0
|
Add exception mapping for Jackson object mapping exceptions - A partial vm json for example should not create an uggly 500, but some meaningful error.
|
process
|
add exception mapping for jackson object mapping exceptions a partial vm json for example should not create an uggly but some meaningful error
| 1
|
2,566
| 5,316,106,015
|
IssuesEvent
|
2017-02-13 19:02:06
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[subtitles] [fr] MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI »
|
Language: French Process: [7] Confirmation sent to Antoine
|
# Video title
MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI »
# URL
https://youtu.be/EDC1-D4qYV8
# Youtube subtitles language
Langue des sous-titres : Français
# Duration
7'39
# Subtitles URL
https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&bl=watch&ref=wt&ui=hd&v=EDC1-D4qYV8&lang=fr&tab=captions
|
1.0
|
[subtitles] [fr] MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI » - # Video title
MÉLENCHON : « MON BULLETIN DE VOTE EST UN COUP DE BALAI »
# URL
https://youtu.be/EDC1-D4qYV8
# Youtube subtitles language
Langue des sous-titres : Français
# Duration
7'39
# Subtitles URL
https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&bl=watch&ref=wt&ui=hd&v=EDC1-D4qYV8&lang=fr&tab=captions
|
process
|
mélenchon « mon bulletin de vote est un coup de balai » video title mélenchon « mon bulletin de vote est un coup de balai » url youtube subtitles language langue des sous titres français duration subtitles url
| 1
|
18,164
| 24,201,554,491
|
IssuesEvent
|
2022-09-24 16:41:22
|
fertadeo/ISPC-2do-Cuat-Proyecto
|
https://api.github.com/repos/fertadeo/ISPC-2do-Cuat-Proyecto
|
closed
|
FrontEnd - #US02 - Crear archivos iniciales del proyecto
|
in process
|
Crear archivo
index.html
style.css
main.js
|
1.0
|
FrontEnd - #US02 - Crear archivos iniciales del proyecto - Crear archivo
index.html
style.css
main.js
|
process
|
frontend crear archivos iniciales del proyecto crear archivo index html style css main js
| 1
|
17,287
| 23,095,899,302
|
IssuesEvent
|
2022-07-26 19:30:48
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
opened
|
Traduzir/remover códigos remanescentes de módulos que utilizem o Pyppeteer
|
[0] Desenvolvimento [2] Média Prioridade [1] Aprimoramento [3] Processamento Dinâmico
|
## Comportamento Esperado
Com a introdução do Playwright no sistema, qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca. Códigos específicos do Pyppeteer que não estejam em uso devem ser removidos, e códigos que estejam em uso devem ser corretamente traduzidos.
## Comportamento Atual
Algumas partes do código ainda utilizam a biblioteca Pyppeteer para realização de interações dinâmicas, entre elas:
- Módulo `antiblock_drivers` (menciona o Pyppeteer no README, mas não utiliza internamente)
- View `load_iframe` (utiliza o módulo `scrapy_puppeteer`, que deve ser removido)
- Módulo `step-by-step` (menciona o Pyppeteer no código e utiliza em alguns exemplos)
- Módulo `entry_probing`
- Módulo `scrapy_puppeteer` (obsoleto e deve ser removido)
|
1.0
|
Traduzir/remover códigos remanescentes de módulos que utilizem o Pyppeteer - ## Comportamento Esperado
Com a introdução do Playwright no sistema, qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca. Códigos específicos do Pyppeteer que não estejam em uso devem ser removidos, e códigos que estejam em uso devem ser corretamente traduzidos.
## Comportamento Atual
Algumas partes do código ainda utilizam a biblioteca Pyppeteer para realização de interações dinâmicas, entre elas:
- Módulo `antiblock_drivers` (menciona o Pyppeteer no README, mas não utiliza internamente)
- View `load_iframe` (utiliza o módulo `scrapy_puppeteer`, que deve ser removido)
- Módulo `step-by-step` (menciona o Pyppeteer no código e utiliza em alguns exemplos)
- Módulo `entry_probing`
- Módulo `scrapy_puppeteer` (obsoleto e deve ser removido)
|
process
|
traduzir remover códigos remanescentes de módulos que utilizem o pyppeteer comportamento esperado com a introdução do playwright no sistema qualquer tipo de processamento dinâmico deve ser feito utilizando essa biblioteca códigos específicos do pyppeteer que não estejam em uso devem ser removidos e códigos que estejam em uso devem ser corretamente traduzidos comportamento atual algumas partes do código ainda utilizam a biblioteca pyppeteer para realização de interações dinâmicas entre elas módulo antiblock drivers menciona o pyppeteer no readme mas não utiliza internamente view load iframe utiliza o módulo scrapy puppeteer que deve ser removido módulo step by step menciona o pyppeteer no código e utiliza em alguns exemplos módulo entry probing módulo scrapy puppeteer obsoleto e deve ser removido
| 1
|
276,847
| 30,554,827,461
|
IssuesEvent
|
2023-07-20 10:57:30
|
CliffCrerar/bohemian-rhapsody
|
https://api.github.com/repos/CliffCrerar/bohemian-rhapsody
|
opened
|
CVE-2023-26136 (Medium) detected in tough-cookie-2.5.0.tgz, tough-cookie-3.0.1.tgz
|
Mend: dependency security vulnerability
|
## CVE-2023-26136 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tough-cookie-2.5.0.tgz</b>, <b>tough-cookie-3.0.1.tgz</b></p></summary>
<p>
<details><summary><b>tough-cookie-2.5.0.tgz</b></p></summary>
<p>RFC6265 Cookies and Cookie Jar for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/tough-cookie/package.json</p>
<p>
Dependency Hierarchy:
- ng-deploy-0.2.3.tgz (Root Library)
- arm-resources-1.1.0.tgz
- ms-rest-js-1.8.14.tgz
- :x: **tough-cookie-2.5.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>tough-cookie-3.0.1.tgz</b></p></summary>
<p>RFC6265 Cookies and Cookie Jar for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@azure/storage-blob/node_modules/tough-cookie/package.json</p>
<p>
Dependency Hierarchy:
- ng-deploy-0.2.3.tgz (Root Library)
- storage-blob-10.5.0.tgz
- ms-rest-js-2.0.5.tgz
- :x: **tough-cookie-3.0.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bohemian-rhapsody/commit/ecce634609a1383c732a17c20e59f30a68158b81">ecce634609a1383c732a17c20e59f30a68158b81</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package tough-cookie before 4.1.3 are vulnerable to Prototype Pollution due to improper handling of Cookies when using CookieJar in rejectPublicSuffixes=false mode. This issue arises from the manner in which the objects are initialized.
<p>Publish Date: 2023-07-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26136>CVE-2023-26136</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-26136">https://www.cve.org/CVERecord?id=CVE-2023-26136</a></p>
<p>Release Date: 2023-07-01</p>
<p>Fix Resolution: tough-cookie - 4.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-26136 (Medium) detected in tough-cookie-2.5.0.tgz, tough-cookie-3.0.1.tgz - ## CVE-2023-26136 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tough-cookie-2.5.0.tgz</b>, <b>tough-cookie-3.0.1.tgz</b></p></summary>
<p>
<details><summary><b>tough-cookie-2.5.0.tgz</b></p></summary>
<p>RFC6265 Cookies and Cookie Jar for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.5.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/tough-cookie/package.json</p>
<p>
Dependency Hierarchy:
- ng-deploy-0.2.3.tgz (Root Library)
- arm-resources-1.1.0.tgz
- ms-rest-js-1.8.14.tgz
- :x: **tough-cookie-2.5.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>tough-cookie-3.0.1.tgz</b></p></summary>
<p>RFC6265 Cookies and Cookie Jar for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-3.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@azure/storage-blob/node_modules/tough-cookie/package.json</p>
<p>
Dependency Hierarchy:
- ng-deploy-0.2.3.tgz (Root Library)
- storage-blob-10.5.0.tgz
- ms-rest-js-2.0.5.tgz
- :x: **tough-cookie-3.0.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/bohemian-rhapsody/commit/ecce634609a1383c732a17c20e59f30a68158b81">ecce634609a1383c732a17c20e59f30a68158b81</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package tough-cookie before 4.1.3 are vulnerable to Prototype Pollution due to improper handling of Cookies when using CookieJar in rejectPublicSuffixes=false mode. This issue arises from the manner in which the objects are initialized.
<p>Publish Date: 2023-07-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26136>CVE-2023-26136</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-26136">https://www.cve.org/CVERecord?id=CVE-2023-26136</a></p>
<p>Release Date: 2023-07-01</p>
<p>Fix Resolution: tough-cookie - 4.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in tough cookie tgz tough cookie tgz cve medium severity vulnerability vulnerable libraries tough cookie tgz tough cookie tgz tough cookie tgz cookies and cookie jar for node js library home page a href path to dependency file package json path to vulnerable library node modules tough cookie package json dependency hierarchy ng deploy tgz root library arm resources tgz ms rest js tgz x tough cookie tgz vulnerable library tough cookie tgz cookies and cookie jar for node js library home page a href path to dependency file package json path to vulnerable library node modules azure storage blob node modules tough cookie package json dependency hierarchy ng deploy tgz root library storage blob tgz ms rest js tgz x tough cookie tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package tough cookie before are vulnerable to prototype pollution due to improper handling of cookies when using cookiejar in rejectpublicsuffixes false mode this issue arises from the manner in which the objects are initialized publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tough cookie step up your open source security game with mend
| 0
|
11,210
| 13,959,170,566
|
IssuesEvent
|
2020-10-24 15:18:21
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
QA ISO images
|
0.kind: bug 6.topic: GNOME 6.topic: nixos 6.topic: qt/kde 6.topic: release process
|
**Describe the bug**
Some amount of manual QA is needed to verify these images are in good shape. In 20.03 we had a pretty silly bug slip through and got dragged on HN :rofl:
Even without that as encouragement this is a good idea.
The main goal is to determine the following:
> It needs to be verified that all ISO's that are using a release blocking desktop boot cleanly in their supported configurations. They must boot live when written to a USB stick with **any** of our supported methods documented in the NixOS manual.
I've drafted some criteria to look out for when testing a desktop. The version for this exact issue can be seen at this [gist](https://gist.github.com/worldofpeace/019e0b77020affbe3c3623d807f73f22).
After having testing please comment with a checklist on the items like
```markdown
- [x] Shutdown, reboot, login, logout
...
```
The following images need testing:
- [x] Plasma5 ISO
- [x] GNOME ISO
Here are links to the latest images:
https://channels.nixos.org/nixos-20.09/latest-nixos-plasma5-x86_64-linux.iso
~https://channels.nixos.org/nixos-20.09/latest-nixos-gnome-x86_64-linux.iso~
The gnome iso isn't added to the channel atm:
You can get it from an eval of it on hydra: https://hydra.nixos.org/job/nixos/release-20.09/nixos.iso_gnome.x86_64-linux
picking a build https://hydra.nixos.org/build/128080348
and selecting `Link` from `Build Products`.
# Remaining Q.A tasks
## Plasma
- [x] Printing https://github.com/NixOS/nixpkgs/issues/98536
- [ ] Keyboard layout configuration
- [ ] unlocking encrypted storage volumes
- [x] console - this is tested automatically
- [ ] When logging in via the default login manager for a release-blocking desktop
- [ ] After logging in to a release-blocking desktop, if the user account does not have its own keyboard layout configuration for that desktop (if there is such a user/desktop-specific configuration, it must be used when that user logs in to that desktop)
- [ ] Desktop Keyring
- [ ] User Switching?
## GNOME
- [x] configured keyboard layout (using nixos options) is used when unlocking encrypted storage volumes
- [x] Desktop Keyring
|
1.0
|
QA ISO images - **Describe the bug**
Some amount of manual QA is needed to verify these images are in good shape. In 20.03 we had a pretty silly bug slip through and got dragged on HN :rofl:
Even without that as encouragement this is a good idea.
The main goal is to determine the following:
> It needs to be verified that all ISO's that are using a release blocking desktop boot cleanly in their supported configurations. They must boot live when written to a USB stick with **any** of our supported methods documented in the NixOS manual.
I've drafted some criteria to look out for when testing a desktop. The version for this exact issue can be seen at this [gist](https://gist.github.com/worldofpeace/019e0b77020affbe3c3623d807f73f22).
After having testing please comment with a checklist on the items like
```markdown
- [x] Shutdown, reboot, login, logout
...
```
The following images need testing:
- [x] Plasma5 ISO
- [x] GNOME ISO
Here are links to the latest images:
https://channels.nixos.org/nixos-20.09/latest-nixos-plasma5-x86_64-linux.iso
~https://channels.nixos.org/nixos-20.09/latest-nixos-gnome-x86_64-linux.iso~
The gnome iso isn't added to the channel atm:
You can get it from an eval of it on hydra: https://hydra.nixos.org/job/nixos/release-20.09/nixos.iso_gnome.x86_64-linux
picking a build https://hydra.nixos.org/build/128080348
and selecting `Link` from `Build Products`.
# Remaining Q.A tasks
## Plasma
- [x] Printing https://github.com/NixOS/nixpkgs/issues/98536
- [ ] Keyboard layout configuration
- [ ] unlocking encrypted storage volumes
- [x] console - this is tested automatically
- [ ] When logging in via the default login manager for a release-blocking desktop
- [ ] After logging in to a release-blocking desktop, if the user account does not have its own keyboard layout configuration for that desktop (if there is such a user/desktop-specific configuration, it must be used when that user logs in to that desktop)
- [ ] Desktop Keyring
- [ ] User Switching?
## GNOME
- [x] configured keyboard layout (using nixos options) is used when unlocking encrypted storage volumes
- [x] Desktop Keyring
|
process
|
qa iso images describe the bug some amount of manual qa is needed to verify these images are in good shape in we had a pretty silly bug slip through and got dragged on hn rofl even without that as encouragement this is a good idea the main goal is to determine the following it needs to be verified that all iso s that are using a release blocking desktop boot cleanly in their supported configurations they must boot live when written to a usb stick with any of our supported methods documented in the nixos manual i ve drafted some criteria to look out for when testing a desktop the version for this exact issue can be seen at this after having testing please comment with a checklist on the items like markdown shutdown reboot login logout the following images need testing iso gnome iso here are links to the latest images the gnome iso isn t added to the channel atm you can get it from an eval of it on hydra picking a build and selecting link from build products remaining q a tasks plasma printing keyboard layout configuration unlocking encrypted storage volumes console this is tested automatically when logging in via the default login manager for a release blocking desktop after logging in to a release blocking desktop if the user account does not have its own keyboard layout configuration for that desktop if there is such a user desktop specific configuration it must be used when that user logs in to that desktop desktop keyring user switching gnome configured keyboard layout using nixos options is used when unlocking encrypted storage volumes desktop keyring
| 1
|
43,897
| 2,893,895,318
|
IssuesEvent
|
2015-06-15 20:23:15
|
Naoghuman/Dream-Better-Worlds
|
https://api.github.com/repos/Naoghuman/Dream-Better-Worlds
|
closed
|
Refactoring logger level in all onActionXy methods from info to debug.
|
development refactoring issue fixed priority low
|
- Refactoring logger level in all onActionXy methods from info to debug.
- (v) Add list from all classes which have onActionXy methods.
|
1.0
|
Refactoring logger level in all onActionXy methods from info to debug. - - Refactoring logger level in all onActionXy methods from info to debug.
- (v) Add list from all classes which have onActionXy methods.
|
non_process
|
refactoring logger level in all onactionxy methods from info to debug refactoring logger level in all onactionxy methods from info to debug v add list from all classes which have onactionxy methods
| 0
|
16,873
| 22,152,749,303
|
IssuesEvent
|
2022-06-03 18:44:53
|
GoogleCloudPlatform/cloud-ops-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
|
closed
|
Brainstorm ways to reduce Cloud Shell dependency
|
type: process priority: p3
|
Backend changes can break sandbox in ways we can't control. We should take some time to brainstorm ways we can prevent these issues in the future.
Related to https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/issues/754, where we discuss moving sandboxctl into the Cloud Shell base image, rather than maintaining a custom image. But even using the base image, we may encounter issues related to Cloud Shell. We should consider more substantial re-designs as well
|
1.0
|
Brainstorm ways to reduce Cloud Shell dependency - Backend changes can break sandbox in ways we can't control. We should take some time to brainstorm ways we can prevent these issues in the future.
Related to https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/issues/754, where we discuss moving sandboxctl into the Cloud Shell base image, rather than maintaining a custom image. But even using the base image, we may encounter issues related to Cloud Shell. We should consider more substantial re-designs as well
|
process
|
brainstorm ways to reduce cloud shell dependency backend changes can break sandbox in ways we can t control we should take some time to brainstorm ways we can prevent these issues in the future related to where we discuss moving sandboxctl into the cloud shell base image rather than maintaining a custom image but even using the base image we may encounter issues related to cloud shell we should consider more substantial re designs as well
| 1
|
7,624
| 10,729,241,390
|
IssuesEvent
|
2019-10-28 15:16:37
|
usgs/libcomcat
|
https://api.github.com/repos/usgs/libcomcat
|
closed
|
Implement nightly builds
|
process
|
Choose a method for building nightly (i.e. Jenkins, Azure and cron job, gitlab runners, etc.).
|
1.0
|
Implement nightly builds - Choose a method for building nightly (i.e. Jenkins, Azure and cron job, gitlab runners, etc.).
|
process
|
implement nightly builds choose a method for building nightly i e jenkins azure and cron job gitlab runners etc
| 1
|
56,801
| 15,373,966,970
|
IssuesEvent
|
2021-03-02 13:16:27
|
Questie/Questie
|
https://api.github.com/repos/Questie/Questie
|
closed
|
Questie tracker in invalid location, resetting....
|
Type - Defect
|
<!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
My questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly. When I open the game it says that is is resetting but never does. When I manually reset it to the center of the screen I can see it (though it is not populated) but when I go to move it anywhere else on the screen it disappears immediately
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->v6.2.5
|
1.0
|
Questie tracker in invalid location, resetting.... - <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
My questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly. When I open the game it says that is is resetting but never does. When I manually reset it to the center of the screen I can see it (though it is not populated) but when I go to move it anywhere else on the screen it disappears immediately
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->v6.2.5
|
non_process
|
questie tracker in invalid location resetting read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description my questie minimap interface is working fine but the tracker with the list of active quests is not functioning properly when i open the game it says that is is resetting but never does when i manually reset it to the center of the screen i can see it though it is not populated but when i go to move it anywhere else on the screen it disappears immediately screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version
| 0
|
432,449
| 12,493,516,066
|
IssuesEvent
|
2020-06-01 09:25:10
|
mesg-foundation/aragon
|
https://api.github.com/repos/mesg-foundation/aragon
|
closed
|
Delete connection
|
high priority
|
1. Aragon app mark as removed the connection on the aragon smart contract
2. A process listen for this removed event from the smart contract and trigger the service process-deployer to delete the process
3. the process-deployer service using the processUrl, download it and calculate its hash (may need to use calculate a hash endpoint). with the hash, it can remove the process from firebase using the delete process endpoint (or something like this).
|
1.0
|
Delete connection - 1. Aragon app mark as removed the connection on the aragon smart contract
2. A process listen for this removed event from the smart contract and trigger the service process-deployer to delete the process
3. the process-deployer service using the processUrl, download it and calculate its hash (may need to use calculate a hash endpoint). with the hash, it can remove the process from firebase using the delete process endpoint (or something like this).
|
non_process
|
delete connection aragon app mark as removed the connection on the aragon smart contract a process listen for this removed event from the smart contract and trigger the service process deployer to delete the process the process deployer service using the processurl download it and calculate its hash may need to use calculate a hash endpoint with the hash it can remove the process from firebase using the delete process endpoint or something like this
| 0
|
219,829
| 17,114,104,913
|
IssuesEvent
|
2021-07-11 00:42:29
|
backend-br/vagas
|
https://api.github.com/repos/backend-br/vagas
|
closed
|
[Remoto] Java Backend Engineer Specialist na AgileProcess
|
AWS CI Docker Git Java MySQL Remoto SQL Scrum Stale Testes Unitários
|
## Descrição da vaga
Buscamos um(a) **Backend Engineer Specialist** que será responsável, junto ao nossos squads de desenvolvimento, por prover a melhor experiência para nossos clientes através de nossas soluções.
RESPONSABILIDADES E ATRIBUIÇÕES
\- Desafiar o status quo e desenvolver soluções inovadoras para problemas complexos;
\- Desenvolver e manter nossos Microserviços de forma ágil, aplicando boas práticas de Engenharia de Software;
\- Contribuir com o desenvolvimento e arquitetura da plataforma, preparando-a para um crescimento acelerado;
\- Construir uma base sólida para o desenvolvimento de novos produtos;
\- Desenvolver sistemas escaláveis, sustentáveis e orientados ao usuário;
\- Trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência;
\- Ajudar o crescimento do time de tecnologia e engenharia.
REQUISITOS E QUALIFICAÇÕES
\- Experiência e conhecimento profundo com desenvolvimento Java 8;
\- Experiência e conhecimento em GitFlow;
\- Experiência e conhecimento profundo em Docker;
\- Ter atuado na construção de testes unitários e integrados;
\- Ter atuado na construção de testes de comportamento (BDD);
\- Conhecimentos em Design Patterns, arquitetura e engenharia de software;
\- Conhecimentos em GitLab CI;
\- Conhecimentos em metodologias ágeis de desenvolvimento (Scrum, Kanban).
INFORMAÇÕES ADICIONAIS
**Nosso Stack:**
\- Java;
\- MySQL;
\- AWS;
\- Git (GitLab).
**DNA Agile <3**
Faz parte do DNA de todo **AGILE** ser ambicioso e buscar entregar sempre mais e melhor. O respeito e a tolerância são indispensáveis para nós, e somos apaixonados pela nossa diversidade. Crescemos como time e celebrando juntos todos os resultados atingidos. Somos Customer Centric e buscamos criar conexão com nossos clientes, com simplicidade e empatia. Compartilhamos de forma direta e honesta todas as informações e opiniões possíveis. Reconhecemos nossas fraquezas e assumimos nossos erros. Somos apaixonados por inovação e usamos da nossa criatividade, autonomia e voz para propor novas formas de fazer as coisas.
Somos representados por cada um e nadamos todos na mesma direção.
**O que você vai encontrar:**
* Onboarding de boas-vindas!
* “All Hands”: nosso encontro semanal com o CEO;
* Dress Code: seja você mesmo(a);
* Flexibilidade de horário;
* VR/VA Flex: R$ 550,00 (mês);
* Plano de saúde (para você e quem você ama);
* Plano odontológico;
* TotalPass;
* Clube de Descontos - NewValue;
* PLR;
* Parceria com ZenKlub;
* e muito mais!
**Redes Sociais**: Venha conferir nossas redes sociais e revolucionar a logística com a gente!
* Instagram: @agileprocess
* LinkedIn: /agileprocess
* Twitter: @agileprocess
_\*Todas as vagas da AgileProcess são abertas para pessoas com deficiência (PcD)._
_\*Nos processos seletivos da AgileProcess, não há qualquer tipo de distinção em virtude de gênero, orientação sexual, religião, crença, condição social ou raça._
## Salário
R$12500 - R$14000
## Local
REMOTO
## Como se candidatar
Accesse a [Vaga de Java Backend Engineer Specialist](https://vaga.contratar.dev/apply/full/06686ffa-44e4-4505-a874-f719b0532a21) para se candidatar.
## Labels
* Remoto
* SQL
* Scrum
* Redes
* AWS
* Docker
* Java
* Git
## Sobre a contratar.dev
A contratar.dev é uma plataforma de recrutamento e seleção automatizada de Desenvolvedores. Ao se candidatar a uma vaga da contratar.dev, você estará participando automaticamente de dezenas de processos seletivos para vagas similares, aumentando em MUITO as suas chances de mudar de vida 😉
|
1.0
|
[Remoto] Java Backend Engineer Specialist na AgileProcess - ## Descrição da vaga
Buscamos um(a) **Backend Engineer Specialist** que será responsável, junto ao nossos squads de desenvolvimento, por prover a melhor experiência para nossos clientes através de nossas soluções.
RESPONSABILIDADES E ATRIBUIÇÕES
\- Desafiar o status quo e desenvolver soluções inovadoras para problemas complexos;
\- Desenvolver e manter nossos Microserviços de forma ágil, aplicando boas práticas de Engenharia de Software;
\- Contribuir com o desenvolvimento e arquitetura da plataforma, preparando-a para um crescimento acelerado;
\- Construir uma base sólida para o desenvolvimento de novos produtos;
\- Desenvolver sistemas escaláveis, sustentáveis e orientados ao usuário;
\- Trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência;
\- Ajudar o crescimento do time de tecnologia e engenharia.
REQUISITOS E QUALIFICAÇÕES
\- Experiência e conhecimento profundo com desenvolvimento Java 8;
\- Experiência e conhecimento em GitFlow;
\- Experiência e conhecimento profundo em Docker;
\- Ter atuado na construção de testes unitários e integrados;
\- Ter atuado na construção de testes de comportamento (BDD);
\- Conhecimentos em Design Patterns, arquitetura e engenharia de software;
\- Conhecimentos em GitLab CI;
\- Conhecimentos em metodologias ágeis de desenvolvimento (Scrum, Kanban).
INFORMAÇÕES ADICIONAIS
**Nosso Stack:**
\- Java;
\- MySQL;
\- AWS;
\- Git (GitLab).
**DNA Agile <3**
Faz parte do DNA de todo **AGILE** ser ambicioso e buscar entregar sempre mais e melhor. O respeito e a tolerância são indispensáveis para nós, e somos apaixonados pela nossa diversidade. Crescemos como time e celebrando juntos todos os resultados atingidos. Somos Customer Centric e buscamos criar conexão com nossos clientes, com simplicidade e empatia. Compartilhamos de forma direta e honesta todas as informações e opiniões possíveis. Reconhecemos nossas fraquezas e assumimos nossos erros. Somos apaixonados por inovação e usamos da nossa criatividade, autonomia e voz para propor novas formas de fazer as coisas.
Somos representados por cada um e nadamos todos na mesma direção.
**O que você vai encontrar:**
* Onboarding de boas-vindas!
* “All Hands”: nosso encontro semanal com o CEO;
* Dress Code: seja você mesmo(a);
* Flexibilidade de horário;
* VR/VA Flex: R$ 550,00 (mês);
* Plano de saúde (para você e quem você ama);
* Plano odontológico;
* TotalPass;
* Clube de Descontos - NewValue;
* PLR;
* Parceria com ZenKlub;
* e muito mais!
**Redes Sociais**: Venha conferir nossas redes sociais e revolucionar a logística com a gente!
* Instagram: @agileprocess
* LinkedIn: /agileprocess
* Twitter: @agileprocess
_\*Todas as vagas da AgileProcess são abertas para pessoas com deficiência (PcD)._
_\*Nos processos seletivos da AgileProcess, não há qualquer tipo de distinção em virtude de gênero, orientação sexual, religião, crença, condição social ou raça._
## Salário
R$12500 - R$14000
## Local
REMOTO
## Como se candidatar
Accesse a [Vaga de Java Backend Engineer Specialist](https://vaga.contratar.dev/apply/full/06686ffa-44e4-4505-a874-f719b0532a21) para se candidatar.
## Labels
* Remoto
* SQL
* Scrum
* Redes
* AWS
* Docker
* Java
* Git
## Sobre a contratar.dev
A contratar.dev é uma plataforma de recrutamento e seleção automatizada de Desenvolvedores. Ao se candidatar a uma vaga da contratar.dev, você estará participando automaticamente de dezenas de processos seletivos para vagas similares, aumentando em MUITO as suas chances de mudar de vida 😉
|
non_process
|
java backend engineer specialist na agileprocess descrição da vaga buscamos um a backend engineer specialist que será responsável junto ao nossos squads de desenvolvimento por prover a melhor experiência para nossos clientes através de nossas soluções responsabilidades e atribuições desafiar o status quo e desenvolver soluções inovadoras para problemas complexos desenvolver e manter nossos microserviços de forma ágil aplicando boas práticas de engenharia de software contribuir com o desenvolvimento e arquitetura da plataforma preparando a para um crescimento acelerado construir uma base sólida para o desenvolvimento de novos produtos desenvolver sistemas escaláveis sustentáveis e orientados ao usuário trabalhar em um ambiente que estimula e valoriza a autonomia e a transparência ajudar o crescimento do time de tecnologia e engenharia requisitos e qualificações experiência e conhecimento profundo com desenvolvimento java experiência e conhecimento em gitflow experiência e conhecimento profundo em docker ter atuado na construção de testes unitários e integrados ter atuado na construção de testes de comportamento bdd conhecimentos em design patterns arquitetura e engenharia de software conhecimentos em gitlab ci conhecimentos em metodologias ágeis de desenvolvimento scrum kanban informações adicionais nosso stack java mysql aws git gitlab dna agile faz parte do dna de todo agile ser ambicioso e buscar entregar sempre mais e melhor o respeito e a tolerância são indispensáveis para nós e somos apaixonados pela nossa diversidade crescemos como time e celebrando juntos todos os resultados atingidos somos customer centric e buscamos criar conexão com nossos clientes com simplicidade e empatia compartilhamos de forma direta e honesta todas as informações e opiniões possíveis reconhecemos nossas fraquezas e assumimos nossos erros somos apaixonados por inovação e usamos da nossa criatividade autonomia e voz para propor novas formas de fazer as coisas somos representados por cada um e nadamos todos na mesma direção o que você vai encontrar onboarding de boas vindas “all hands” nosso encontro semanal com o ceo dress code seja você mesmo a flexibilidade de horário vr va flex r mês plano de saúde para você e quem você ama plano odontológico totalpass clube de descontos newvalue plr parceria com zenklub e muito mais redes sociais venha conferir nossas redes sociais e revolucionar a logística com a gente instagram agileprocess linkedin agileprocess twitter agileprocess todas as vagas da agileprocess são abertas para pessoas com deficiência pcd nos processos seletivos da agileprocess não há qualquer tipo de distinção em virtude de gênero orientação sexual religião crença condição social ou raça salário r r local remoto como se candidatar accesse a para se candidatar labels remoto sql scrum redes aws docker java git sobre a contratar dev a contratar dev é uma plataforma de recrutamento e seleção automatizada de desenvolvedores ao se candidatar a uma vaga da contratar dev você estará participando automaticamente de dezenas de processos seletivos para vagas similares aumentando em muito as suas chances de mudar de vida 😉
| 0
|
2,937
| 5,920,860,911
|
IssuesEvent
|
2017-05-22 21:20:43
|
ncbo/bioportal-project
|
https://api.github.com/repos/ncbo/bioportal-project
|
closed
|
HIV: fails to parse
|
in progress ontology processing problem
|
Submission 2.6 of the [HIV ontology](http://bioportal.bioontology.org/ontologies/HIV), uploaded on March 20, 2017 failed to parse. BioPortal UI shows status of "Error Rdf".
Preliminary investigation shows several issues:
1. The user incorrectly designated the submission file format as OWL when it should be OBO. This shouldn't affect the parsing step though. The format should be corrected when the next submission is created.
2. Line 2 of their ontology source file uses an incorrect date format. The [OBO Flat File Format Guide](https://owlcollab.github.io/oboformat/doc/GO.format.obo-1_4.html) states that dates must be in a format of dd:MM:yyyy HH:mm. If you parse this ontology with version 4.2.5 of the OWL API, this exception is thrown:
```
Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@3888cf9
Stack trace:
LINENO: 2 - Could not parse date from string: December 27, 2016
LINE: date: December 27, 2016 org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60)
```
To fix this exception, change line 2 of the ontology source file to this:
```
date: 27:12:2016 00:00
```
3. After fixing line 2 of the ontology, re-parsing results in further errors, e.g.:
```
Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@15563bcf
Stack trace:
LINENO: 38 - Could not find tag for: part_of
LINE: part_of: HIV:29 ! retrovirus org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyFactoryImpl.loadOWLOntology(OWLOntologyFactoryImpl.java:197)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.actualParse(OWLOntologyManagerImpl.java:1061)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntology(OWLOntologyManagerImpl.java:1017)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntologyFromOntologyDocument(OWLOntologyManagerImpl.java:967)
edu.stanford.ncbo.bioportal.SimpleLoadTest.testLoadHIVOntology_WithDocumentFormat(SimpleLoadTest.java:50)
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
java.lang.reflect.Method.invoke(Method.java:606)
```
These exceptions are thrown with version 4.2.5 of the OWL API, which BioPortal currently uses internally. Version 4.2.6 (and above) of the OWL API successfully parse the ontology with no exceptions. The user reported that they could open the ontology in the Protege ontology editor without issues. The latest version of Protege (5.2.0) uses version 4.2.8 of the OWL API, thus the lack of issue opening the HIV ontology in their environment.
|
1.0
|
HIV: fails to parse - Submission 2.6 of the [HIV ontology](http://bioportal.bioontology.org/ontologies/HIV), uploaded on March 20, 2017 failed to parse. BioPortal UI shows status of "Error Rdf".
Preliminary investigation shows several issues:
1. The user incorrectly designated the submission file format as OWL when it should be OBO. This shouldn't affect the parsing step though. The format should be corrected when the next submission is created.
2. Line 2 of their ontology source file uses an incorrect date format. The [OBO Flat File Format Guide](https://owlcollab.github.io/oboformat/doc/GO.format.obo-1_4.html) states that dates must be in a format of dd:MM:yyyy HH:mm. If you parse this ontology with version 4.2.5 of the OWL API, this exception is thrown:
```
Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@3888cf9
Stack trace:
LINENO: 2 - Could not parse date from string: December 27, 2016
LINE: date: December 27, 2016 org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60)
```
To fix this exception, change line 2 of the ontology source file to this:
```
date: 27:12:2016 00:00
```
3. After fixing line 2 of the ontology, re-parsing results in further errors, e.g.:
```
Parser: org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser@15563bcf
Stack trace:
LINENO: 38 - Could not find tag for: part_of
LINE: part_of: HIV:29 ! retrovirus org.semanticweb.owlapi.oboformat.OBOFormatOWLAPIParser.parse(OBOFormatOWLAPIParser.java:60)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyFactoryImpl.loadOWLOntology(OWLOntologyFactoryImpl.java:197)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.actualParse(OWLOntologyManagerImpl.java:1061)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntology(OWLOntologyManagerImpl.java:1017)
uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.loadOntologyFromOntologyDocument(OWLOntologyManagerImpl.java:967)
edu.stanford.ncbo.bioportal.SimpleLoadTest.testLoadHIVOntology_WithDocumentFormat(SimpleLoadTest.java:50)
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
java.lang.reflect.Method.invoke(Method.java:606)
```
These exceptions are thrown with version 4.2.5 of the OWL API, which BioPortal currently uses internally. Version 4.2.6 (and above) of the OWL API successfully parse the ontology with no exceptions. The user reported that they could open the ontology in the Protege ontology editor without issues. The latest version of Protege (5.2.0) uses version 4.2.8 of the OWL API, thus the lack of issue opening the HIV ontology in their environment.
|
process
|
hiv fails to parse submission of the uploaded on march failed to parse bioportal ui shows status of error rdf preliminary investigation shows several issues the user incorrectly designated the submission file format as owl when it should be obo this shouldn t affect the parsing step though the format should be corrected when the next submission is created line of their ontology source file uses an incorrect date format the states that dates must be in a format of dd mm yyyy hh mm if you parse this ontology with version of the owl api this exception is thrown parser org semanticweb owlapi oboformat oboformatowlapiparser stack trace lineno could not parse date from string december line date december org semanticweb owlapi oboformat oboformatowlapiparser parse oboformatowlapiparser java to fix this exception change line of the ontology source file to this date after fixing line of the ontology re parsing results in further errors e g parser org semanticweb owlapi oboformat oboformatowlapiparser stack trace lineno could not find tag for part of line part of hiv retrovirus org semanticweb owlapi oboformat oboformatowlapiparser parse oboformatowlapiparser java uk ac manchester cs owl owlapi owlontologyfactoryimpl loadowlontology owlontologyfactoryimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl actualparse owlontologymanagerimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl loadontology owlontologymanagerimpl java uk ac manchester cs owl owlapi owlontologymanagerimpl loadontologyfromontologydocument owlontologymanagerimpl java edu stanford ncbo bioportal simpleloadtest testloadhivontology withdocumentformat simpleloadtest java sun reflect nativemethodaccessorimpl native method sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java java lang reflect method invoke method java these exceptions are thrown with version of the owl api which bioportal currently uses internally version and above of the owl api successfully parse the ontology with no exceptions the user reported that they could open the ontology in the protege ontology editor without issues the latest version of protege uses version of the owl api thus the lack of issue opening the hiv ontology in their environment
| 1
|
9,277
| 12,302,870,443
|
IssuesEvent
|
2020-05-11 17:42:35
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
RPM for openSUSE / SUSE Enterprise Linux / Need for dependency listing.
|
team-EngProd type: process under investigation untriaged
|
* Operating System: Linux
* Bazel version (output of `bazel info release`):
0.8.0
#2701 is related
I'm developing an rpm for openSUSE / SUSE which will be built on https://build.opensuse.org and hosted in a repo with a pretty permanent address on download.opensuse.org
Before getting it officially in the distro, we need to be able to build bazel without the bundled third party bits. To do so, I would need a full list of the dependencies either as jars or libraries. Is there a way to get bazel to use the distro version of protobuf ? That is already in the distro and I suspect most of the bundled jars.
|
1.0
|
RPM for openSUSE / SUSE Enterprise Linux / Need for dependency listing. - * Operating System: Linux
* Bazel version (output of `bazel info release`):
0.8.0
#2701 is related
I'm developing an rpm for openSUSE / SUSE which will be built on https://build.opensuse.org and hosted in a repo with a pretty permanent address on download.opensuse.org
Before getting it officially in the distro, we need to be able to build bazel without the bundled third party bits. To do so, I would need a full list of the dependencies either as jars or libraries. Is there a way to get bazel to use the distro version of protobuf ? That is already in the distro and I suspect most of the bundled jars.
|
process
|
rpm for opensuse suse enterprise linux need for dependency listing operating system linux bazel version output of bazel info release is related i m developing an rpm for opensuse suse which will be built on and hosted in a repo with a pretty permanent address on download opensuse org before getting it officially in the distro we need to be able to build bazel without the bundled third party bits to do so i would need a full list of the dependencies either as jars or libraries is there a way to get bazel to use the distro version of protobuf that is already in the distro and i suspect most of the bundled jars
| 1
|
15,696
| 19,848,160,629
|
IssuesEvent
|
2022-01-21 09:16:26
|
ooi-data/CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered
|
https://api.github.com/repos/ooi-data/CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:16:25.469693.
## Details
Flow name: `CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:16:25.469693.
## Details
Flow name: `CE09OSPM-WFP01-02-DOFSTK000-recovered_wfp-dofst_k_wfp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp dofst k wfp instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
153,202
| 12,137,397,699
|
IssuesEvent
|
2020-04-23 15:40:25
|
deathlyrage/pot-demo-bugs
|
https://api.github.com/repos/deathlyrage/pot-demo-bugs
|
closed
|
Deinonychus slides when it gets up from rest
|
needs testing
|
Deinonychus slides when it gets up from rest/movement doesn't match animation.
|
1.0
|
Deinonychus slides when it gets up from rest - Deinonychus slides when it gets up from rest/movement doesn't match animation.
|
non_process
|
deinonychus slides when it gets up from rest deinonychus slides when it gets up from rest movement doesn t match animation
| 0
|
6,066
| 8,902,729,898
|
IssuesEvent
|
2019-01-17 08:33:08
|
Juris-M/citeproc-js
|
https://api.github.com/repos/Juris-M/citeproc-js
|
closed
|
Bug: creator label formatting
|
fix in process
|
I am working with this style:
https://gist.github.com/bwiernik/f33d5ebf8793fbe663f826d478298c76
In the `"format"` macro, I use `names` to call `reviewed-author`:
```
<names variable="reviewed-author">
<label form="verb-short" suffix=" "/>
<name and="symbol" initialize-with=". " delimiter=", "/>
</names>
```
The expected output for this item:
```
[
{
"id": "http://zotero.org/groups/2205533/items/IJBGS4BQ",
"type": "post-weblog",
"title": "The future for Deaf individuals is not that bleak",
"URL": "http://www.personalityresearch.org/papers/hall.html",
"note": "reviewed-title: Decrease of Deaf potential in a mainstreamed environment\ngenre: Peer commentary\nreviewed-author: Hall || W.C.",
"language": "en",
"author": [
{
"family": "Wolf",
"given": "K.S."
}
],
"issued": {
"date-parts": [
[
"2005"
]
]
}
}
]
```
is:
Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, by W. C. Hall]. Retrieved from http://www.personalityresearch.org/papers/hall.html
However, what renders instead is:
Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, W. C. Hallby ]. Retrieved from http://www.personalityresearch.org/papers/hall.html
With the `label` coming after `name`, even though it is specified to come before. If I move the `names` element out of the `"format"` macro (e.g., directly into the bibliography layout or into the `"title"` macro), it renders correctly. If I change `reviewed-author` to another creator variable, the formatting of `label` still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable (or following the `name` if it is not specified elsewhere).
The cause is that `<text macro="format">` is included in the `substitute` list of the `"author-short"` macro. If I delete the three instances of `<text macro="format">` from the `"author-short"` macro (all three must be deleted, even the two in conditional branches that do not apply to the item), the `label` formatting is rendered correctly.
@adam3smith
|
1.0
|
Bug: creator label formatting - I am working with this style:
https://gist.github.com/bwiernik/f33d5ebf8793fbe663f826d478298c76
In the `"format"` macro, I use `names` to call `reviewed-author`:
```
<names variable="reviewed-author">
<label form="verb-short" suffix=" "/>
<name and="symbol" initialize-with=". " delimiter=", "/>
</names>
```
The expected output for this item:
```
[
{
"id": "http://zotero.org/groups/2205533/items/IJBGS4BQ",
"type": "post-weblog",
"title": "The future for Deaf individuals is not that bleak",
"URL": "http://www.personalityresearch.org/papers/hall.html",
"note": "reviewed-title: Decrease of Deaf potential in a mainstreamed environment\ngenre: Peer commentary\nreviewed-author: Hall || W.C.",
"language": "en",
"author": [
{
"family": "Wolf",
"given": "K.S."
}
],
"issued": {
"date-parts": [
[
"2005"
]
]
}
}
]
```
is:
Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, by W. C. Hall]. Retrieved from http://www.personalityresearch.org/papers/hall.html
However, what renders instead is:
Wolf, K. S. (2005). The future for Deaf individuals is not that bleak [Peer commentary, _Decrease of Deaf potential in a mainstreamed environment_, W. C. Hallby ]. Retrieved from http://www.personalityresearch.org/papers/hall.html
With the `label` coming after `name`, even though it is specified to come before. If I move the `names` element out of the `"format"` macro (e.g., directly into the bibliography layout or into the `"title"` macro), it renders correctly. If I change `reviewed-author` to another creator variable, the formatting of `label` still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable (or following the `name` if it is not specified elsewhere).
The cause is that `<text macro="format">` is included in the `substitute` list of the `"author-short"` macro. If I delete the three instances of `<text macro="format">` from the `"author-short"` macro (all three must be deleted, even the two in conditional branches that do not apply to the item), the `label` formatting is rendered correctly.
@adam3smith
|
process
|
bug creator label formatting i am working with this style in the format macro i use names to call reviewed author the expected output for this item id type post weblog title the future for deaf individuals is not that bleak url note reviewed title decrease of deaf potential in a mainstreamed environment ngenre peer commentary nreviewed author hall w c language en author family wolf given k s issued date parts is wolf k s the future for deaf individuals is not that bleak retrieved from however what renders instead is wolf k s the future for deaf individuals is not that bleak retrieved from with the label coming after name even though it is specified to come before if i move the names element out of the format macro e g directly into the bibliography layout or into the title macro it renders correctly if i change reviewed author to another creator variable the formatting of label still ignores the formatting specified here and instead uses the formatting given elsewhere in the style for that variable or following the name if it is not specified elsewhere the cause is that is included in the substitute list of the author short macro if i delete the three instances of from the author short macro all three must be deleted even the two in conditional branches that do not apply to the item the label formatting is rendered correctly
| 1
|
439,241
| 12,679,747,385
|
IssuesEvent
|
2020-06-19 12:25:49
|
nokazn/spotify-player
|
https://api.github.com/repos/nokazn/spotify-player
|
closed
|
公式のランキングのプレイリストの追加日
|
0. low-priority 2. improvement 💪
|
## 問題
spotifycharts が作成したプレイリスト (バイラルチャートなど) では追加日が `1970-01-01-T00:00:00Z`になっている。
|
1.0
|
公式のランキングのプレイリストの追加日 - ## 問題
spotifycharts が作成したプレイリスト (バイラルチャートなど) では追加日が `1970-01-01-T00:00:00Z`になっている。
|
non_process
|
公式のランキングのプレイリストの追加日 問題 spotifycharts が作成したプレイリスト バイラルチャートなど では追加日が になっている。
| 0
|
3,696
| 3,222,840,851
|
IssuesEvent
|
2015-10-09 05:26:35
|
DynamoRIO/dynamorio
|
https://api.github.com/repos/DynamoRIO/dynamorio
|
closed
|
strange drgui release build error with gcc 5.1+
|
Component-Build OpSys-Linux
|
Here's another release build error with gcc 5.1.1 (xref #1796), first mentioned in #1788:
```
/work/dr/git/src/ext/drgui/drgui_main_window.cpp: In destructor ‘drgui_main_window_t::~drgui_main_window_t()’:
/work/dr/git/src/ext/drgui/drgui_main_window.cpp:100:1: error: assuming signed overflow does not occur when assuming that (X - c) <= X is always true [-Werror=strict-overflow]
drgui_main_window_t::~drgui_main_window_t(void)
```
It took a while to figure out where this was coming from: it's the line
"plugins.pop_back()" and it's QVector<T>::reallocData() being inlined
(pop_back() calls removeLast() which calls reallocData()). This is the
line in reallocData() in /usr/include/qt5/QtCore/qvector.h:
```
if (asize <= d->size) {
```
Changing it to this removes the warning:
```
if ((uint)asize <= d->size) {
```
Of course we don't want everyone to go edit their system headers so we'll
turn off this warning.
|
1.0
|
strange drgui release build error with gcc 5.1+ - Here's another release build error with gcc 5.1.1 (xref #1796), first mentioned in #1788:
```
/work/dr/git/src/ext/drgui/drgui_main_window.cpp: In destructor ‘drgui_main_window_t::~drgui_main_window_t()’:
/work/dr/git/src/ext/drgui/drgui_main_window.cpp:100:1: error: assuming signed overflow does not occur when assuming that (X - c) <= X is always true [-Werror=strict-overflow]
drgui_main_window_t::~drgui_main_window_t(void)
```
It took a while to figure out where this was coming from: it's the line
"plugins.pop_back()" and it's QVector<T>::reallocData() being inlined
(pop_back() calls removeLast() which calls reallocData()). This is the
line in reallocData() in /usr/include/qt5/QtCore/qvector.h:
```
if (asize <= d->size) {
```
Changing it to this removes the warning:
```
if ((uint)asize <= d->size) {
```
Of course we don't want everyone to go edit their system headers so we'll
turn off this warning.
|
non_process
|
strange drgui release build error with gcc here s another release build error with gcc xref first mentioned in work dr git src ext drgui drgui main window cpp in destructor ‘drgui main window t drgui main window t ’ work dr git src ext drgui drgui main window cpp error assuming signed overflow does not occur when assuming that x c x is always true drgui main window t drgui main window t void it took a while to figure out where this was coming from it s the line plugins pop back and it s qvector reallocdata being inlined pop back calls removelast which calls reallocdata this is the line in reallocdata in usr include qtcore qvector h if asize size changing it to this removes the warning if uint asize size of course we don t want everyone to go edit their system headers so we ll turn off this warning
| 0
|
8,496
| 11,659,800,299
|
IssuesEvent
|
2020-03-03 01:11:49
|
MicrosoftDocs/vsts-docs
|
https://api.github.com/repos/MicrosoftDocs/vsts-docs
|
closed
|
dependencies for stages?
|
Pri1 devops-cicd-process/tech devops/prod doc-bug
|
How does dependencies work between stages? Can we use output variables from jobs in previous stages?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#type-casting)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
dependencies for stages? - How does dependencies work between stages? Can we use output variables from jobs in previous stages?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#type-casting)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
dependencies for stages how does dependencies work between stages can we use output variables from jobs in previous stages document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
20,347
| 27,002,946,217
|
IssuesEvent
|
2023-02-10 09:23:27
|
inmanta/inmanta-core
|
https://api.github.com/repos/inmanta/inmanta-core
|
opened
|
configure dependabot to use the widen strategy
|
process
|
dependabot/dependabot-core#6632 implements the `widen` strategy for Python packages. The behavior for this strategy matches the behavior we want and have gotten up till now. However, the default behavior wasn't necessarily meant to be like this (dependabot/dependabot-core#6631) and might change in the future. Once the next version of `dependabot` is released (`>0.215.0`) we'd best configure dependabot explicitly to use the desired strategy for all relevant repos (at least product and component repos, probably all our tools (pytest-inmanta, vscode-inmanta, ...), I'm not sure about modules).
|
1.0
|
configure dependabot to use the widen strategy - dependabot/dependabot-core#6632 implements the `widen` strategy for Python packages. The behavior for this strategy matches the behavior we want and have gotten up till now. However, the default behavior wasn't necessarily meant to be like this (dependabot/dependabot-core#6631) and might change in the future. Once the next version of `dependabot` is released (`>0.215.0`) we'd best configure dependabot explicitly to use the desired strategy for all relevant repos (at least product and component repos, probably all our tools (pytest-inmanta, vscode-inmanta, ...), I'm not sure about modules).
|
process
|
configure dependabot to use the widen strategy dependabot dependabot core implements the widen strategy for python packages the behavior for this strategy matches the behavior we want and have gotten up till now however the default behavior wasn t necessarily meant to be like this dependabot dependabot core and might change in the future once the next version of dependabot is released we d best configure dependabot explicitly to use the desired strategy for all relevant repos at least product and component repos probably all our tools pytest inmanta vscode inmanta i m not sure about modules
| 1
|
336,254
| 24,491,011,260
|
IssuesEvent
|
2022-10-10 01:51:21
|
sytone/obsidian-remote
|
https://api.github.com/repos/sytone/obsidian-remote
|
closed
|
Missing TZ variable
|
documentation
|
**Is your feature request related to a problem? Please describe.**
There are plugins that use the current time of the system to work. Right now it shows an hour less than my Timezone.
**Describe the solution you'd like**
It would be awesome to set it as a TZ variable in the docker container.
|
1.0
|
Missing TZ variable - **Is your feature request related to a problem? Please describe.**
There are plugins that use the current time of the system to work. Right now it shows an hour less than my Timezone.
**Describe the solution you'd like**
It would be awesome to set it as a TZ variable in the docker container.
|
non_process
|
missing tz variable is your feature request related to a problem please describe there are plugins that use the current time of the system to work right now it shows an hour less than my timezone describe the solution you d like it would be awesome to set it as a tz variable in the docker container
| 0
|
25,454
| 7,708,972,621
|
IssuesEvent
|
2018-05-22 07:40:44
|
xamarin/xamarin-android
|
https://api.github.com/repos/xamarin/xamarin-android
|
closed
|
Getting build failure for Cheesesquare sample in both debug and release mode
|
Area: App+Library Build
|
Moving VSTS bug [BUG 614180](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/614180) to this repo
Affected Branch: master
Affected Build: Version 7.6 Preview (7.6 build 750)
Affected Product Language:
While building "Cheesesquare" sample in VSFM , build failure is observed with below error
"/Users/globallogic/Downloads/monodroid-samples-master/android5.0/Cheesesquare/obj/Debug/generated/Resources/layout/list_item.xml(38,38): Error CS0234: The type or namespace name 'Android' does not exist in the namespace 'Cheesesquare' (are you missing an assembly reference?) (CS0234) (Cheesesquare)"
Build JSON : [https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json](https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json)
Sample link : [https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare](https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare)
Steps to reproduce:
Download or clone the above sample and open it in VS for MAC
Set build config as debug and build the project
Repeat same for Release mode
Expected : App should built successfully
Actual : Build failure is observed with error mentioned above
Build info : [https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0](https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0)
Reproducibility : 100%
Failure is observed in Automation run as well
[http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt](http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt
)
Screen-recording : [http://g.recordit.co/LK1X0YUlvd.gif](http://g.recordit.co/LK1X0YUlvd.gif)
[build and Ide logs.zip](https://github.com/xamarin/xamarin-android/files/1991562/build.and.Ide.logs.zip)
Attaching IDE and Build logs too
|
1.0
|
Getting build failure for Cheesesquare sample in both debug and release mode - Moving VSTS bug [BUG 614180](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/614180) to this repo
Affected Branch: master
Affected Build: Version 7.6 Preview (7.6 build 750)
Affected Product Language:
While building "Cheesesquare" sample in VSFM , build failure is observed with below error
"/Users/globallogic/Downloads/monodroid-samples-master/android5.0/Cheesesquare/obj/Debug/generated/Resources/layout/list_item.xml(38,38): Error CS0234: The type or namespace name 'Android' does not exist in the namespace 'Cheesesquare' (are you missing an assembly reference?) (CS0234) (Cheesesquare)"
Build JSON : [https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json](https://github.com/xamarin/release-manifests/blob/nightly-master/manifests/0_nightly_smoke_test_master.json)
Sample link : [https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare](https://github.com/xamarin/monodroid-samples/tree/master/android5.0/Cheesesquare)
Steps to reproduce:
Download or clone the above sample and open it in VS for MAC
Set build config as debug and build the project
Repeat same for Release mode
Expected : App should built successfully
Actual : Build failure is observed with error mentioned above
Build info : [https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0](https://gist.github.com/parasAmbhore/c78a44483529b357b626830cd0e9e4d0)
Reproducibility : 100%
Failure is observed in Automation run as well
[http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt](http://xqa.blob.core.windows.net/gist/TestReport-4da6427447aa4bdb8d52345201da9430.txt
)
Screen-recording : [http://g.recordit.co/LK1X0YUlvd.gif](http://g.recordit.co/LK1X0YUlvd.gif)
[build and Ide logs.zip](https://github.com/xamarin/xamarin-android/files/1991562/build.and.Ide.logs.zip)
Attaching IDE and Build logs too
|
non_process
|
getting build failure for cheesesquare sample in both debug and release mode moving vsts bug to this repo affected branch master affected build version preview build affected product language while building cheesesquare sample in vsfm build failure is observed with below error users globallogic downloads monodroid samples master cheesesquare obj debug generated resources layout list item xml error the type or namespace name android does not exist in the namespace cheesesquare are you missing an assembly reference cheesesquare build json sample link steps to reproduce download or clone the above sample and open it in vs for mac set build config as debug and build the project repeat same for release mode expected app should built successfully actual build failure is observed with error mentioned above build info reproducibility failure is observed in automation run as well screen recording attaching ide and build logs too
| 0
|
71,918
| 30,927,999,339
|
IssuesEvent
|
2023-08-06 18:24:29
|
ArjunaCluster/ArjunaUsers
|
https://api.github.com/repos/ArjunaCluster/ArjunaUsers
|
opened
|
Arjuna is Offline (Unplanned Outage)
|
service-outage notify
|
Arjuna is currently offline (likely storm related), and we have disabled ssh access to the cluster while we return the nodes to service.
We will update this thread with more information as it becomes available.
Best,
The Arjuna Admin Team
|
1.0
|
Arjuna is Offline (Unplanned Outage) - Arjuna is currently offline (likely storm related), and we have disabled ssh access to the cluster while we return the nodes to service.
We will update this thread with more information as it becomes available.
Best,
The Arjuna Admin Team
|
non_process
|
arjuna is offline unplanned outage arjuna is currently offline likely storm related and we have disabled ssh access to the cluster while we return the nodes to service we will update this thread with more information as it becomes available best the arjuna admin team
| 0
|
781,267
| 27,430,210,787
|
IssuesEvent
|
2023-03-02 00:19:07
|
coral-xyz/backpack
|
https://api.github.com/repos/coral-xyz/backpack
|
closed
|
trailing malformed characters for nft names
|
help wanted good first issue priority 1 polish
|
This is probably a result of metaplex's deserialization of null bytes of the metadata.

|
1.0
|
trailing malformed characters for nft names - This is probably a result of metaplex's deserialization of null bytes of the metadata.

|
non_process
|
trailing malformed characters for nft names this is probably a result of metaplex s deserialization of null bytes of the metadata
| 0
|
95,656
| 16,096,857,278
|
IssuesEvent
|
2021-04-27 01:59:09
|
AnhaaD/auth-server
|
https://api.github.com/repos/AnhaaD/auth-server
|
opened
|
CVE-2021-31597 (Medium) detected in xmlhttprequest-ssl-1.5.3.tgz
|
security vulnerability
|
## CVE-2021-31597 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.3.tgz</b></p></summary>
<p>XMLHttpRequest for Node</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz</a></p>
<p>Path to dependency file: /auth-server/package.json</p>
<p>Path to vulnerable library: auth-server/node_modules/xmlhttprequest-ssl/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- socket.io-1.7.3.tgz
- socket.io-client-1.7.3.tgz
- engine.io-client-1.8.3.tgz
- :x: **xmlhttprequest-ssl-1.5.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The xmlhttprequest-ssl package before 1.6.1 for Node.js disables SSL certificate validation by default, because rejectUnauthorized (when the property exists but is undefined) is considered to be false within the https.request function of Node.js. In other words, no certificate is ever rejected.
<p>Publish Date: 2021-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31597>CVE-2021-31597</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597</a></p>
<p>Release Date: 2021-04-23</p>
<p>Fix Resolution: xmlhttprequest-ssl - 1.6.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-31597 (Medium) detected in xmlhttprequest-ssl-1.5.3.tgz - ## CVE-2021-31597 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.3.tgz</b></p></summary>
<p>XMLHttpRequest for Node</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.3.tgz</a></p>
<p>Path to dependency file: /auth-server/package.json</p>
<p>Path to vulnerable library: auth-server/node_modules/xmlhttprequest-ssl/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- socket.io-1.7.3.tgz
- socket.io-client-1.7.3.tgz
- engine.io-client-1.8.3.tgz
- :x: **xmlhttprequest-ssl-1.5.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The xmlhttprequest-ssl package before 1.6.1 for Node.js disables SSL certificate validation by default, because rejectUnauthorized (when the property exists but is undefined) is considered to be false within the https.request function of Node.js. In other words, no certificate is ever rejected.
<p>Publish Date: 2021-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31597>CVE-2021-31597</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-31597</a></p>
<p>Release Date: 2021-04-23</p>
<p>Fix Resolution: xmlhttprequest-ssl - 1.6.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in xmlhttprequest ssl tgz cve medium severity vulnerability vulnerable library xmlhttprequest ssl tgz xmlhttprequest for node library home page a href path to dependency file auth server package json path to vulnerable library auth server node modules xmlhttprequest ssl package json dependency hierarchy karma tgz root library socket io tgz socket io client tgz engine io client tgz x xmlhttprequest ssl tgz vulnerable library vulnerability details the xmlhttprequest ssl package before for node js disables ssl certificate validation by default because rejectunauthorized when the property exists but is undefined is considered to be false within the https request function of node js in other words no certificate is ever rejected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmlhttprequest ssl step up your open source security game with whitesource
| 0
|
20,378
| 27,031,457,435
|
IssuesEvent
|
2023-02-12 08:43:11
|
Open-Data-Product-Initiative/open-data-product-spec
|
https://api.github.com/repos/Open-Data-Product-Initiative/open-data-product-spec
|
closed
|
Product management: Data product portfolio, product line, product series etc.
|
enhancement unprocessed
|
Data product document level attribute contains information about the product portfolio, family, line, series etc. to which it belongs.
This requires a discussion, especially from a product management perspective.
|
1.0
|
Product management: Data product portfolio, product line, product series etc. - Data product document level attribute contains information about the product portfolio, family, line, series etc. to which it belongs.
This requires a discussion, especially from a product management perspective.
|
process
|
product management data product portfolio product line product series etc data product document level attribute contains information about the product portfolio family line series etc to which it belongs this requires a discussion especially from a product management perspective
| 1
|
20,932
| 27,778,609,847
|
IssuesEvent
|
2023-03-16 19:08:30
|
googleapis/python-bigquery-reservation
|
https://api.github.com/repos/googleapis/python-bigquery-reservation
|
opened
|
Warning: a recent release failed
|
type: process
|
The following release PRs may have failed:
* #49 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #42 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
|
1.0
|
Warning: a recent release failed - The following release PRs may have failed:
* #49 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #42 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
|
process
|
warning a recent release failed the following release prs may have failed the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published
| 1
|
9,565
| 12,519,484,773
|
IssuesEvent
|
2020-06-03 14:30:33
|
code4romania/expert-consultation-api
|
https://api.github.com/repos/code4romania/expert-consultation-api
|
closed
|
[Users] Import a list of users by copy/paste from Excel file
|
document processing documents java nice to have spring
|
As an admin user of the Legal Consultation platform I want to be able to import a list of users by copy pasting from an excel file.

|
1.0
|
[Users] Import a list of users by copy/paste from Excel file - As an admin user of the Legal Consultation platform I want to be able to import a list of users by copy pasting from an excel file.

|
process
|
import a list of users by copy paste from excel file as an admin user of the legal consultation platform i want to be able to import a list of users by copy pasting from an excel file
| 1
|
5,440
| 8,304,002,889
|
IssuesEvent
|
2018-09-21 19:33:16
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
Release google-cloud-bigquery 1.5.1
|
api: bigquery type: process
|
We've had a recent bug fix PR that needs to get released.
I forget if we've had any feature PRs since the last one.
|
1.0
|
Release google-cloud-bigquery 1.5.1 - We've had a recent bug fix PR that needs to get released.
I forget if we've had any feature PRs since the last one.
|
process
|
release google cloud bigquery we ve had a recent bug fix pr that needs to get released i forget if we ve had any feature prs since the last one
| 1
|
4,998
| 7,834,347,697
|
IssuesEvent
|
2018-06-16 12:41:41
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Add documentation for Python Scripts (Trac #768)
|
Migrated from Trac enhancement post_processing weberjk@uwm.edu
|
Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/768
```json
{
"status": "closed",
"changetime": "2015-07-15T02:14:44",
"description": "Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. ",
"reporter": "weberjk@uwm.edu",
"cc": "vlarson@uwm.edu",
"resolution": "worksforme",
"_ts": "1436926484509531",
"component": "post_processing",
"summary": "Add documentation for Python Scripts",
"priority": "minor",
"keywords": "",
"time": "2015-04-23T15:30:13",
"milestone": "",
"owner": "weberjk@uwm.edu",
"type": "enhancement"
}
```
|
1.0
|
Add documentation for Python Scripts (Trac #768) - Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/768
```json
{
"status": "closed",
"changetime": "2015-07-15T02:14:44",
"description": "Justin added some python scripts that he used for his thesis. We may want to use them in the future. Let's add to their documentation. Use this ticket when making svn commits. ",
"reporter": "weberjk@uwm.edu",
"cc": "vlarson@uwm.edu",
"resolution": "worksforme",
"_ts": "1436926484509531",
"component": "post_processing",
"summary": "Add documentation for Python Scripts",
"priority": "minor",
"keywords": "",
"time": "2015-04-23T15:30:13",
"milestone": "",
"owner": "weberjk@uwm.edu",
"type": "enhancement"
}
```
|
process
|
add documentation for python scripts trac justin added some python scripts that he used for his thesis we may want to use them in the future let s add to their documentation use this ticket when making svn commits attachments migrated from json status closed changetime description justin added some python scripts that he used for his thesis we may want to use them in the future let s add to their documentation use this ticket when making svn commits reporter weberjk uwm edu cc vlarson uwm edu resolution worksforme ts component post processing summary add documentation for python scripts priority minor keywords time milestone owner weberjk uwm edu type enhancement
| 1
|
12,043
| 14,738,753,031
|
IssuesEvent
|
2021-01-07 05:38:07
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Receiving error message in SAB
|
anc-ops anc-process anp-1 ant-bug ant-support has attachment
|
In GitLab by @kdjstudios on Jul 20, 2018, 08:23
**Submitted by:** "Sarah Baptist" <sarah.baptist@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-19-29832/conversation
**Server:** Internal
**Client/Site:** Stockton
**Account:** NA
**Issue:**
When trying to update/edit client account information in SAB I keep receiving this error. The client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time.

|
1.0
|
Receiving error message in SAB - In GitLab by @kdjstudios on Jul 20, 2018, 08:23
**Submitted by:** "Sarah Baptist" <sarah.baptist@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-19-29832/conversation
**Server:** Internal
**Client/Site:** Stockton
**Account:** NA
**Issue:**
When trying to update/edit client account information in SAB I keep receiving this error. The client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time.

|
process
|
receiving error message in sab in gitlab by kdjstudios on jul submitted by sarah baptist helpdesk server internal client site stockton account na issue when trying to update edit client account information in sab i keep receiving this error the client is also receiving an error that the their login information is invalid and the account is unable to be accessed at this time uploads sab error png
| 1
|
11,397
| 14,234,427,052
|
IssuesEvent
|
2020-11-18 13:35:05
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
fatal error: 'sys/inotify.h' file not found
|
P3 team-Rules-CPP type: support / not a bug (process)
|
### Description of the problem / feature request:
when I run `bazel build //test:test`, sometimes I got the error info ,
```
INFO: Analyzed target //test:test (4 packages loaded, 73 targets configured).
INFO: Found 1 target...
ERROR: /private/var/tmp/_bazel_core/c1311ca8ca6f559bd475572aa40dd5b5/external/glfw/BUILD.bazel:120:11: C++ compilation of rule '@glfw//:linux_joystick' failed (Exit 1) wrapped_clang failed: error executing command external/local_config_cc/wrapped_clang '-D_FORTIFY_SOURCE=1' -fstack-protector -fcolor-diagnostics -Wall -Wthread-safety -Wself-assign -fno-omit-frame-pointer -O0 -DDEBUG -iquote external/glfw -iquote ... (remaining 29 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox
external/glfw/src/linux_joystick.c:34:10: fatal error: 'sys/inotify.h' file not found
#include <sys/inotify.h>
^~~~~~~~~~~~~~~
1 error generated.
Target //test:test failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 8.039s, Critical Path: 4.03s
INFO: 14 processes: 5 internal, 9 darwin-sandbox.
FAILED: Build did NOT complete successfully
```
So my problem is , **how to import `<sys/inotify.h>`**
### Feature requests: what underlying problem are you trying to solve with this feature?
import glfw library by bazel.First of all , I solve the problem `fatal error : X11/Xlib.h not found`.But this problem occurs randomly
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
in my `third_party/glfw`, the docs structure is :
```
.
├── BUILD
├── glfw.BUILD
└── workspace.bzl
```
in my `workspace.bzl`:
```
def clean_dep(dep):
return str(Label(dep))
def repo():
//download glfw to local
native.new_local_repository(
name = "glfw",
path = "/Users/core/Documents/Github/glfw",
build_file = clean_dep("//third_party/glfw:glfw.BUILD"),
)
```
in my glfw.BUILD :
```
package(
default_visibility = ["//visibility:public"],
)
DEFINES = [
"_GLFW_X11",
"_GLFW_HAS_XF86VM",
]
cc_library(
name = "includes",
hdrs = glob([
"include/GLFW/*.h",
"src/*.h",
]),
)
cc_library(
name = "context",
srcs = ["src/context.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "init",
srcs = ["src/init.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "input",
srcs = ["src/input.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "monitor",
srcs = ["src/monitor.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "vulkan",
srcs = ["src/vulkan.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "window",
srcs = ["src/window.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_init",
srcs = ["src/x11_init.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_monitor",
srcs = ["src/x11_monitor.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_window",
srcs = ["src/x11_window.c"],
defines = DEFINES,
deps = [
":egl_context",
":glx_context",
":includes",
"@x11",
],
)
cc_library(
name = "xkb_unicode",
srcs = ["src/xkb_unicode.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "linux_joystick",
srcs = ["src/linux_joystick.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "posix_time",
srcs = ["src/posix_time.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "glx_context",
srcs = ["src/glx_context.c"],
defines = DEFINES,
deps = [
":context",
":includes",
# ":posix_tls",
"@x11",
],
)
cc_library(
name = "egl_context",
srcs = ["src/egl_context.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "glfw",
hdrs = glob(["include/GLFW/*.h"]),
defines = DEFINES,
strip_include_prefix = "include",
visibility = ["//visibility:public"],
deps = [
":context",
":egl_context",
":glx_context",
":init",
":input",
":linux_joystick",
":monitor",
":posix_time",
#":posix_tls",
":vulkan",
":window",
":x11_init",
":x11_monitor",
":x11_window",
":xkb_unicode",
],
)
```
### What operating system are you running Bazel on?
Macos10.14.6
### What's the output of `bazel info release`?
release 3.6.0-homebrew
|
1.0
|
fatal error: 'sys/inotify.h' file not found - ### Description of the problem / feature request:
when I run `bazel build //test:test`, sometimes I got the error info ,
```
INFO: Analyzed target //test:test (4 packages loaded, 73 targets configured).
INFO: Found 1 target...
ERROR: /private/var/tmp/_bazel_core/c1311ca8ca6f559bd475572aa40dd5b5/external/glfw/BUILD.bazel:120:11: C++ compilation of rule '@glfw//:linux_joystick' failed (Exit 1) wrapped_clang failed: error executing command external/local_config_cc/wrapped_clang '-D_FORTIFY_SOURCE=1' -fstack-protector -fcolor-diagnostics -Wall -Wthread-safety -Wself-assign -fno-omit-frame-pointer -O0 -DDEBUG -iquote external/glfw -iquote ... (remaining 29 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox
external/glfw/src/linux_joystick.c:34:10: fatal error: 'sys/inotify.h' file not found
#include <sys/inotify.h>
^~~~~~~~~~~~~~~
1 error generated.
Target //test:test failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 8.039s, Critical Path: 4.03s
INFO: 14 processes: 5 internal, 9 darwin-sandbox.
FAILED: Build did NOT complete successfully
```
So my problem is , **how to import `<sys/inotify.h>`**
### Feature requests: what underlying problem are you trying to solve with this feature?
import glfw library by bazel.First of all , I solve the problem `fatal error : X11/Xlib.h not found`.But this problem occurs randomly
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
in my `third_party/glfw`, the docs structure is :
```
.
├── BUILD
├── glfw.BUILD
└── workspace.bzl
```
in my `workspace.bzl`:
```
def clean_dep(dep):
return str(Label(dep))
def repo():
//download glfw to local
native.new_local_repository(
name = "glfw",
path = "/Users/core/Documents/Github/glfw",
build_file = clean_dep("//third_party/glfw:glfw.BUILD"),
)
```
in my glfw.BUILD :
```
package(
default_visibility = ["//visibility:public"],
)
DEFINES = [
"_GLFW_X11",
"_GLFW_HAS_XF86VM",
]
cc_library(
name = "includes",
hdrs = glob([
"include/GLFW/*.h",
"src/*.h",
]),
)
cc_library(
name = "context",
srcs = ["src/context.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "init",
srcs = ["src/init.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "input",
srcs = ["src/input.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "monitor",
srcs = ["src/monitor.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "vulkan",
srcs = ["src/vulkan.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "window",
srcs = ["src/window.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_init",
srcs = ["src/x11_init.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_monitor",
srcs = ["src/x11_monitor.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "x11_window",
srcs = ["src/x11_window.c"],
defines = DEFINES,
deps = [
":egl_context",
":glx_context",
":includes",
"@x11",
],
)
cc_library(
name = "xkb_unicode",
srcs = ["src/xkb_unicode.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "linux_joystick",
srcs = ["src/linux_joystick.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "posix_time",
srcs = ["src/posix_time.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "glx_context",
srcs = ["src/glx_context.c"],
defines = DEFINES,
deps = [
":context",
":includes",
# ":posix_tls",
"@x11",
],
)
cc_library(
name = "egl_context",
srcs = ["src/egl_context.c"],
defines = DEFINES,
deps = [
":includes",
"@x11",
],
)
cc_library(
name = "glfw",
hdrs = glob(["include/GLFW/*.h"]),
defines = DEFINES,
strip_include_prefix = "include",
visibility = ["//visibility:public"],
deps = [
":context",
":egl_context",
":glx_context",
":init",
":input",
":linux_joystick",
":monitor",
":posix_time",
#":posix_tls",
":vulkan",
":window",
":x11_init",
":x11_monitor",
":x11_window",
":xkb_unicode",
],
)
```
### What operating system are you running Bazel on?
Macos10.14.6
### What's the output of `bazel info release`?
release 3.6.0-homebrew
|
process
|
fatal error sys inotify h file not found description of the problem feature request when i run bazel build test test sometimes i got the error info info analyzed target test test packages loaded targets configured info found target error private var tmp bazel core external glfw build bazel c compilation of rule glfw linux joystick failed exit wrapped clang failed error executing command external local config cc wrapped clang d fortify source fstack protector fcolor diagnostics wall wthread safety wself assign fno omit frame pointer ddebug iquote external glfw iquote remaining argument s skipped use sandbox debug to see verbose messages from the sandbox external glfw src linux joystick c fatal error sys inotify h file not found include error generated target test test failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal darwin sandbox failed build did not complete successfully so my problem is how to import feature requests what underlying problem are you trying to solve with this feature import glfw library by bazel first of all i solve the problem fatal error xlib h not found but this problem occurs randomly bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible in my third party glfw the docs structure is ├── build ├── glfw build └── workspace bzl in my workspace bzl def clean dep dep return str label dep def repo download glfw to local native new local repository name glfw path users core documents github glfw build file clean dep third party glfw glfw build in my glfw build package default visibility defines glfw glfw has cc library name includes hdrs glob include glfw h src h cc library name context srcs defines defines deps includes cc library name init srcs defines defines deps includes cc library name input srcs defines defines deps includes cc library name monitor srcs defines defines deps includes cc library name vulkan srcs defines defines deps includes cc library name window srcs defines defines deps includes cc library name init srcs defines defines deps includes cc library name monitor srcs defines defines deps includes cc library name window srcs defines defines deps egl context glx context includes cc library name xkb unicode srcs defines defines deps includes cc library name linux joystick srcs defines defines deps includes cc library name posix time srcs defines defines deps includes cc library name glx context srcs defines defines deps context includes posix tls cc library name egl context srcs defines defines deps includes cc library name glfw hdrs glob defines defines strip include prefix include visibility deps context egl context glx context init input linux joystick monitor posix time posix tls vulkan window init monitor window xkb unicode what operating system are you running bazel on what s the output of bazel info release release homebrew
| 1
|
7,318
| 10,452,928,460
|
IssuesEvent
|
2019-09-19 15:33:15
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
opened
|
Bug: Experience dates display incorrectly on application review page
|
Apply Process Bug State Dept.
|
Environment: Production
Issue: Work experience start and end dates are displaying incorrectly on the application review page
Steps to reproduce.
Related ticket:
|
1.0
|
Bug: Experience dates display incorrectly on application review page - Environment: Production
Issue: Work experience start and end dates are displaying incorrectly on the application review page
Steps to reproduce.
Related ticket:
|
process
|
bug experience dates display incorrectly on application review page environment production issue work experience start and end dates are displaying incorrectly on the application review page steps to reproduce related ticket
| 1
|
17,276
| 23,065,817,808
|
IssuesEvent
|
2022-07-25 13:50:59
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
Need help writing scrip 9s12
|
Type: Bug Feature: Processor/HCS12 Status: Internal
|
Ok -- getting to know Ghidra, (old IDA user). I have an issue with the CPU that I am working with (9s12) where allot of my the references to ram addresses were disassembled pointing to
the wrong region of memory (CPU has a paged window from 0x8000->0xB000 that code can run in, and that code maps to physical addresses anywhere from FF:8000 -> FFB000 to E1:8000 -> E1:B000) The mapped in page is set by the ppgae register - but all ram accesses (regardless of the ppage) are always at 0x2000 -> 0x4000. Ghidra made the assumption (probably something I did wrong) that all ram accesses have the page pre-pended to the address. So, for example, if a branch is made to 0x FE:8000 and at that location there is a LDAA 0x2000 it creates that load reference as FE2000 -- which is wrong.
for example:
784470 7d 29 86 STY offset DAT_0fe986
should really be:
784470 7d 29 86 STY DAT_002986
I can do this manually, by hitting the R key, and replacing the
FE with the value 2, but i really want to write a script for that.
I want to write a script to examine all of the references, and fix them up. can someone get me started in the right direction?
|
1.0
|
Need help writing scrip 9s12 - Ok -- getting to know Ghidra, (old IDA user). I have an issue with the CPU that I am working with (9s12) where allot of my the references to ram addresses were disassembled pointing to
the wrong region of memory (CPU has a paged window from 0x8000->0xB000 that code can run in, and that code maps to physical addresses anywhere from FF:8000 -> FFB000 to E1:8000 -> E1:B000) The mapped in page is set by the ppgae register - but all ram accesses (regardless of the ppage) are always at 0x2000 -> 0x4000. Ghidra made the assumption (probably something I did wrong) that all ram accesses have the page pre-pended to the address. So, for example, if a branch is made to 0x FE:8000 and at that location there is a LDAA 0x2000 it creates that load reference as FE2000 -- which is wrong.
for example:
784470 7d 29 86 STY offset DAT_0fe986
should really be:
784470 7d 29 86 STY DAT_002986
I can do this manually, by hitting the R key, and replacing the
FE with the value 2, but i really want to write a script for that.
I want to write a script to examine all of the references, and fix them up. can someone get me started in the right direction?
|
process
|
need help writing scrip ok getting to know ghidra old ida user i have an issue with the cpu that i am working with where allot of my the references to ram addresses were disassembled pointing to the wrong region of memory cpu has a paged window from that code can run in and that code maps to physical addresses anywhere from ff to the mapped in page is set by the ppgae register but all ram accesses regardless of the ppage are always at ghidra made the assumption probably something i did wrong that all ram accesses have the page pre pended to the address so for example if a branch is made to fe and at that location there is a ldaa it creates that load reference as which is wrong for example sty offset dat should really be sty dat i can do this manually by hitting the r key and replacing the fe with the value but i really want to write a script for that i want to write a script to examine all of the references and fix them up can someone get me started in the right direction
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.