Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
65,990
| 7,948,303,982
|
IssuesEvent
|
2018-07-11 07:39:00
|
unee-t/frontend
|
https://api.github.com/repos/unee-t/frontend
|
closed
|
No possibility to create a case for a unit which has no case in the list (Ex: empty list of cases)
|
critical design/ux enhancement production
|
# The problem:
If a user has no open cases, he/she will see an empty page with no clear call to action

@kiatlim can you think of a solution to solve that?
We could send the user to a different page (the 'Unit' page maybe?) if there is no open cases

Or maybe we can bring back the 'Add Case' button if there are no open cases 🤔
|
1.0
|
No possibility to create a case for a unit which has no case in the list (Ex: empty list of cases) - # The problem:
If a user has no open cases, he/she will see an empty page with no clear call to action

@kiatlim can you think of a solution to solve that?
We could send the user to a different page (the 'Unit' page maybe?) if there is no open cases

Or maybe we can bring back the 'Add Case' button if there are no open cases 🤔
|
non_test
|
no possibility to create a case for a unit which has no case in the list ex empty list of cases the problem if a user has no open cases he she will see an empty page with no clear call to action kiatlim can you think of a solution to solve that we could send the user to a different page the unit page maybe if there is no open cases or maybe we can bring back the add case button if there are no open cases 🤔
| 0
|
160,889
| 25,250,472,808
|
IssuesEvent
|
2022-11-15 14:20:41
|
Sage/carbon
|
https://api.github.com/repos/Sage/carbon
|
closed
|
Change background color for FlatTableCell
|
Enhancement triage Design System Review Required flat-table
|
### Desired behaviour
FlatTableCell should have prop background-color
### Current behaviour
FlatTableCell don't have prop background-color
### Suggested Solution
_No response_
### CodeSandbox or Storybook URL
_No response_
### Anything else we should know?
_No response_
### Confidentiality
- [X] I confirm there is no confidential or commercially sensitive information included.
|
1.0
|
Change background color for FlatTableCell - ### Desired behaviour
FlatTableCell should have prop background-color
### Current behaviour
FlatTableCell don't have prop background-color
### Suggested Solution
_No response_
### CodeSandbox or Storybook URL
_No response_
### Anything else we should know?
_No response_
### Confidentiality
- [X] I confirm there is no confidential or commercially sensitive information included.
|
non_test
|
change background color for flattablecell desired behaviour flattablecell should have prop background color current behaviour flattablecell don t have prop background color suggested solution no response codesandbox or storybook url no response anything else we should know no response confidentiality i confirm there is no confidential or commercially sensitive information included
| 0
|
19,947
| 10,564,176,528
|
IssuesEvent
|
2019-10-04 23:55:29
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Flutter SSL Memory Leaks
|
customer: gold dependency: dart engine severe: performance
|
## Steps to Reproduce
1. Start Flutter app.
2. Push several pages the Navigator
3. Pop several pages.
4. In Xcode use Debug Memory Graph to capture memory graph
5. In memory filter the leaks.
## leaks
<img width="1341" alt="ssl_leaks" src="https://user-images.githubusercontent.com/2551915/43937206-ce03e7ea-9c8e-11e8-92e9-72cdbf89ff67.png">
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
```
<!-- Finally, paste the output of running `flutter doctor -v` here. -->
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel xy_beta_v0.5.6, v0.5.6-pre.112, on Mac OS X 10.13.5 17F77, locale zh-Hans-CN)
[✓] Android toolchain - develop for Android devices (Android SDK 27.0.3)
[✓] iOS toolchain - develop for iOS devices (Xcode 9.4.1)
[✓] Android Studio (version 3.1)
[✓] VS Code (version 1.25.1)
[✓] Connected devices (1 available)
• No issues found!
```
|
True
|
Flutter SSL Memory Leaks -
## Steps to Reproduce
1. Start Flutter app.
2. Push several pages the Navigator
3. Pop several pages.
4. In Xcode use Debug Memory Graph to capture memory graph
5. In memory filter the leaks.
## leaks
<img width="1341" alt="ssl_leaks" src="https://user-images.githubusercontent.com/2551915/43937206-ce03e7ea-9c8e-11e8-92e9-72cdbf89ff67.png">
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
```
<!-- Finally, paste the output of running `flutter doctor -v` here. -->
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel xy_beta_v0.5.6, v0.5.6-pre.112, on Mac OS X 10.13.5 17F77, locale zh-Hans-CN)
[✓] Android toolchain - develop for Android devices (Android SDK 27.0.3)
[✓] iOS toolchain - develop for iOS devices (Xcode 9.4.1)
[✓] Android Studio (version 3.1)
[✓] VS Code (version 1.25.1)
[✓] Connected devices (1 available)
• No issues found!
```
|
non_test
|
flutter ssl memory leaks steps to reproduce start flutter app push several pages the navigator pop several pages in xcode use debug memory graph to capture memory graph in memory filter the leaks leaks img width alt ssl leaks src run flutter analyze and attach any output of that command below if there are any analysis errors try resolving them before filing this issue doctor summary to see all details run flutter doctor v flutter channel xy beta pre on mac os x locale zh hans cn android toolchain develop for android devices android sdk ios toolchain develop for ios devices xcode android studio version vs code version connected devices available • no issues found
| 0
|
263,119
| 23,036,618,286
|
IssuesEvent
|
2022-07-22 19:34:19
|
OvercastCommunity/public-competitive
|
https://api.github.com/repos/OvercastCommunity/public-competitive
|
closed
|
5CP - 5v5 - Backstreets
|
contest
|
### Checklist
Check what applies to you. *Add an X between the brackets or click the checkboxes when you have submitted the issue.*
- [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map
- [X] I have agreed with assigning the [CC BY-SA 4.0 license](https://creativecommons.org/licenses/by-sa/4.0/) to this map
- [X] I have provided an XML file
- [X] I have uploaded the map zip file to a file sharing service
# Backstreets
## 5CP
## XML
https://gist.github.com/Tywnis/decb996a192fc57b45b2c1bba7ecc8da

[Backstreets.zip](https://github.com/OvercastCommunity/public-competitive/files/8504182/Backstreets.zip)
|
1.0
|
5CP - 5v5 - Backstreets - ### Checklist
Check what applies to you. *Add an X between the brackets or click the checkboxes when you have submitted the issue.*
- [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map
- [X] I have agreed with assigning the [CC BY-SA 4.0 license](https://creativecommons.org/licenses/by-sa/4.0/) to this map
- [X] I have provided an XML file
- [X] I have uploaded the map zip file to a file sharing service
# Backstreets
## 5CP
## XML
https://gist.github.com/Tywnis/decb996a192fc57b45b2c1bba7ecc8da

[Backstreets.zip](https://github.com/OvercastCommunity/public-competitive/files/8504182/Backstreets.zip)
|
test
|
backstreets checklist check what applies to you add an x between the brackets or click the checkboxes when you have submitted the issue i have the map i have agreed with assigning the to this map i have provided an xml file i have uploaded the map zip file to a file sharing service backstreets xml
| 1
|
109,062
| 16,827,998,683
|
IssuesEvent
|
2021-06-17 21:32:11
|
kevins01/Java3
|
https://api.github.com/repos/kevins01/Java3
|
opened
|
CVE-2020-8840 (High) detected in jackson-databind-2.8.8.jar
|
security vulnerability
|
## CVE-2020-8840 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Java3/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kevins01/Java3/commit/8398785463cdebbe0e5c0e503e89f594f59554d5">8398785463cdebbe0e5c0e503e89f594f59554d5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.0.0 through 2.9.10.2 lacks certain xbean-reflect/JNDI blocking, as demonstrated by org.apache.xbean.propertyeditor.JndiConverter.
<p>Publish Date: 2020-02-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8840>CVE-2020-8840</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2620">https://github.com/FasterXML/jackson-databind/issues/2620</a></p>
<p>Release Date: 2020-02-10</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8840 (High) detected in jackson-databind-2.8.8.jar - ## CVE-2020-8840 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Java3/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kevins01/Java3/commit/8398785463cdebbe0e5c0e503e89f594f59554d5">8398785463cdebbe0e5c0e503e89f594f59554d5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.0.0 through 2.9.10.2 lacks certain xbean-reflect/JNDI blocking, as demonstrated by org.apache.xbean.propertyeditor.JndiConverter.
<p>Publish Date: 2020-02-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8840>CVE-2020-8840</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2620">https://github.com/FasterXML/jackson-databind/issues/2620</a></p>
<p>Release Date: 2020-02-10</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind through lacks certain xbean reflect jndi blocking as demonstrated by org apache xbean propertyeditor jndiconverter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
| 0
|
78,798
| 9,796,465,630
|
IssuesEvent
|
2019-06-11 07:42:38
|
wazuh/wazuh-kibana-app
|
https://api.github.com/repos/wazuh/wazuh-kibana-app
|
closed
|
Adapt menu directive design to Kibana 7.x new style
|
UI/UX design frontend
|
Hi team,
**Description**
Since Kibana 7, our menu directive looks weird, the design breaks all the views.
Here is how it looks with the default theme:

And here is how it looks with the dark theme:

**Proposal**
Adapt the directive to a new fresh design. It must include both light and dark themes.
Something like the Kibana menu would work:

Regards
|
1.0
|
Adapt menu directive design to Kibana 7.x new style - Hi team,
**Description**
Since Kibana 7, our menu directive looks weird, the design breaks all the views.
Here is how it looks with the default theme:

And here is how it looks with the dark theme:

**Proposal**
Adapt the directive to a new fresh design. It must include both light and dark themes.
Something like the Kibana menu would work:

Regards
|
non_test
|
adapt menu directive design to kibana x new style hi team description since kibana our menu directive looks weird the design breaks all the views here is how it looks with the default theme and here is how it looks with the dark theme proposal adapt the directive to a new fresh design it must include both light and dark themes something like the kibana menu would work regards
| 0
|
504,551
| 14,620,187,935
|
IssuesEvent
|
2020-12-22 19:14:50
|
googleapis/elixir-google-api
|
https://api.github.com/repos/googleapis/elixir-google-api
|
closed
|
Synthesis failed for CloudBilling
|
api: cloudbilling autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate CloudBilling. :broken_heart:
Here's the output from running `synth.py`:
```
rkspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh CloudBilling
DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmpg5v2hc16/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh CloudBilling
/workspace /workspace
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
Resolving Hex dependencies...
Dependency resolution completed:
Unchanged:
certifi 2.5.1
google_api_discovery 0.7.0
google_gax 0.3.2
hackney 1.15.2
idna 6.0.0
jason 1.2.1
metrics 1.0.1
mime 1.3.1
mimerl 1.2.0
oauth2 0.9.4
parse_trans 3.3.0
poison 3.1.0
ssl_verify_fun 1.1.5
temp 0.4.7
tesla 1.3.3
unicode_util_compat 0.4.1
* Getting google_api_discovery (Hex package)
* Getting tesla (Hex package)
* Getting oauth2 (Hex package)
* Getting temp (Hex package)
* Getting jason (Hex package)
* Getting poison (Hex package)
* Getting hackney (Hex package)
* Getting certifi (Hex package)
* Getting idna (Hex package)
* Getting metrics (Hex package)
* Getting mimerl (Hex package)
* Getting ssl_verify_fun (Hex package)
* Getting unicode_util_compat (Hex package)
* Getting parse_trans (Hex package)
* Getting mime (Hex package)
* Getting google_gax (Hex package)
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
==> temp
Compiling 3 files (.ex)
Generated temp app
===> Compiling parse_trans
===> Compiling mimerl
===> Compiling metrics
===> Compiling unicode_util_compat
===> Compiling idna
==> jason
Compiling 8 files (.ex)
Generated jason app
warning: String.strip/1 is deprecated. Use String.trim/1 instead
/workspace/deps/poison/mix.exs:4
==> poison
Compiling 4 files (.ex)
warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead
lib/poison/encoder.ex:173
Generated poison app
==> ssl_verify_fun
Compiling 7 files (.erl)
Generated ssl_verify_fun app
===> Compiling certifi
===> Compiling hackney
==> oauth2
Compiling 13 files (.ex)
Generated oauth2 app
==> mime
Compiling 2 files (.ex)
Generated mime app
==> tesla
Compiling 26 files (.ex)
Generated tesla app
==> google_gax
Compiling 5 files (.ex)
Generated google_gax app
==> google_api_discovery
Compiling 21 files (.ex)
Generated google_api_discovery app
==> google_apis
Compiling 28 files (.ex)
Generated google_apis app
13:29:39.840 [info] FETCHING: https://cloudbilling.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v1
13:29:40.163 [info] FETCHING: https://cloudbilling.googleapis.com/$discovery/rest?version=v1
13:29:40.378 [info] FOUND: https://cloudbilling.googleapis.com/$discovery/rest?version=v1
Revision check: old=20201017, new=20201017, generating=true
Creating leading directories
Writing AggregationInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/aggregation_info.ex.
Writing AuditConfig to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/audit_config.ex.
Writing AuditLogConfig to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/audit_log_config.ex.
Writing BillingAccount to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/billing_account.ex.
Writing Binding to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/binding.ex.
Writing Category to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/category.ex.
Writing Expr to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/expr.ex.
Writing GeoTaxonomy to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/geo_taxonomy.ex.
Writing ListBillingAccountsResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_billing_accounts_response.ex.
Writing ListProjectBillingInfoResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_project_billing_info_response.ex.
Writing ListServicesResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_services_response.ex.
Writing ListSkusResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_skus_response.ex.
Writing Money to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/money.ex.
Writing Policy to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/policy.ex.
Writing PricingExpression to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/pricing_expression.ex.
Writing PricingInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/pricing_info.ex.
Writing ProjectBillingInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/project_billing_info.ex.
Writing Service to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/service.ex.
Writing SetIamPolicyRequest to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/set_iam_policy_request.ex.
Writing Sku to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/sku.ex.
Writing TestIamPermissionsRequest to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/test_iam_permissions_request.ex.
Writing TestIamPermissionsResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/test_iam_permissions_response.ex.
Writing TierRate to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/tier_rate.ex.
Writing BillingAccounts to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/billing_accounts.ex.
Writing Projects to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/projects.ex.
Writing Services to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/services.ex.
Writing connection.ex.
Writing metadata.ex.
Writing mix.exs
Writing README.md
Writing LICENSE
Writing .gitignore
Writing config/config.exs
Writing test/test_helper.exs
13:29:40.816 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR.
fixing file permissions
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 252, in __exit__
self.observer.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 361, in on_thread_stop
self.unschedule_all()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 357, in unschedule_all
self._clear_emitters()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 231, in _clear_emitters
emitter.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify.py", line 121, in on_thread_stop
self._inotify.close()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 50, in close
self.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 46, in on_thread_stop
self._inotify.close()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_c.py", line 277, in close
os.close(self._inotify_fd)
OSError: [Errno 9] Bad file descriptor
2020-12-18 05:29:43,909 autosynth [ERROR] > Synthesis failed
2020-12-18 05:29:43,909 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 291, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/cloud_billing/synth.metadata', 'synth.py', '--', 'CloudBilling']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/3a455424-7540-46a9-bb77-8265c8f04c06/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
|
1.0
|
Synthesis failed for CloudBilling - Hello! Autosynth couldn't regenerate CloudBilling. :broken_heart:
Here's the output from running `synth.py`:
```
rkspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh CloudBilling
DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmpg5v2hc16/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh CloudBilling
/workspace /workspace
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
Resolving Hex dependencies...
Dependency resolution completed:
Unchanged:
certifi 2.5.1
google_api_discovery 0.7.0
google_gax 0.3.2
hackney 1.15.2
idna 6.0.0
jason 1.2.1
metrics 1.0.1
mime 1.3.1
mimerl 1.2.0
oauth2 0.9.4
parse_trans 3.3.0
poison 3.1.0
ssl_verify_fun 1.1.5
temp 0.4.7
tesla 1.3.3
unicode_util_compat 0.4.1
* Getting google_api_discovery (Hex package)
* Getting tesla (Hex package)
* Getting oauth2 (Hex package)
* Getting temp (Hex package)
* Getting jason (Hex package)
* Getting poison (Hex package)
* Getting hackney (Hex package)
* Getting certifi (Hex package)
* Getting idna (Hex package)
* Getting metrics (Hex package)
* Getting mimerl (Hex package)
* Getting ssl_verify_fun (Hex package)
* Getting unicode_util_compat (Hex package)
* Getting parse_trans (Hex package)
* Getting mime (Hex package)
* Getting google_gax (Hex package)
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
==> temp
Compiling 3 files (.ex)
Generated temp app
===> Compiling parse_trans
===> Compiling mimerl
===> Compiling metrics
===> Compiling unicode_util_compat
===> Compiling idna
==> jason
Compiling 8 files (.ex)
Generated jason app
warning: String.strip/1 is deprecated. Use String.trim/1 instead
/workspace/deps/poison/mix.exs:4
==> poison
Compiling 4 files (.ex)
warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead
lib/poison/encoder.ex:173
Generated poison app
==> ssl_verify_fun
Compiling 7 files (.erl)
Generated ssl_verify_fun app
===> Compiling certifi
===> Compiling hackney
==> oauth2
Compiling 13 files (.ex)
Generated oauth2 app
==> mime
Compiling 2 files (.ex)
Generated mime app
==> tesla
Compiling 26 files (.ex)
Generated tesla app
==> google_gax
Compiling 5 files (.ex)
Generated google_gax app
==> google_api_discovery
Compiling 21 files (.ex)
Generated google_api_discovery app
==> google_apis
Compiling 28 files (.ex)
Generated google_apis app
13:29:39.840 [info] FETCHING: https://cloudbilling.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v1
13:29:40.163 [info] FETCHING: https://cloudbilling.googleapis.com/$discovery/rest?version=v1
13:29:40.378 [info] FOUND: https://cloudbilling.googleapis.com/$discovery/rest?version=v1
Revision check: old=20201017, new=20201017, generating=true
Creating leading directories
Writing AggregationInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/aggregation_info.ex.
Writing AuditConfig to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/audit_config.ex.
Writing AuditLogConfig to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/audit_log_config.ex.
Writing BillingAccount to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/billing_account.ex.
Writing Binding to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/binding.ex.
Writing Category to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/category.ex.
Writing Expr to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/expr.ex.
Writing GeoTaxonomy to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/geo_taxonomy.ex.
Writing ListBillingAccountsResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_billing_accounts_response.ex.
Writing ListProjectBillingInfoResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_project_billing_info_response.ex.
Writing ListServicesResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_services_response.ex.
Writing ListSkusResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/list_skus_response.ex.
Writing Money to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/money.ex.
Writing Policy to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/policy.ex.
Writing PricingExpression to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/pricing_expression.ex.
Writing PricingInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/pricing_info.ex.
Writing ProjectBillingInfo to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/project_billing_info.ex.
Writing Service to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/service.ex.
Writing SetIamPolicyRequest to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/set_iam_policy_request.ex.
Writing Sku to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/sku.ex.
Writing TestIamPermissionsRequest to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/test_iam_permissions_request.ex.
Writing TestIamPermissionsResponse to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/test_iam_permissions_response.ex.
Writing TierRate to clients/cloud_billing/lib/google_api/cloud_billing/v1/model/tier_rate.ex.
Writing BillingAccounts to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/billing_accounts.ex.
Writing Projects to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/projects.ex.
Writing Services to clients/cloud_billing/lib/google_api/cloud_billing/v1/api/services.ex.
Writing connection.ex.
Writing metadata.ex.
Writing mix.exs
Writing README.md
Writing LICENSE
Writing .gitignore
Writing config/config.exs
Writing test/test_helper.exs
13:29:40.816 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR.
fixing file permissions
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 252, in __exit__
self.observer.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 361, in on_thread_stop
self.unschedule_all()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 357, in unschedule_all
self._clear_emitters()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 231, in _clear_emitters
emitter.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify.py", line 121, in on_thread_stop
self._inotify.close()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 50, in close
self.stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop
self.on_thread_stop()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 46, in on_thread_stop
self._inotify.close()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_c.py", line 277, in close
os.close(self._inotify_fd)
OSError: [Errno 9] Bad file descriptor
2020-12-18 05:29:43,909 autosynth [ERROR] > Synthesis failed
2020-12-18 05:29:43,909 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 291, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/cloud_billing/synth.metadata', 'synth.py', '--', 'CloudBilling']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/3a455424-7540-46a9-bb77-8265c8f04c06/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
|
non_test
|
synthesis failed for cloudbilling hello autosynth couldn t regenerate cloudbilling broken heart here s the output from running synth py rkspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh cloudbilling debug synthtool running docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh cloudbilling workspace workspace mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data resolving hex dependencies dependency resolution completed unchanged certifi google api discovery google gax hackney idna jason metrics mime mimerl parse trans poison ssl verify fun temp tesla unicode util compat getting google api discovery hex package getting tesla hex package getting hex package getting temp hex package getting jason hex package getting poison hex package getting hackney hex package getting certifi hex package getting idna hex package getting metrics hex package getting mimerl hex package getting ssl verify fun hex package getting unicode util compat hex package getting parse trans hex package getting mime hex package getting google gax hex package mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data temp compiling files ex generated temp app compiling parse trans compiling mimerl compiling metrics compiling unicode util compat compiling idna jason compiling files ex generated jason app warning string strip is deprecated use string trim instead workspace deps poison mix exs poison compiling files ex warning integer to char list is deprecated use integer to charlist instead lib poison encoder ex generated poison app ssl verify fun compiling files erl generated ssl verify fun app compiling certifi compiling hackney compiling files ex generated app mime compiling files ex generated mime app tesla compiling files ex generated tesla app google gax compiling files ex generated google gax app google api discovery compiling files ex generated google api discovery app google apis compiling files ex generated google apis app fetching fetching found revision check old new generating true creating leading directories writing aggregationinfo to clients cloud billing lib google api cloud billing model aggregation info ex writing auditconfig to clients cloud billing lib google api cloud billing model audit config ex writing auditlogconfig to clients cloud billing lib google api cloud billing model audit log config ex writing billingaccount to clients cloud billing lib google api cloud billing model billing account ex writing binding to clients cloud billing lib google api cloud billing model binding ex writing category to clients cloud billing lib google api cloud billing model category ex writing expr to clients cloud billing lib google api cloud billing model expr ex writing geotaxonomy to clients cloud billing lib google api cloud billing model geo taxonomy ex writing listbillingaccountsresponse to clients cloud billing lib google api cloud billing model list billing accounts response ex writing listprojectbillinginforesponse to clients cloud billing lib google api cloud billing model list project billing info response ex writing listservicesresponse to clients cloud billing lib google api cloud billing model list services response ex writing listskusresponse to clients cloud billing lib google api cloud billing model list skus response ex writing money to clients cloud billing lib google api cloud billing model money ex writing policy to clients cloud billing lib google api cloud billing model policy ex writing pricingexpression to clients cloud billing lib google api cloud billing model pricing expression ex writing pricinginfo to clients cloud billing lib google api cloud billing model pricing info ex writing projectbillinginfo to clients cloud billing lib google api cloud billing model project billing info ex writing service to clients cloud billing lib google api cloud billing model service ex writing setiampolicyrequest to clients cloud billing lib google api cloud billing model set iam policy request ex writing sku to clients cloud billing lib google api cloud billing model sku ex writing testiampermissionsrequest to clients cloud billing lib google api cloud billing model test iam permissions request ex writing testiampermissionsresponse to clients cloud billing lib google api cloud billing model test iam permissions response ex writing tierrate to clients cloud billing lib google api cloud billing model tier rate ex writing billingaccounts to clients cloud billing lib google api cloud billing api billing accounts ex writing projects to clients cloud billing lib google api cloud billing api projects ex writing services to clients cloud billing lib google api cloud billing api services ex writing connection ex writing metadata ex writing mix exs writing readme md writing license writing gitignore writing config config exs writing test test helper exs found only discovery revision and or formatting changes not significant enough for a pr fixing file permissions traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit self observer stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers api py line in on thread stop self unschedule all file tmpfs src github synthtool env lib site packages watchdog observers api py line in unschedule all self clear emitters file tmpfs src github synthtool env lib site packages watchdog observers api py line in clear emitters emitter stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers inotify py line in on thread stop self inotify close file tmpfs src github synthtool env lib site packages watchdog observers inotify buffer py line in close self stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers inotify buffer py line in on thread stop self inotify close file tmpfs src github synthtool env lib site packages watchdog observers inotify c py line in close os close self inotify fd oserror bad file descriptor autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
278,211
| 24,133,996,424
|
IssuesEvent
|
2022-09-21 09:44:12
|
Gilson-Dourado10/challenge_mentoring-_spdv
|
https://api.github.com/repos/Gilson-Dourado10/challenge_mentoring-_spdv
|
opened
|
Testar Peças
|
test
|
- [ ] Eu COMO administrador do sistema, DESEJO testar a tabela peças, PARA verificar se o sistema de dados está coberto com testes automatizados.
- [ ] testar model
- [ ] testar controller
|
1.0
|
Testar Peças - - [ ] Eu COMO administrador do sistema, DESEJO testar a tabela peças, PARA verificar se o sistema de dados está coberto com testes automatizados.
- [ ] testar model
- [ ] testar controller
|
test
|
testar peças eu como administrador do sistema desejo testar a tabela peças para verificar se o sistema de dados está coberto com testes automatizados testar model testar controller
| 1
|
80,169
| 10,154,757,302
|
IssuesEvent
|
2019-08-06 08:47:08
|
mantidproject/mantid
|
https://api.github.com/repos/mantidproject/mantid
|
closed
|
Fix C++ Exercise documentation
|
Documentation
|
### Expected behavior
- EX01 explains what is expected of hyphenated words.
- EX02 is consistent with its use of area.
### Actual behavior
- Unclear and inconsistent.
### Steps to reproduce the behavior
1. Visit [page](https://www.mantidproject.org/New_Starter_C%2B%2B_introduction)
### Platforms affected
All
|
1.0
|
Fix C++ Exercise documentation - ### Expected behavior
- EX01 explains what is expected of hyphenated words.
- EX02 is consistent with its use of area.
### Actual behavior
- Unclear and inconsistent.
### Steps to reproduce the behavior
1. Visit [page](https://www.mantidproject.org/New_Starter_C%2B%2B_introduction)
### Platforms affected
All
|
non_test
|
fix c exercise documentation expected behavior explains what is expected of hyphenated words is consistent with its use of area actual behavior unclear and inconsistent steps to reproduce the behavior visit platforms affected all
| 0
|
92,547
| 8,367,598,446
|
IssuesEvent
|
2018-10-04 12:43:48
|
ValveSoftware/csgo-osx-linux
|
https://api.github.com/repos/ValveSoftware/csgo-osx-linux
|
closed
|
Multicore rendering not working with Linux
|
Linux Need Retest
|
#### Your system information
Informations de l'ordinateur :
Fabricant : Unknown
Modèle : Unknown
Type : Ordinateur portable
Aucun écran tactile détecté
Processeur :
Fabricant du CPU : GenuineIntel
Marque du processeur : Intel(R) Core(TM) i3-3217U CPU @ 1.80GHz
Famille du processeur : 0x6
Modèle du processeur : 0x3a
Révision processeur : 0x9
Type de processeur : 0x0
Fréquence : 1800 Mhz
4 processeurs logiques
2 processeurs physiques
HyperThreading : Pris en charge
FCMOV : Pris en charge
SSE2 : Pris en charge
SSE3 : Pris en charge
SSSE3 : Pris en charge
SSE4a : Non pris en charge
SSE41 : Pris en charge
SSE42 : Pris en charge
AES : Non pris en charge
AVX : Pris en charge
CMPXCHG16B : Pris en charge
LAHF/SAHF : Pris en charge
PrefetchW : Non pris en charge
Version du système d'exploitation :
Debian GNU/Linux testing (buster) (64 bits)
Nom du noyau : Linux
Version du noyau : 4.17.0-3-amd64
Éditeur du serveur X : The X.Org Foundation
Version du serveur X : 12001000
Gestionnaire de fenêtres X : GNOME Shell
Version d'exécution Steam : steam-runtime-beta-release_2018-06-14
Carte graphique :
Pilote : Intel Open Source Technology Center Mesa DRI Intel(R) Ivybridge Mobile x86/MMX/SSE2
Version du pilote : 3.0 Mesa 18.1.7
Version OpenGL : 3.0
Profondeur de couleur du bureau : 24 bits par pixel
Taux de rafraîchissement du moniteur : 59 Hz
ID fabricant : 0x8086
ID périphérique : 0x166
Révision non détectée
Nombre de moniteurs : 1
Nombre de cartes vidéo logiques : 1
Résolution de l'affichage principal : 1366 x 768
Résolution du bureau : 1366 x 768
Taille de l'affichage principal : 13.78" x 7.48" (15.67" diag)
35.0cm x 19.0cm (39.8cm diag)
Mémoire vidéo principale non détecté
Carte son :
Périphérique audio : Realtek ALC270
Mémoire :
RAM : 7426 Mo
Divers :
Langue de l'interface : Français
LANGUE : fr_FR.UTF-8
Espace disque total disponible : 932356 Mo
Plus grand bloc disponible sur le disque dur : 44911 Mo
Casque VR : aucun détecté
Rapports d'échec récent :
* Have you checked for system updates?: [Yes]
#### Please describe your issue in as much detail as possible:
Hi,
First, sorry for my bad english, i'm french.
I have a problem with the game csgo.
One of my core is always at 100% of use.
I tried the option -thread 4, but this dont work.
I also tried with other game like "half-life 2" and it's work perfectly.
I tried with the Intel graphic card, but same problem.
So the problem come from csgo... I tried to desactivate "multicore rendering" in csgo options, but nothing change.
Please help me.
I have the driver Bumblebee because i have two graphic card, one Intel, and one Nvidia.
uname -a
Linux localhost.localdomain 4.17.0-3-amd64 #1 SMP Debian 4.17.17-1 (2018-08-18) x86_64 GNU/Linux
optirun nvidia-settings -c :8 -q GPUPerfModes -t
perf=0, nvclock=162, nvclockmin=162, nvclockmax=270, nvclockeditable=0,
memclock=405, memclockmin=405, memclockmax=405, memclockeditable=0,
memTransferRate=810, memTransferRatemin=810, memTransferRatemax=810,
memTransferRateeditable=0, processorclock=324, processorclockmin=324,
processorclockmax=540, processorclockeditable=0 ; perf=1, nvclock=162,
nvclockmin=162, nvclockmax=800, nvclockeditable=0, memclock=900,
memclockmin=900, memclockmax=900, memclockeditable=0, memTransferRate=1800,
memTransferRatemin=1800, memTransferRatemax=1800, memTransferRateeditable=0,
processorclock=324, processorclockmin=324, processorclockmax=1600,
processorclockeditable=0
perf=0, nvclock=162, nvclockmin=162, nvclockmax=270, nvclockeditable=0,
memclock=405, memclockmin=405, memclockmax=405, memclockeditable=0,
memTransferRate=810, memTransferRatemin=810, memTransferRatemax=810,
memTransferRateeditable=0, processorclock=324, processorclockmin=324,
processorclockmax=540, processorclockeditable=0 ; perf=1, nvclock=162,
nvclockmin=162, nvclockmax=800, nvclockeditable=0, memclock=900,
memclockmin=900, memclockmax=900, memclockeditable=0, memTransferRate=1800,
memTransferRatemin=1800, memTransferRatemax=1800, memTransferRateeditable=0,
processorclock=324, processorclockmin=324, processorclockmax=1600,
processorclockeditable=0
lspci -k | egrep -iA2 'vga|3d'
00:02.0 VGA compatible controller: Intel Corporation 3rd Gen Core processor Graphics Controller (rev 09)
Subsystem: ASUSTeK Computer Inc. 3rd Gen Core processor Graphics Controller
Kernel driver in use: i915
--
01:00.0 3D controller: NVIDIA Corporation GF117M [GeForce 610M/710M/810M/820M / GT 620M/625M/630M/720M] (rev ff)
Kernel modules: nvidia
|
1.0
|
Multicore rendering not working with Linux - #### Your system information
Informations de l'ordinateur :
Fabricant : Unknown
Modèle : Unknown
Type : Ordinateur portable
Aucun écran tactile détecté
Processeur :
Fabricant du CPU : GenuineIntel
Marque du processeur : Intel(R) Core(TM) i3-3217U CPU @ 1.80GHz
Famille du processeur : 0x6
Modèle du processeur : 0x3a
Révision processeur : 0x9
Type de processeur : 0x0
Fréquence : 1800 Mhz
4 processeurs logiques
2 processeurs physiques
HyperThreading : Pris en charge
FCMOV : Pris en charge
SSE2 : Pris en charge
SSE3 : Pris en charge
SSSE3 : Pris en charge
SSE4a : Non pris en charge
SSE41 : Pris en charge
SSE42 : Pris en charge
AES : Non pris en charge
AVX : Pris en charge
CMPXCHG16B : Pris en charge
LAHF/SAHF : Pris en charge
PrefetchW : Non pris en charge
Version du système d'exploitation :
Debian GNU/Linux testing (buster) (64 bits)
Nom du noyau : Linux
Version du noyau : 4.17.0-3-amd64
Éditeur du serveur X : The X.Org Foundation
Version du serveur X : 12001000
Gestionnaire de fenêtres X : GNOME Shell
Version d'exécution Steam : steam-runtime-beta-release_2018-06-14
Carte graphique :
Pilote : Intel Open Source Technology Center Mesa DRI Intel(R) Ivybridge Mobile x86/MMX/SSE2
Version du pilote : 3.0 Mesa 18.1.7
Version OpenGL : 3.0
Profondeur de couleur du bureau : 24 bits par pixel
Taux de rafraîchissement du moniteur : 59 Hz
ID fabricant : 0x8086
ID périphérique : 0x166
Révision non détectée
Nombre de moniteurs : 1
Nombre de cartes vidéo logiques : 1
Résolution de l'affichage principal : 1366 x 768
Résolution du bureau : 1366 x 768
Taille de l'affichage principal : 13.78" x 7.48" (15.67" diag)
35.0cm x 19.0cm (39.8cm diag)
Mémoire vidéo principale non détecté
Carte son :
Périphérique audio : Realtek ALC270
Mémoire :
RAM : 7426 Mo
Divers :
Langue de l'interface : Français
LANGUE : fr_FR.UTF-8
Espace disque total disponible : 932356 Mo
Plus grand bloc disponible sur le disque dur : 44911 Mo
Casque VR : aucun détecté
Rapports d'échec récent :
* Have you checked for system updates?: [Yes]
#### Please describe your issue in as much detail as possible:
Hi,
First, sorry for my bad english, i'm french.
I have a problem with the game csgo.
One of my core is always at 100% of use.
I tried the option -thread 4, but this dont work.
I also tried with other game like "half-life 2" and it's work perfectly.
I tried with the Intel graphic card, but same problem.
So the problem come from csgo... I tried to desactivate "multicore rendering" in csgo options, but nothing change.
Please help me.
I have the driver Bumblebee because i have two graphic card, one Intel, and one Nvidia.
uname -a
Linux localhost.localdomain 4.17.0-3-amd64 #1 SMP Debian 4.17.17-1 (2018-08-18) x86_64 GNU/Linux
optirun nvidia-settings -c :8 -q GPUPerfModes -t
perf=0, nvclock=162, nvclockmin=162, nvclockmax=270, nvclockeditable=0,
memclock=405, memclockmin=405, memclockmax=405, memclockeditable=0,
memTransferRate=810, memTransferRatemin=810, memTransferRatemax=810,
memTransferRateeditable=0, processorclock=324, processorclockmin=324,
processorclockmax=540, processorclockeditable=0 ; perf=1, nvclock=162,
nvclockmin=162, nvclockmax=800, nvclockeditable=0, memclock=900,
memclockmin=900, memclockmax=900, memclockeditable=0, memTransferRate=1800,
memTransferRatemin=1800, memTransferRatemax=1800, memTransferRateeditable=0,
processorclock=324, processorclockmin=324, processorclockmax=1600,
processorclockeditable=0
perf=0, nvclock=162, nvclockmin=162, nvclockmax=270, nvclockeditable=0,
memclock=405, memclockmin=405, memclockmax=405, memclockeditable=0,
memTransferRate=810, memTransferRatemin=810, memTransferRatemax=810,
memTransferRateeditable=0, processorclock=324, processorclockmin=324,
processorclockmax=540, processorclockeditable=0 ; perf=1, nvclock=162,
nvclockmin=162, nvclockmax=800, nvclockeditable=0, memclock=900,
memclockmin=900, memclockmax=900, memclockeditable=0, memTransferRate=1800,
memTransferRatemin=1800, memTransferRatemax=1800, memTransferRateeditable=0,
processorclock=324, processorclockmin=324, processorclockmax=1600,
processorclockeditable=0
lspci -k | egrep -iA2 'vga|3d'
00:02.0 VGA compatible controller: Intel Corporation 3rd Gen Core processor Graphics Controller (rev 09)
Subsystem: ASUSTeK Computer Inc. 3rd Gen Core processor Graphics Controller
Kernel driver in use: i915
--
01:00.0 3D controller: NVIDIA Corporation GF117M [GeForce 610M/710M/810M/820M / GT 620M/625M/630M/720M] (rev ff)
Kernel modules: nvidia
|
test
|
multicore rendering not working with linux your system information informations de l ordinateur fabricant unknown modèle unknown type ordinateur portable aucun écran tactile détecté processeur fabricant du cpu genuineintel marque du processeur intel r core tm cpu famille du processeur modèle du processeur révision processeur type de processeur fréquence mhz processeurs logiques processeurs physiques hyperthreading pris en charge fcmov pris en charge pris en charge pris en charge pris en charge non pris en charge pris en charge pris en charge aes non pris en charge avx pris en charge pris en charge lahf sahf pris en charge prefetchw non pris en charge version du système d exploitation debian gnu linux testing buster bits nom du noyau linux version du noyau éditeur du serveur x the x org foundation version du serveur x gestionnaire de fenêtres x gnome shell version d exécution steam steam runtime beta release carte graphique pilote intel open source technology center mesa dri intel r ivybridge mobile mmx version du pilote mesa version opengl profondeur de couleur du bureau bits par pixel taux de rafraîchissement du moniteur hz id fabricant id périphérique révision non détectée nombre de moniteurs nombre de cartes vidéo logiques résolution de l affichage principal x résolution du bureau x taille de l affichage principal x diag x diag mémoire vidéo principale non détecté carte son périphérique audio realtek mémoire ram mo divers langue de l interface français langue fr fr utf espace disque total disponible mo plus grand bloc disponible sur le disque dur mo casque vr aucun détecté rapports d échec récent have you checked for system updates please describe your issue in as much detail as possible hi first sorry for my bad english i m french i have a problem with the game csgo one of my core is always at of use i tried the option thread but this dont work i also tried with other game like half life and it s work perfectly i tried with the intel graphic card but same problem so the problem come from csgo i tried to desactivate multicore rendering in csgo options but nothing change please help me i have the driver bumblebee because i have two graphic card one intel and one nvidia uname a linux localhost localdomain smp debian gnu linux optirun nvidia settings c q gpuperfmodes t perf nvclock nvclockmin nvclockmax nvclockeditable memclock memclockmin memclockmax memclockeditable memtransferrate memtransferratemin memtransferratemax memtransferrateeditable processorclock processorclockmin processorclockmax processorclockeditable perf nvclock nvclockmin nvclockmax nvclockeditable memclock memclockmin memclockmax memclockeditable memtransferrate memtransferratemin memtransferratemax memtransferrateeditable processorclock processorclockmin processorclockmax processorclockeditable perf nvclock nvclockmin nvclockmax nvclockeditable memclock memclockmin memclockmax memclockeditable memtransferrate memtransferratemin memtransferratemax memtransferrateeditable processorclock processorclockmin processorclockmax processorclockeditable perf nvclock nvclockmin nvclockmax nvclockeditable memclock memclockmin memclockmax memclockeditable memtransferrate memtransferratemin memtransferratemax memtransferrateeditable processorclock processorclockmin processorclockmax processorclockeditable lspci k egrep vga vga compatible controller intel corporation gen core processor graphics controller rev subsystem asustek computer inc gen core processor graphics controller kernel driver in use controller nvidia corporation rev ff kernel modules nvidia
| 1
|
14,714
| 8,676,256,816
|
IssuesEvent
|
2018-11-30 13:35:29
|
typelead/eta
|
https://api.github.com/repos/typelead/eta
|
opened
|
Allocate a dedicated thread to detect deadlocks in the runtime system
|
performance rts
|
In certain extreme cases, it is possible for the runtime system to deadlock and lightweight threads to stall. An example of an extreme case is when all of the runtime system allocated threads are all blocked on a Java FFI call. In such a case, it is beneficial to have a separate, dedicated thread that monitors the progress of lightweight threads and will spawn new heavy threads as required if no progress is being made.
|
True
|
Allocate a dedicated thread to detect deadlocks in the runtime system - In certain extreme cases, it is possible for the runtime system to deadlock and lightweight threads to stall. An example of an extreme case is when all of the runtime system allocated threads are all blocked on a Java FFI call. In such a case, it is beneficial to have a separate, dedicated thread that monitors the progress of lightweight threads and will spawn new heavy threads as required if no progress is being made.
|
non_test
|
allocate a dedicated thread to detect deadlocks in the runtime system in certain extreme cases it is possible for the runtime system to deadlock and lightweight threads to stall an example of an extreme case is when all of the runtime system allocated threads are all blocked on a java ffi call in such a case it is beneficial to have a separate dedicated thread that monitors the progress of lightweight threads and will spawn new heavy threads as required if no progress is being made
| 0
|
248,170
| 21,000,965,458
|
IssuesEvent
|
2022-03-29 17:24:12
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
[test-failed]: Chrome UI Functional Tests1.test/functional/apps/home/_welcome·ts - homepage app Welcome interstitial is displayed on a fresh on-prem install
|
failed-test test-cloud
|
**Version: 7.17.2**
**Class: Chrome UI Functional Tests1.test/functional/apps/home/_welcome·ts**
**Stack Trace:**
```
TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="homeWelcomeInterstitial"])
Wait timed out after 10126ms
at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/basicGrp1/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:901:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5) {
remoteStacktrace: ''
}
```
**Other test failures:**
- homepage app Welcome interstitial clicking on "Explore on my own" redirects to the "home" page
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/3054/testReport/_
|
2.0
|
[test-failed]: Chrome UI Functional Tests1.test/functional/apps/home/_welcome·ts - homepage app Welcome interstitial is displayed on a fresh on-prem install - **Version: 7.17.2**
**Class: Chrome UI Functional Tests1.test/functional/apps/home/_welcome·ts**
**Stack Trace:**
```
TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="homeWelcomeInterstitial"])
Wait timed out after 10126ms
at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/basicGrp1/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:901:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5) {
remoteStacktrace: ''
}
```
**Other test failures:**
- homepage app Welcome interstitial clicking on "Explore on my own" redirects to the "home" page
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/3054/testReport/_
|
test
|
chrome ui functional test functional apps home welcome·ts homepage app welcome interstitial is displayed on a fresh on prem install version class chrome ui functional test functional apps home welcome·ts stack trace timeouterror waiting for element to be located by css selector wait timed out after at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules selenium webdriver lib webdriver js at runmicrotasks at processticksandrejections node internal process task queues remotestacktrace other test failures homepage app welcome interstitial clicking on explore on my own redirects to the home page test report
| 1
|
104,420
| 8,972,309,518
|
IssuesEvent
|
2019-01-29 17:56:39
|
nasa-gibs/worldview
|
https://api.github.com/repos/nasa-gibs/worldview
|
closed
|
Add cssnext, a postcss plugin, to the build process
|
enhancement good first issue help wanted testing
|
### Description
Right now we are using these `postcss` plugins to build our CSS: `postcssImport`, `autoprefixer` and `cssnano`.
I suggest we add the [cssnext](http://cssnext.io/) plugin to this pipeline. `cssnext` will allow us to use features found in SASS and LESS without having to write in a separate style sheet language. Using `postcss` will make our CSS future-proof since browsers will eventually support these features.
|
1.0
|
Add cssnext, a postcss plugin, to the build process - ### Description
Right now we are using these `postcss` plugins to build our CSS: `postcssImport`, `autoprefixer` and `cssnano`.
I suggest we add the [cssnext](http://cssnext.io/) plugin to this pipeline. `cssnext` will allow us to use features found in SASS and LESS without having to write in a separate style sheet language. Using `postcss` will make our CSS future-proof since browsers will eventually support these features.
|
test
|
add cssnext a postcss plugin to the build process description right now we are using these postcss plugins to build our css postcssimport autoprefixer and cssnano i suggest we add the plugin to this pipeline cssnext will allow us to use features found in sass and less without having to write in a separate style sheet language using postcss will make our css future proof since browsers will eventually support these features
| 1
|
57,229
| 8,166,082,940
|
IssuesEvent
|
2018-08-25 04:00:33
|
zcash/zcash
|
https://api.github.com/repos/zcash/zcash
|
opened
|
Clean up RPC help messages to have consistent naming of shielded funds
|
RPC interface documentation
|
Re: "private funds" or "shielded funds"?
from comment
https://github.com/zcash/zcash/pull/3436#discussion_r212747800
|
1.0
|
Clean up RPC help messages to have consistent naming of shielded funds - Re: "private funds" or "shielded funds"?
from comment
https://github.com/zcash/zcash/pull/3436#discussion_r212747800
|
non_test
|
clean up rpc help messages to have consistent naming of shielded funds re private funds or shielded funds from comment
| 0
|
394,332
| 11,641,537,277
|
IssuesEvent
|
2020-02-29 03:16:23
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
skaffold dev - deployment fails, but no events
|
area/dev kind/question priority/p2
|
<!--
Issues without logs and details are more complicated to fix.
Please help us by filling the template below!
-->
### Expected behavior
I can see following message after "skaffold dev"
````bash
0/3 nodes are available: 3 node(s) didn't find available persistent volumes to bind.
````
### Actual behavior
````bash
- serviceaccount/commander-svc configured
- role.rbac.authorization.k8s.io/commander-role configured
- rolebinding.rbac.authorization.k8s.io/commander-role-binding configured
- storageclass.storage.k8s.io/local-storage configured
- persistentvolumeclaim/supplier-pvc configured
- secret/x configured
- deployment.apps/redis configured
Watching for changes...
````
### Information
- Skaffold version: v1.3.0
- Operating system: Windows 10
- Contents of skaffold.yaml:
````yaml
apiVersion: skaffold/v2alpha3
kind: Config
metadata:
name: myname
build:
artifacts:
- image: x/a
context: commander
- image: x/b
context: supplier
deploy:
kubectl:
manifests:
- deployment.yaml
````
### Steps to reproduce the behavior
````yaml
apiVersion: v1
kind: ServiceAccount
metadata:
name: commander-svc
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: commander-role
rules:
- apiGroups: [""]
resources: ["pods"]
verbs: ["create", "delete", "list"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
name: commander-role-binding
subjects:
- kind: ServiceAccount
name: commander-svc
roleRef:
kind: Role
apiGroup: rbac.authorization.k8s.io
name: commander-role
---
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
name: local-storage
provisioner: kubernetes.io/no-provisioner
volumeBindingMode: WaitForFirstConsumer
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: supplier-pvc
spec:
storageClassName: local-storage
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 30Gi
---
apiVersion: v1
kind: Secret
metadata:
name: mysecret
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: redis
spec:
replicas: 1
selector:
matchLabels:
app: redis
template:
metadata:
labels:
app: redis
spec:
serviceAccountName: commander-svc
volumes:
- name: supplier-pvc
persistentVolumeClaim:
claimName: supplier-pvc
containers:
- name: redis
image: redis:alpine
ports:
- containerPort: 6379
- name: supplier
image: x/b
envFrom:
- secretRef:
name: mysecret
volumeMounts:
- mountPath: /supplier/supplies
name: supplier-pvc
subPath: supplier/supplies
- name: commander
image: x/a
envFrom:
- secretRef:
name: mysecret
volumeMounts:
- mountPath: /supplier/supplies
name: supplier-pvc
subPath: supplier/supplies
env:
- name: NAMESPACE
value: "default"
````
2. `skaffold dev`
|
1.0
|
skaffold dev - deployment fails, but no events - <!--
Issues without logs and details are more complicated to fix.
Please help us by filling the template below!
-->
### Expected behavior
I can see following message after "skaffold dev"
````bash
0/3 nodes are available: 3 node(s) didn't find available persistent volumes to bind.
````
### Actual behavior
````bash
- serviceaccount/commander-svc configured
- role.rbac.authorization.k8s.io/commander-role configured
- rolebinding.rbac.authorization.k8s.io/commander-role-binding configured
- storageclass.storage.k8s.io/local-storage configured
- persistentvolumeclaim/supplier-pvc configured
- secret/x configured
- deployment.apps/redis configured
Watching for changes...
````
### Information
- Skaffold version: v1.3.0
- Operating system: Windows 10
- Contents of skaffold.yaml:
````yaml
apiVersion: skaffold/v2alpha3
kind: Config
metadata:
name: myname
build:
artifacts:
- image: x/a
context: commander
- image: x/b
context: supplier
deploy:
kubectl:
manifests:
- deployment.yaml
````
### Steps to reproduce the behavior
````yaml
apiVersion: v1
kind: ServiceAccount
metadata:
name: commander-svc
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: commander-role
rules:
- apiGroups: [""]
resources: ["pods"]
verbs: ["create", "delete", "list"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
name: commander-role-binding
subjects:
- kind: ServiceAccount
name: commander-svc
roleRef:
kind: Role
apiGroup: rbac.authorization.k8s.io
name: commander-role
---
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
name: local-storage
provisioner: kubernetes.io/no-provisioner
volumeBindingMode: WaitForFirstConsumer
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: supplier-pvc
spec:
storageClassName: local-storage
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 30Gi
---
apiVersion: v1
kind: Secret
metadata:
name: mysecret
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: redis
spec:
replicas: 1
selector:
matchLabels:
app: redis
template:
metadata:
labels:
app: redis
spec:
serviceAccountName: commander-svc
volumes:
- name: supplier-pvc
persistentVolumeClaim:
claimName: supplier-pvc
containers:
- name: redis
image: redis:alpine
ports:
- containerPort: 6379
- name: supplier
image: x/b
envFrom:
- secretRef:
name: mysecret
volumeMounts:
- mountPath: /supplier/supplies
name: supplier-pvc
subPath: supplier/supplies
- name: commander
image: x/a
envFrom:
- secretRef:
name: mysecret
volumeMounts:
- mountPath: /supplier/supplies
name: supplier-pvc
subPath: supplier/supplies
env:
- name: NAMESPACE
value: "default"
````
2. `skaffold dev`
|
non_test
|
skaffold dev deployment fails but no events issues without logs and details are more complicated to fix please help us by filling the template below expected behavior i can see following message after skaffold dev bash nodes are available node s didn t find available persistent volumes to bind actual behavior bash serviceaccount commander svc configured role rbac authorization io commander role configured rolebinding rbac authorization io commander role binding configured storageclass storage io local storage configured persistentvolumeclaim supplier pvc configured secret x configured deployment apps redis configured watching for changes information skaffold version operating system windows contents of skaffold yaml yaml apiversion skaffold kind config metadata name myname build artifacts image x a context commander image x b context supplier deploy kubectl manifests deployment yaml steps to reproduce the behavior yaml apiversion kind serviceaccount metadata name commander svc apiversion rbac authorization io kind role metadata name commander role rules apigroups resources verbs apiversion rbac authorization io kind rolebinding metadata name commander role binding subjects kind serviceaccount name commander svc roleref kind role apigroup rbac authorization io name commander role apiversion storage io kind storageclass metadata name local storage provisioner kubernetes io no provisioner volumebindingmode waitforfirstconsumer apiversion kind persistentvolumeclaim metadata name supplier pvc spec storageclassname local storage accessmodes readwriteonce resources requests storage apiversion kind secret metadata name mysecret apiversion apps kind deployment metadata name redis spec replicas selector matchlabels app redis template metadata labels app redis spec serviceaccountname commander svc volumes name supplier pvc persistentvolumeclaim claimname supplier pvc containers name redis image redis alpine ports containerport name supplier image x b envfrom secretref name mysecret volumemounts mountpath supplier supplies name supplier pvc subpath supplier supplies name commander image x a envfrom secretref name mysecret volumemounts mountpath supplier supplies name supplier pvc subpath supplier supplies env name namespace value default skaffold dev
| 0
|
212,864
| 16,485,811,938
|
IssuesEvent
|
2021-05-24 17:47:05
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
Add validatePodmanEnv subtest for crio runtime to TestFunctional
|
area/testing co/runtime/crio kind/feature lifecycle/rotten priority/important-longterm
|
This goes hand in hand with validateDockerEnv which currently runs for the docker runtime
|
1.0
|
Add validatePodmanEnv subtest for crio runtime to TestFunctional - This goes hand in hand with validateDockerEnv which currently runs for the docker runtime
|
test
|
add validatepodmanenv subtest for crio runtime to testfunctional this goes hand in hand with validatedockerenv which currently runs for the docker runtime
| 1
|
262,287
| 22,829,445,191
|
IssuesEvent
|
2022-07-12 11:37:25
|
Bouboule-Corp/thurii-mobile-kotlin
|
https://api.github.com/repos/Bouboule-Corp/thurii-mobile-kotlin
|
opened
|
uppgrade coverage %
|
enhancement missing test
|
- [ ] dossier DoubleAuth
- [ ] dossier EmailRegistration
- [ ] dossier HomePage
- [ ] dossier Login
- [ ] dossier LoginSignInMenu
- [ ] dossier Registration
- [ ] dossier Settings
- [ ] fichier MainActivity.kt
|
1.0
|
uppgrade coverage % - - [ ] dossier DoubleAuth
- [ ] dossier EmailRegistration
- [ ] dossier HomePage
- [ ] dossier Login
- [ ] dossier LoginSignInMenu
- [ ] dossier Registration
- [ ] dossier Settings
- [ ] fichier MainActivity.kt
|
test
|
uppgrade coverage dossier doubleauth dossier emailregistration dossier homepage dossier login dossier loginsigninmenu dossier registration dossier settings fichier mainactivity kt
| 1
|
105,778
| 13,216,931,175
|
IssuesEvent
|
2020-08-17 05:31:00
|
nextcloud/server
|
https://api.github.com/repos/nextcloud/server
|
closed
|
Internal / private link
|
client: 💻 desktop client: 🤖🍏 mobile design feature: sharing high overview
|
- [x] Android @tobiasKaminsky
- [x] Desktop @camilasan
- [x] iOS @marinofaggiana
- [x] Server @nextcloud/server-triage
There is a feature request to have the internal/private link (clipboard on the right) also on android.

I discussed this with Jan and he is not very satisfied about the current situation.
Also Marie mentioned that she by accident copied this link to another person and this one could not open the file (as not shared and not even on server).
So, how could this get improved? Once we have a decision, this can also be done the same on all clients.
@nextcloud/designers
@marinofaggiana
@camilasan: newest 2.5.0 has this in option menu: "copy private link" / "copy public link"
|
1.0
|
Internal / private link - - [x] Android @tobiasKaminsky
- [x] Desktop @camilasan
- [x] iOS @marinofaggiana
- [x] Server @nextcloud/server-triage
There is a feature request to have the internal/private link (clipboard on the right) also on android.

I discussed this with Jan and he is not very satisfied about the current situation.
Also Marie mentioned that she by accident copied this link to another person and this one could not open the file (as not shared and not even on server).
So, how could this get improved? Once we have a decision, this can also be done the same on all clients.
@nextcloud/designers
@marinofaggiana
@camilasan: newest 2.5.0 has this in option menu: "copy private link" / "copy public link"
|
non_test
|
internal private link android tobiaskaminsky desktop camilasan ios marinofaggiana server nextcloud server triage there is a feature request to have the internal private link clipboard on the right also on android i discussed this with jan and he is not very satisfied about the current situation also marie mentioned that she by accident copied this link to another person and this one could not open the file as not shared and not even on server so how could this get improved once we have a decision this can also be done the same on all clients nextcloud designers marinofaggiana camilasan newest has this in option menu copy private link copy public link
| 0
|
89,071
| 15,823,696,540
|
IssuesEvent
|
2021-04-06 01:25:19
|
f-peng/centreon
|
https://api.github.com/repos/f-peng/centreon
|
opened
|
CVE-2020-7608 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-13.1.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- webpack-dev-server-3.2.1.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- webpack-cli-3.3.9.tgz (Root Library)
- yargs-13.2.4.tgz
- :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/sass-graph/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.12.0.tgz (Root Library)
- sass-graph-2.2.4.tgz
- yargs-7.1.0.tgz
- :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-06-05</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-13.1.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- webpack-dev-server-3.2.1.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- webpack-cli-3.3.9.tgz (Root Library)
- yargs-13.2.4.tgz
- :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p>
<p>Path to dependency file: centreon/package.json</p>
<p>Path to vulnerable library: centreon/node_modules/sass-graph/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.12.0.tgz (Root Library)
- sass-graph-2.2.4.tgz
- yargs-7.1.0.tgz
- :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-06-05</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file centreon package json path to vulnerable library centreon node modules webpack dev server node modules yargs parser package json dependency hierarchy react scripts tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file centreon package json path to vulnerable library centreon node modules yargs parser package json dependency hierarchy webpack cli tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file centreon package json path to vulnerable library centreon node modules sass graph node modules yargs parser package json dependency hierarchy node sass tgz root library sass graph tgz yargs tgz x yargs parser tgz vulnerable library vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
123,670
| 10,278,931,168
|
IssuesEvent
|
2019-08-25 18:29:47
|
dexpenses/dexpenses-extract
|
https://api.github.com/repos/dexpenses/dexpenses-extract
|
opened
|
Implement test receipt normal/goe-cafe-del-sol
|
enhancement test-data
|
Receipt to implement:

|
1.0
|
Implement test receipt normal/goe-cafe-del-sol - Receipt to implement:

|
test
|
implement test receipt normal goe cafe del sol receipt to implement normal goe cafe del sol
| 1
|
327,659
| 28,075,599,473
|
IssuesEvent
|
2023-03-29 23:09:37
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix creation.test_native_array
|
Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543293886/jobs/8007800682" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-03-28T13:12:43.7353006Z E TypeError: to_native() got an unexpected keyword argument 'dtype'
2023-03-28T13:12:43.7353395Z E Falsifying example: test_native_array(
2023-03-28T13:12:43.7353843Z E dtype_and_x_and_cast_dtype=(['int64'],
2023-03-28T13:12:43.7354145Z E [array(0)],
2023-03-28T13:12:43.7354506Z E ('int64', [array(0)], 'int64')),
2023-03-28T13:12:43.7354822Z E test_flags=FunctionTestFlags(
2023-03-28T13:12:43.7355150Z E num_positional_args=1,
2023-03-28T13:12:43.7355453Z E with_out=False,
2023-03-28T13:12:43.7355894Z E instance_method=False,
2023-03-28T13:12:43.7356195Z E test_gradients=False,
2023-03-28T13:12:43.7356480Z E test_compile=None,
2023-03-28T13:12:43.7357004Z E as_variable=[False],
2023-03-28T13:12:43.7357310Z E native_arrays=[False],
2023-03-28T13:12:43.7357598Z E container=[False],
2023-03-28T13:12:43.7357875Z E ),
2023-03-28T13:12:43.7358213Z E ground_truth_backend='tensorflow',
2023-03-28T13:12:43.7358585Z E fn_name='native_array',
2023-03-28T13:12:43.7359140Z E backend_fw=<module 'ivy.functional.backends.numpy' from '/ivy/ivy/functional/backends/numpy/__init__.py'>,
2023-03-28T13:12:43.7360012Z E on_device='cpu',
2023-03-28T13:12:43.7361258Z E )
2023-03-28T13:12:43.7361539Z E
2023-03-28T13:12:43.7362293Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AAEAAAAAAAEAAAAAAA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
|
1.0
|
Fix creation.test_native_array - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543293886/jobs/8007800682" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4520842325/jobs/7962215017" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-03-28T13:12:43.7353006Z E TypeError: to_native() got an unexpected keyword argument 'dtype'
2023-03-28T13:12:43.7353395Z E Falsifying example: test_native_array(
2023-03-28T13:12:43.7353843Z E dtype_and_x_and_cast_dtype=(['int64'],
2023-03-28T13:12:43.7354145Z E [array(0)],
2023-03-28T13:12:43.7354506Z E ('int64', [array(0)], 'int64')),
2023-03-28T13:12:43.7354822Z E test_flags=FunctionTestFlags(
2023-03-28T13:12:43.7355150Z E num_positional_args=1,
2023-03-28T13:12:43.7355453Z E with_out=False,
2023-03-28T13:12:43.7355894Z E instance_method=False,
2023-03-28T13:12:43.7356195Z E test_gradients=False,
2023-03-28T13:12:43.7356480Z E test_compile=None,
2023-03-28T13:12:43.7357004Z E as_variable=[False],
2023-03-28T13:12:43.7357310Z E native_arrays=[False],
2023-03-28T13:12:43.7357598Z E container=[False],
2023-03-28T13:12:43.7357875Z E ),
2023-03-28T13:12:43.7358213Z E ground_truth_backend='tensorflow',
2023-03-28T13:12:43.7358585Z E fn_name='native_array',
2023-03-28T13:12:43.7359140Z E backend_fw=<module 'ivy.functional.backends.numpy' from '/ivy/ivy/functional/backends/numpy/__init__.py'>,
2023-03-28T13:12:43.7360012Z E on_device='cpu',
2023-03-28T13:12:43.7361258Z E )
2023-03-28T13:12:43.7361539Z E
2023-03-28T13:12:43.7362293Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AAEAAAAAAAEAAAAAAA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_creation.py::test_native_array[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-25T20:13:11.8090895Z E AssertionError: There are no parameters in the inputs to connect the outputs to
2023-03-25T20:13:11.8091389Z E Falsifying example: test_native_array(
2023-03-25T20:13:11.8091955Z E dtype_and_x_and_cast_dtype=(['uint16'],
2023-03-25T20:13:11.8092385Z E [array(0, dtype=uint16)],
2023-03-25T20:13:11.8092910Z E ('uint16', [array(0, dtype=uint16)], 'complex128')),
2023-03-25T20:13:11.8093420Z E ground_truth_backend='tensorflow',
2023-03-25T20:13:11.8093845Z E test_flags=FunctionTestFlags(
2023-03-25T20:13:11.8094237Z E num_positional_args=1,
2023-03-25T20:13:11.8094597Z E with_out=False,
2023-03-25T20:13:11.8094956Z E instance_method=False,
2023-03-25T20:13:11.8095323Z E test_gradients=False,
2023-03-25T20:13:11.8095700Z E test_compile=True,
2023-03-25T20:13:11.8096067Z E as_variable=[False],
2023-03-25T20:13:11.8096438Z E native_arrays=[False],
2023-03-25T20:13:11.8096800Z E container=[False],
2023-03-25T20:13:11.8097130Z E ),
2023-03-25T20:13:11.8097519Z E fn_name='native_array',
2023-03-25T20:13:11.8098173Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-25T20:13:11.8098720Z E on_device='cpu',
2023-03-25T20:13:11.8099051Z E )
2023-03-25T20:13:11.8099737Z E
2023-03-25T20:13:11.8100440Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AAQAAAAAAAABAAA=') as a decorator on your test case
</details>
|
test
|
fix creation test native array tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test core test creation py test native array e assertionerror there are no parameters in the inputs to connect the outputs to e falsifying example test native array e dtype and x and cast dtype e e e ground truth backend tensorflow e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile true e as variable e native arrays e container e e fn name native array e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b aaqaaaaaaaabaaa as a decorator on your test case failed ivy tests test ivy test functional test core test creation py test native array e assertionerror there are no parameters in the inputs to connect the outputs to e falsifying example test native array e dtype and x and cast dtype e e e ground truth backend tensorflow e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile true e as variable e native arrays e container e e fn name native array e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b aaqaaaaaaaabaaa as a decorator on your test case failed ivy tests test ivy test functional test core test creation py test native array e typeerror to native got an unexpected keyword argument dtype e falsifying example test native array e dtype and x and cast dtype e e e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile none e as variable e native arrays e container e e ground truth backend tensorflow e fn name native array e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b aaeaaaaaaaeaaaaaaa as a decorator on your test case failed ivy tests test ivy test functional test core test creation py test native array e assertionerror there are no parameters in the inputs to connect the outputs to e falsifying example test native array e dtype and x and cast dtype e e e ground truth backend tensorflow e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile true e as variable e native arrays e container e e fn name native array e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b aaqaaaaaaaabaaa as a decorator on your test case
| 1
|
165,911
| 12,884,028,752
|
IssuesEvent
|
2020-07-13 01:24:10
|
dhenry-KCI/FredCo-Post-Go-Live-
|
https://api.github.com/repos/dhenry-KCI/FredCo-Post-Go-Live-
|
closed
|
R4C - Permit & Planning Application Submission Issues - Extreme Ticket #14547500
|
High Priority Ready for Testing in TEST
|
A customer has had consistent issues when trying to submit her ResUse - Deck permit via the R4C portal. See the applicant's description of her issue below.

It appears as though the application submits successfully when the submit button is clicked for the first time but potential loading and timeout issues are causing R4C to not load correctly and makes the applicant think it wasn't submitted. As a result applicant continue to try to upload multiple times. A few more examples are visible in the Lookup Application window.

|
2.0
|
R4C - Permit & Planning Application Submission Issues - Extreme Ticket #14547500 - A customer has had consistent issues when trying to submit her ResUse - Deck permit via the R4C portal. See the applicant's description of her issue below.

It appears as though the application submits successfully when the submit button is clicked for the first time but potential loading and timeout issues are causing R4C to not load correctly and makes the applicant think it wasn't submitted. As a result applicant continue to try to upload multiple times. A few more examples are visible in the Lookup Application window.

|
test
|
permit planning application submission issues extreme ticket a customer has had consistent issues when trying to submit her resuse deck permit via the portal see the applicant s description of her issue below it appears as though the application submits successfully when the submit button is clicked for the first time but potential loading and timeout issues are causing to not load correctly and makes the applicant think it wasn t submitted as a result applicant continue to try to upload multiple times a few more examples are visible in the lookup application window
| 1
|
208,746
| 23,646,355,201
|
IssuesEvent
|
2022-08-25 22:49:40
|
ZcashFoundation/zebra
|
https://api.github.com/repos/ZcashFoundation/zebra
|
closed
|
Simplify initial network setup so that peer queries can be answered sooner
|
C-enhancement S-needs-triage P-Low :snowflake: C-security I-slow A-network
|
## Motivation
Ziggurat's testing report on Zebra `v1.0.0-beta.8` discovered that Zebra ignores inbound peer queries during network setup.
### Designs
There are a few ways we can resolve this issue:
1. Split `Inbound` initialisation into fast (state) and slow (network/verifier), so that some peer queries can be answered sooner
2. Move peer queries to the `ReadStateService`, so they can be served faster
3. Remove the initial connection phase from network setup, and just use the dialer to connect to all peers (needs a security analysis of Zebra's peer connection order)
Due to Rust's ownership rules, and Zebra's architecture, network setup will always have a short time when it can't answer peer queries. And we don't want to drop peers due to overload during that time.
But these changes would make that as short as possible.
|
True
|
Simplify initial network setup so that peer queries can be answered sooner - ## Motivation
Ziggurat's testing report on Zebra `v1.0.0-beta.8` discovered that Zebra ignores inbound peer queries during network setup.
### Designs
There are a few ways we can resolve this issue:
1. Split `Inbound` initialisation into fast (state) and slow (network/verifier), so that some peer queries can be answered sooner
2. Move peer queries to the `ReadStateService`, so they can be served faster
3. Remove the initial connection phase from network setup, and just use the dialer to connect to all peers (needs a security analysis of Zebra's peer connection order)
Due to Rust's ownership rules, and Zebra's architecture, network setup will always have a short time when it can't answer peer queries. And we don't want to drop peers due to overload during that time.
But these changes would make that as short as possible.
|
non_test
|
simplify initial network setup so that peer queries can be answered sooner motivation ziggurat s testing report on zebra beta discovered that zebra ignores inbound peer queries during network setup designs there are a few ways we can resolve this issue split inbound initialisation into fast state and slow network verifier so that some peer queries can be answered sooner move peer queries to the readstateservice so they can be served faster remove the initial connection phase from network setup and just use the dialer to connect to all peers needs a security analysis of zebra s peer connection order due to rust s ownership rules and zebra s architecture network setup will always have a short time when it can t answer peer queries and we don t want to drop peers due to overload during that time but these changes would make that as short as possible
| 0
|
17,828
| 23,768,850,174
|
IssuesEvent
|
2022-09-01 14:45:36
|
ArneBinder/pie-utils
|
https://api.github.com/repos/ArneBinder/pie-utils
|
closed
|
create a partition via regex
|
document processor
|
Implement a document processor that creates a partition via a regex split pattern. This should take advantage from [previous implementation](https://github.com/ArneBinder/pytorch-ie-sam-template/blob/main/src/document_processors/partition.py). This should also collect the distribution of the lengths of the parts (partition entries) and the full texts (to compare against) and also the number of parts per document.
|
1.0
|
create a partition via regex - Implement a document processor that creates a partition via a regex split pattern. This should take advantage from [previous implementation](https://github.com/ArneBinder/pytorch-ie-sam-template/blob/main/src/document_processors/partition.py). This should also collect the distribution of the lengths of the parts (partition entries) and the full texts (to compare against) and also the number of parts per document.
|
non_test
|
create a partition via regex implement a document processor that creates a partition via a regex split pattern this should take advantage from this should also collect the distribution of the lengths of the parts partition entries and the full texts to compare against and also the number of parts per document
| 0
|
180,697
| 30,549,230,807
|
IssuesEvent
|
2023-07-20 07:19:52
|
DeveloperAcademy-POSTECH/MC3-Team8-Aing
|
https://api.github.com/repos/DeveloperAcademy-POSTECH/MC3-Team8-Aing
|
closed
|
[Design] 카메라뷰 레이아웃 제작
|
Design 클리프
|
## 📸 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
- 카메라 뷰 레이아웃 제작
- 카메라 뷰 페이지 전체는 UIKit으로 제작한 뒤 Representable로 래핑
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [x] 컴포넌트 배치
- [x] 오토 레이아웃 및 Constraints 설정
|
1.0
|
[Design] 카메라뷰 레이아웃 제작 - ## 📸 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
- 카메라 뷰 레이아웃 제작
- 카메라 뷰 페이지 전체는 UIKit으로 제작한 뒤 Representable로 래핑
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [x] 컴포넌트 배치
- [x] 오토 레이아웃 및 Constraints 설정
|
non_test
|
카메라뷰 레이아웃 제작 📸 issue 카메라 뷰 레이아웃 제작 카메라 뷰 페이지 전체는 uikit으로 제작한 뒤 representable로 래핑 📝 to do 컴포넌트 배치 오토 레이아웃 및 constraints 설정
| 0
|
249,964
| 21,219,486,395
|
IssuesEvent
|
2022-04-11 10:30:26
|
LimeChain/hashport-validator
|
https://api.github.com/repos/LimeChain/hashport-validator
|
closed
|
Unit tests for app/process/handler/nft/transfer/handler.go
|
unit tests
|
- Implement unit tests for **app/process/handler/nft/transfer/handler.go**
|
1.0
|
Unit tests for app/process/handler/nft/transfer/handler.go - - Implement unit tests for **app/process/handler/nft/transfer/handler.go**
|
test
|
unit tests for app process handler nft transfer handler go implement unit tests for app process handler nft transfer handler go
| 1
|
293,285
| 25,281,668,496
|
IssuesEvent
|
2022-11-16 16:11:45
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
opened
|
Add request analysisd configuration test
|
team/qa test/integration type/test-development status/not-tracked subteam/qa-main
|
| Target version | Related issue |
|--------------------|--------------------|
| 4.4 | #3112 |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
We need to add a test that checks if the analysisd configuration, requested with the API is the expected.
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
- [ ] Get and check the default configuration
- [ ] Get and check custom configuration
|
2.0
|
Add request analysisd configuration test - | Target version | Related issue |
|--------------------|--------------------|
| 4.4 | #3112 |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
We need to add a test that checks if the analysisd configuration, requested with the API is the expected.
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
- [ ] Get and check the default configuration
- [ ] Get and check custom configuration
|
test
|
add request analysisd configuration test target version related issue description we need to add a test that checks if the analysisd configuration requested with the api is the expected proposed test cases get and check the default configuration get and check custom configuration
| 1
|
41,168
| 16,652,475,851
|
IssuesEvent
|
2021-06-05 00:09:14
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
az webapp identity assign fails with `NoneType` object is not callable
|
Managed Identity Service Attention Web Apps
|
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az webapp identity assign`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
'NoneType' object is not callable
Traceback (most recent call last):
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\knack/cli.py", line 231, in invoke
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 657, in execute
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 720, in _run_jobs_serially
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 691, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 328, in __call__
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 121, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/appservice/custom.py", line 789, in assign_identity
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/arm.py", line 706, in assign_identity
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/appservice/custom.py", line 782, in setter
TypeError: 'NoneType' object is not callable
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az webapp identity assign --name {} --resource-group {} --slot {} --identities {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.8.9
Installer: MSI
azure-cli 2.23.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
1.0
|
az webapp identity assign fails with `NoneType` object is not callable - ### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az webapp identity assign`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
'NoneType' object is not callable
Traceback (most recent call last):
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\knack/cli.py", line 231, in invoke
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 657, in execute
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 720, in _run_jobs_serially
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 691, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 328, in __call__
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 121, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/appservice/custom.py", line 789, in assign_identity
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/arm.py", line 706, in assign_identity
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/appservice/custom.py", line 782, in setter
TypeError: 'NoneType' object is not callable
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az webapp identity assign --name {} --resource-group {} --slot {} --identities {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.8.9
Installer: MSI
azure-cli 2.23.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
non_test
|
az webapp identity assign fails with nonetype object is not callable this is autogenerated please review and update as needed describe the bug command name az webapp identity assign errors the command failed with an unexpected error here is the traceback nonetype object is not callable traceback most recent call last file d a s build scripts windows artifacts cli lib site packages knack cli py line in invoke file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in execute file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in run jobs serially file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in run job file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in call file d a s build scripts windows artifacts cli lib site packages azure cli core commands command operation py line in handler file d a s build scripts windows artifacts cli lib site packages azure cli command modules appservice custom py line in assign identity file d a s build scripts windows artifacts cli lib site packages azure cli core commands arm py line in assign identity file d a s build scripts windows artifacts cli lib site packages azure cli command modules appservice custom py line in setter typeerror nonetype object is not callable to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az webapp identity assign name resource group slot identities expected behavior environment summary windows python installer msi azure cli additional context
| 0
|
443,754
| 12,799,098,009
|
IssuesEvent
|
2020-07-02 14:53:35
|
TheRandomLabs/RandomPatches
|
https://api.github.com/repos/TheRandomLabs/RandomPatches
|
closed
|
[MC1.12.2 - 1.18.1.1] TickNextTick error still occurs, world crashing
|
bug help wanted low priority
|
---- Minecraft Crash Report ----
WARNING: coremods are present:
RandomPatches (randompatches-1.12.2-1.18.1.1.jar)
SplashAnimationCoremod (SplashAnimation-0.2.1.jar)
BedPatch (bedpatch-2.2-1.12.2.jar)
Do not report to Forge! (If you haven't disabled the FoamFix coremod, try disabling it in the config! Note that this bit of text will still appear.) (foamfix-0.10.5-1.12.2.jar)
Quark Plugin (Quark-r1.6-177.jar)
OTGCorePlugin (OTG-Core.jar)
OpenModsCorePlugin (OpenModsLib-1.12.2-0.12.2.jar)
TransformerLoader (OpenComputers-MC1.12.2-1.7.4.153.jar)
EnderCorePlugin (EnderCore-1.12.2-0.5.65-core.jar)
CTMCorePlugin (CTM-MC1.12.2-0.3.3.22.jar)
p455w0rdc0re (p455w0rdslib-1.12.2-2.2.156.jar)
ForgelinPlugin (Forgelin-1.8.3.jar)
Plugin (NotEnoughIDs-1.5.4.4.jar)
CorePlugin (ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar)
IELoadingPlugin (ImmersiveEngineering-core-0.12-89.jar)
Inventory Tweaks Coremod (InventoryTweaks-1.63.jar)
ShetiPhian-ASM (ShetiPhian-ASM-1.12.0.jar)
MekanismCoremod (Mekanism-1.12.2-9.8.1.383.jar)
LoadingPlugin (ResourceLoader-MC1.12.1-1.5.3.jar)
CoreMod (Aroma1997Core-1.12.2-2.0.0.2.b165.jar)
DynamicSurroundingsCore (DynamicSurroundings-core-1.12.2-3.5.4.3.jar)
AdvancedRocketryPlugin (AdvancedRocketry-1.12.2-1.6.0-164-universal.jar)
Contact their authors BEFORE contacting forge
// Don't do that.
Time: 9/12/19 6:44 PM
Description: Exception ticking world
java.lang.IllegalStateException: TickNextTick list out of synch
at net.minecraft.world.WorldServer.func_72955_a(WorldServer.java:691)
at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:223)
at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:756)
at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397)
at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668)
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526)
at java.lang.Thread.run(Thread.java:748)
A detailed walkthrough of the error, its code path and all known details is as follows:
---------------------------------------------------------------------------------------
-- Head --
Thread: Server thread
Stacktrace:
at net.minecraft.world.WorldServer.func_72955_a(WorldServer.java:691)
at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:223)
-- Affected level --
Details:
Level name: VanillaVistas
All players: 1 total; [EntityPlayerMP['firemx'/566, l='VanillaVistas', x=115.22, y=85.79, z=-1632.99]]
Chunk stats: ServerChunkCache: 1357 Drop: 0
Level seed: -8290667671068293349
Level generator: ID 11 - OTG, ver 0. Features enabled: true
Level generator options: OpenTerrainGenerator
Level spawn location: World: (365,64,-1295), Chunk: (at 13,4,1 in 22,-81; contains blocks 352,0,-1296 to 367,255,-1281), Region: (0,-3; contains chunks 0,-96 to 31,-65, blocks 0,0,-1536 to 511,255,-1025)
Level time: 20033484 game time, 2840813 day time
Level dimension: 0
Level storage version: 0x04ABD - Anvil
Level weather: Rain time: 40550 (now: false), thunder time: 13025 (now: true)
Level game mode: Game mode: survival (ID 0). Hardcore: false. Cheats: false
Stacktrace:
at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:756)
at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397)
at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668)
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526)
at java.lang.Thread.run(Thread.java:748)
-- System Details --
Details:
Minecraft Version: 1.12.2
Operating System: Linux (amd64) version 3.10.0-957.21.3.el7.x86_64
Java Version: 1.8.0_212, Oracle Corporation
Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Oracle Corporation
Memory: 5802951920 bytes (5534 MB) / 7057965056 bytes (6731 MB) up to 7057965056 bytes (6731 MB)
JVM Flags: 3 total; -Xmx7168M -Xms7168M -XX:MaxPermSize=256m
IntCache: cache: 0, tcache: 0, allocated: 0, tallocated: 0
FML: MCP 9.42 Powered by Forge 14.23.5.2838 207 mods loaded, 207 mods active
States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored
| State | ID | Version | Source | Signature |
|:------ |:--------------------------------- |:------------------------ |:--------------------------------------------------- |:---------------------------------------- |
| LCHIJA | minecraft | 1.12.2 | minecraft.jar | None |
| LCHIJA | mcp | 9.42 | minecraft.jar | None |
| LCHIJA | FML | 8.0.99.99 | forge1122.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 |
| LCHIJA | forge | 14.23.5.2838 | forge1122.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 |
| LCHIJA | advancedrocketrycore | 1 | minecraft.jar | None |
| LCHIJA | otgcore | 1.12.2 - v7 | minecraft.jar | None |
| LCHIJA | openmodscore | 0.12.2 | minecraft.jar | None |
| LCHIJA | foamfixcore | 7.7.4 | minecraft.jar | None |
| LCHIJA | opencomputers|core | 1.7.4.153 | minecraft.jar | None |
| LCHIJA | dsurroundcore | 3.5.4.3 | minecraft.jar | None |
| LCHIJA | randompatches | 1.12.2-1.18.1.1 | randompatches-1.12.2-1.18.1.1.jar | None |
| LCHIJA | fastbench | 1.7.2 | FastWorkbench-1.12.2-1.7.2.jar | None |
| LCHIJA | actuallyadditions | 1.12.2-r149 | ActuallyAdditions-1.12.2-r149.jar | None |
| LCHIJA | baubles | 1.5.2 | Baubles-1.12-1.5.2.jar | None |
| LCHIJA | actuallybaubles | 1.1 | ActuallyBaubles-1.12-1.1.jar | None |
| LCHIJA | forgeendertech | 1.12.2-4.4.22.2 | ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar | None |
| LCHIJA | adlods | 1.12.2-1.0.6.2 | AdLods-1.12.2-1.0.6.2-build.0408.jar | None |
| LCHIJA | redstoneflux | 2.1.0 | RedstoneFlux-1.12-2.1.0.6-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | cofhcore | 4.6.3 | CoFHCore-1.12.2-4.6.3.27-universal.jar | None |
| LCHIJA | libvulpes | 0.4.1.-62 | LibVulpes-1.12.2-0.4.1-62-universal.jar | None |
| LCHIJA | advancedrocketry | 1.6.0.-164 | AdvancedRocketry-1.12.2-1.6.0-164-universal.jar | None |
| LCHIJA | appliedenergistics2 | rv6-stable-7 | appliedenergistics2-rv6-stable-7.jar | dfa4d3ac143316c6f32aa1a1beda1e34d42132e5 |
| LCHIJA | bdlib | 1.14.3.12 | bdlib-1.14.3.12-mc1.12.2.jar | None |
| LCHIJA | ae2stuff | 0.7.0.4 | ae2stuff-0.7.0.4-mc1.12.2.jar | None |
| LCHIJA | endercore | 1.12.2-0.5.65 | EnderCore-1.12.2-0.5.65.jar | None |
| LCHIJA | crafttweaker | 4.1.19 | CraftTweaker2-1.12-4.1.19.jar | None |
| LCHIJA | mtlib | 3.0.6 | MTLib-3.0.6.jar | None |
| LCHIJA | modtweaker | 4.0.17 | modtweaker-4.0.17.jar | None |
| LCHIJA | jei | 4.15.0.289 | jei_1.12.2-4.15.0.289.jar | None |
| LCHIJA | codechickenlib | 3.2.3.357 | CodeChickenLib-1.12.2-3.2.3.357-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | brandonscore | 2.4.17 | BrandonsCore-1.12.2-2.4.17.208-universal.jar | None |
| LCHIJA | cofhworld | 1.3.1 | CoFHWorld-1.12.2-1.3.1.7-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermalfoundation | 2.6.3 | ThermalFoundation-1.12.2-2.6.3.27-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | draconicevolution | 2.3.24 | Draconic-Evolution-1.12.2-2.3.24.349-universal.jar | None |
| LCHIJA | thermalexpansion | 5.5.4 | ThermalExpansion-1.12.2-5.5.4.43-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | enderio | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | mantle | 1.12-1.3.3.55 | Mantle-1.12-1.3.3.55.jar | None |
| LCHIJA | chisel | MC1.12.2-0.2.1.35 | Chisel-MC1.12.2-0.2.1.35.jar | None |
| LCHIJA | enderiointegrationtic | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | tconstruct | 1.12.2-2.12.0.157 | TConstruct-1.12.2-2.12.0.157.jar | None |
| LCHIJA | p455w0rdslib | 2.2.156 | p455w0rdslib-1.12.2-2.2.156.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | ae2wtlib | 1.0.30 | AE2WTLib-1.12.2-1.0.30.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | infinitylib | 1.12.2-1.12.0 | infinitylib-1.12.0.jar | None |
| LCHIJA | agricraft | 2.12.0-1.12.0-a6 | AgriCraft-2.12.0-1.12.0-a6.jar | None |
| LCHIJA | akashictome | 1.2-12 | AkashicTome-1.2-12.jar | None |
| LCHIJA | appleskin | 1.0.9 | AppleSkin-mc1.12-1.0.9.jar | None |
| LCHIJA | architecturecraft | @VERSION@ | architecturecraft-1.12-3.98.jar | None |
| LCHIJA | aroma1997core | 2.0.0.2.b165 | Aroma1997Core-1.12.2-2.0.0.2.b165.jar | dfbfe4c473253d8c5652417689848f650b2cbe32 |
| LCHIJA | aroma1997sdimension | 2.0.0.2.b89 | Aroma1997s-Dimensional-World-1.12.2-2.0.0.2.b89.jar | dfbfe4c473253d8c5652417689848f650b2cbe32 |
| LCHIJA | morphtool | 1.2-21 | Morph-o-Tool-1.2-21.jar | None |
| LCHIJA | quark | r1.6-177 | Quark-r1.6-177.jar | None |
| LCHIJA | autoreglib | 1.3-32 | AutoRegLib-1.3-32.jar | None |
| LCHIJA | badwithernocookiereloaded | 1.12.2-3.3.16 | badwithernocookiereloaded-1.12.2-3.3.16.jar | None |
| LCHIJA | base | 3.13.0 | base-1.12.2-3.13.0.jar | None |
| LCHIJA | bedpatch | 2.2 | bedpatch-2.2-1.12.2.jar | 6bf7527e690fb5e8719b9832bce5000a3e87dfe6 |
| LCHIJA | betterbuilderswands | 0.13.2 | BetterBuildersWands-1.12.2-0.13.2.271+5997513.jar | None |
| LCHIJA | blockcraftery | 1.12.2-1.3.1 | blockcraftery-1.12.2-1.3.1.jar | None |
| LCHIJA | bookshelf | 2.3.585 | Bookshelf-1.12.2-2.3.585.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | botania | r1.10-361 | Botania r1.10-361.jar | None |
| LCHIJA | chameleon | 1.12-4.1.3 | Chameleon-1.12-4.1.3.jar | None |
| LCHIJA | chesttransporter | 2.8.8 | ChestTransporter-1.12.2-2.8.8.jar | None |
| LCHIJA | chiselsandbits | 14.33 | chiselsandbits-14.33.jar | None |
| LCHIJA | clumps | 3.1.2 | Clumps-3.1.2.jar | None |
| LCHIJA | compactmachines3 | 3.0.18 | compactmachines3-1.12.2-3.0.18-b278.jar | None |
| LCHIJA | conarm | 1.2.4 | conarm-1.12.2-1.2.4.jar | 5d5b8aee896a4f5ea3f3114784742662a67ad32f |
| LCHIJA | cookingforblockheads | 6.4.71 | CookingForBlockheads_1.12.2-6.4.71.jar | None |
| LCHIJA | crafttweakerjei | 2.0.3 | CraftTweaker2-1.12-4.1.19.jar | None |
| LCHIJA | theoneprobe | 1.4.28 | theoneprobe-1.12-1.4.28.jar | None |
| LCHIJA | danknull | 1.6.77 | DankNull-1.12.2-1.6.77.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | darkutils | 1.8.226 | DarkUtils-1.12.2-1.8.226.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | diethopper | 1.1 | diethopper-1.1.jar | None |
| LCHIJA | dirt2path | 1.7.1 | dirt2path-1.8.0.jar | None |
| LCHIJA | dynamictrees | 1.12.2-0.9.5 | DynamicTrees-1.12.2-0.9.5.jar | None |
| LCHIJA | eleccore | 1.9.452 | ElecCore-1.12.2-1.9.452.jar | None |
| LCHIJA | enderiobase | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduits | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduitsappliedenergistics | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | opencomputers | 1.7.4.153 | OpenComputers-MC1.12.2-1.7.4.153.jar | None |
| LCHIJA | enderioconduitsopencomputers | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduitsrefinedstorage | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiointegrationforestry | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiointegrationticlate | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | ftblib | 5.4.3.132 | FTBLib-5.4.3.132.jar | None |
| LCHIJA | enderiomachines | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiopowertools | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderstorage | 2.4.6.137 | EnderStorage-1.12.2-2.4.6.137-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | immersiveengineering | 0.12-89 | ImmersiveEngineering-0.12-89.jar | 4cb49fcde3b43048c9889e0a3d083225da926334 |
| LCHIJA | engineersdoors | 0.8.0 | engineers_doors-1.12.2-0.8.0.jar | None |
| LCHIJA | valkyrielib | 1.12.2-2.0.20.1 | valkyrielib-1.12.2-2.0.20.1.jar | None |
| LCHIJA | environmentaltech | 1.12.2-2.0.20.1 | environmentaltech-1.12.2-2.0.20.1.jar | None |
| LCHIJA | etlunar | 1.12.2-2.0.20.1 | etlunar-1.12.2-2.0.20.1.jar | None |
| LCHIJA | mcmultipart | 2.5.3 | MCMultiPart-2.5.3.jar | None |
| LCHIJA | mekanism | 1.12.2-9.8.1.383 | Mekanism-1.12.2-9.8.1.383.jar | None |
| LCHIJA | exchangers | 1.12.2-2.8 | Exchangers-1.12.2-2.8.jar | 4ffa87db52cf086d00ecc4853a929367b1c39b5c |
| LCHIJA | extracells | 2.6.2 | ExtraCells-1.12.2-2.6.2a.jar | None |
| LCHIJA | shadowmc | 3.8.0 | ShadowMC-1.12-3.8.0.jar | None |
| LCHIJA | extrarails | 1.3.0 | ExtraRails-1.12-1.3.0.jar | None |
| LCHIJA | extrautils2 | 1.0 | extrautils2-1.12-1.9.9.jar | None |
| LCHIJA | zerocore | 1.12.2-0.1.2.8 | zerocore-1.12.2-0.1.2.8.jar | None |
| LCHIJA | bigreactors | 1.12.2-0.4.5.67 | ExtremeReactors-1.12.2-0.4.5.67.jar | None |
| LCHIJA | fencejumper | 1.0.5 | fencejumper-1.12-1.0.5.jar | None |
| LCHIJA | flatcoloredblocks | mc1.12-6.8 | flatcoloredblocks-mc1.12-6.8.jar | None |
| LCHIJA | sonarcore | 5.0.19 | sonarcore-1.12.2-5.0.19-20.jar | None |
| LCHIJA | fluxnetworks | 3.0.19 | fluxnetworks-1.12.2-3.0.19-21.jar | None |
| LCHIJA | foamfix | 0.10.5-1.12.2 | foamfix-0.10.5-1.12.2.jar | None |
| LCHIJA | forgelin | 1.8.3 | Forgelin-1.8.3.jar | None |
| LCHIJA | forgemultipartcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | microblockcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | None |
| LCHIJA | minecraftmultipartcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | None |
| LCHIJA | ftbguides | 1.0.2.33 | FTBGuides-1.0.2.33.jar | None |
| LCHIJA | ftbutilities | 5.4.0.102 | FTBUtilities-5.4.0.102.jar | None |
| LCHIJA | ftbbackups | 0.0.0.ftbbackups | FTBUtilitiesBackups-1.0.0.5.jar | None |
| LCHIJA | modularrouters | 1.12.2-3.2.2 | modular-routers-1.12.2-3.2.2.jar | None |
| LCHIJA | guideapi | 1.12-2.1.8-63 | Guide-API-1.12-2.1.8-63.jar | None |
| LCHIJA | harvest | 1.12-1.2.7-20 | Harvest-1.12-1.2.7-20.jar | None |
| LCHIJA | ichunutil | 7.2.1 | iChunUtil-1.12.2-7.2.1.jar | 4db5c2bd1b556f252a5b8b54b256d381b2a0a6b8 |
| LCHIJA | immersivepetroleum | 1.1.9 | immersivepetroleum-1.12.2-1.1.9.jar | None |
| LCHIJA | immersivetech | 1.3.10 | immersivetech-1.12-1.3.10.jar | None |
| LCHIJA | teslacorelib | 1.0.15 | tesla-core-lib-1.12.2-1.0.15.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | industrialforegoing | 1.12.2-1.12.2 | industrialforegoing-1.12.2-1.12.13-237.jar | None |
| LCHIJA | initialinventory | 2.0.2 | InitialInventory-3.0.0.jar | None |
| LCHIJA | natura | 1.12.2-4.3.2.69 | natura-1.12.2-4.3.2.69.jar | None |
| LCHIJA | openterraingenerator | v6 | OpenTerrainGenerator-1.12.2 - v6.jar | e9f7847a78c5342af5b0a9e04e5abc0b554d69e0 |
| LCHIJA | nuclearcraft | 2.18f | NuclearCraft-2.18f-1.12.2.jar | None |
| LCHIJA | mcjtylib_ng | 3.5.4 | mcjtylib-1.12-3.5.4.jar | None |
| LCHIJA | rftools | 7.72 | rftools-1.12-7.72.jar | None |
| LCHIJA | rustic | 1.1.1 | rustic-1.1.1.jar | None |
| LCHIJA | integrationforegoing | 1.12.2-1.9 | IntegrationForegoing-1.12.2-1.9.jar | 4ffa87db52cf086d00ecc4853a929367b1c39b5c |
| LCHIJA | inventorytweaks | 1.63+release.109.220f184 | InventoryTweaks-1.63.jar | 55d2cd4f5f0961410bf7b91ef6c6bf00a766dcbe |
| LCHIJA | ironbackpacks | 1.12.2-3.0.8-12 | IronBackpacks-1.12.2-3.0.8-12.jar | None |
| LCHIJA | ironchest | 1.12.2-7.0.67.844 | ironchest-1.12.2-7.0.72.847.jar | None |
| LCHIJA | jaopca | 1.12.2-2.2.8.103 | JAOPCA-1.12.2-2.2.8.103.jar | None |
| LCHIJA | oredictinit | 1.12.2-2.2.1.71 | JAOPCA-1.12.2-2.2.8.103.jar | None |
| LCHIJA | journeymap | 1.12.2-5.5.5 | journeymap-1.12.2-5.5.5.jar | None |
| LCHIJA | kleeslabs | 5.4.12 | KleeSlabs_1.12.2-5.4.12.jar | None |
| LCHIJA | laggoggles | FAT-1.12.2-4.8 | LagGoggles-FAT-1.12.2-4.8.jar | None |
| LCHIJA | letsencryptcraft | @VERSION@ | letsencryptcraft-1.10.2-1.2.0.jar | None |
| LCHIJA | longfallboots | 1.2.1a | longfallboots-1.2.1b.jar | None |
| LCHIJA | lostcities | 2.0.21 | lostcities-1.12-2.0.21.jar | None |
| LCHIJA | mekanismgenerators | 1.12.2-9.8.1.383 | MekanismGenerators-1.12.2-9.8.1.383.jar | None |
| LCHIJA | minetogether | unspecified | minetogether-1.10.2-2.2.0-RC6.jar | None |
| LCHIJA | minetogetherserver | unspecified | minetogether-1.10.2-2.2.0-RC6.jar | None |
| LCHIJA | morpheus | 1.12.2-3.5.106 | Morpheus-1.12.2-3.5.106.jar | None |
| LCHIJA | mrtjpcore | 2.1.4.43 | MrTJPCore-1.12.2-2.1.4.43-universal.jar | None |
| LCHIJA | naturescompass | 1.5.1 | NaturesCompass-1.12.2-1.5.1.jar | None |
| LCHIJA | netherportalfix | 5.3.17 | NetherPortalFix_1.12.1-5.3.17.jar | None |
| LCHIJA | nice | 0.4.0 | nice-1.12-0.4.0.jar | None |
| LCHIJA | neid | 1.5.4.4 | NotEnoughIDs-1.5.4.4.jar | None |
| LCHIJA | notenoughwands | 1.8.1 | notenoughwands-1.12-1.8.1.jar | None |
| LCHIJA | openmods | 0.12.2 | OpenModsLib-1.12.2-0.12.2.jar | d2a9a8e8440196e26a268d1f3ddc01b2e9c572a5 |
| LCHIJA | openblocks | 1.8.1 | OpenBlocks-1.12.2-1.8.1.jar | d2a9a8e8440196e26a268d1f3ddc01b2e9c572a5 |
| LCHIJA | placebo | 1.6.0 | Placebo-1.12.2-1.6.0.jar | None |
| LCHIJA | shetiphiancore | 3.5.9 | shetiphiancore-1.12.0-3.5.9.jar | None |
| LCHIJA | platforms | 1.4.6 | platforms-1.12.0-1.4.6.jar | None |
| LCHIJA | portalgun | 7.1.0 | PortalGun-1.12.2-7.1.0.jar | 4db5c2bd1b556f252a5b8b54b256d381b2a0a6b8 |
| LCHIJA | projectred-core | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-Base.jar | None |
| LCHIJA | projectred-integration | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-integration.jar | None |
| LCHIJA | projectred-transmission | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-integration.jar | None |
| LCHIJA | projectred-fabrication | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-fabrication.jar | None |
| LCHIJA | projectred-illumination | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-lighting.jar | None |
| LCHIJA | projectred-expansion | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-relocation | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-transportation | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-exploration | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-world.jar | None |
| LCHIJA | quickleafdecay | 1.2.4 | QuickLeafDecay-MC1.12.1-1.2.4.jar | None |
| LCHIJA | rangedpumps | 0.5 | rangedpumps-0.5.jar | None |
| LCHIJA | reborncore | 3.16.0.469 | RebornCore-1.12.2-3.16.0.469-universal.jar | 8727a3141c8ec7f173b87aa78b9b9807867c4e6b |
| LCHIJA | redstonearsenal | 2.6.3 | RedstoneArsenal-1.12.2-2.6.3.18-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | rftoolscontrol | 2.0.2 | rftoolsctrl-1.12-2.0.2.jar | None |
| LCHIJA | rftoolspower | 1.2.0 | rftoolspower-1.12-1.2.0.jar | None |
| LCHIJA | signals | 1.4.1-30 | Signals-1.12.2-1.4.1-30-universal.jar | None |
| LCHIJA | simplevoidworld | 1.2.0.9 | Simple-Void-World-1.12-1.2.0.9-universal.jar | None |
| LCHIJA | snad | 1.12.1-1.7.09.16a | Snad-1.12.1-1.7.09.16a.jar | None |
| LCHIJA | storagedrawers | 1.12-5.3.5 | StorageDrawers-1.12.2-5.3.8.jar | None |
| LCHIJA | storagedrawersextra | @VERSION@ | StorageDrawersExtras-1.12-3.1.0.jar | None |
| LCHIJA | thermalcultivation | 0.3.3 | ThermalCultivation-1.12.2-0.3.3.12-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermaldynamics | 2.5.5 | ThermalDynamics-1.12.2-2.5.5.21-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermalinnovation | 0.3.3 | ThermalInnovation-1.12.2-0.3.3.12-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thutcore | 5.21.0 | thutcore-1.12.2-5.21.0.jar | None |
| LCHIJA | thutcore_compat | 1.0 | thutcore-1.12.2-5.21.0.jar | None |
| LCHIJA | thuttech | 6.1.6 | thuttech-1.12.2-6.1.6.jar | None |
| LCHIJA | tinkertoolleveling | 1.12.2-1.1.0.DEV.b23e769 | TinkerToolLeveling-1.12.2-1.1.0.jar | None |
| LCHIJA | topaddons | 1.12.2-1.11.0 | topaddons-1.12.2-1.11.0.jar | None |
| LCHIJA | torchmaster | 1.8.0.79 | torchmaster_1.12.2-1.8.0.79.jar | 5e9a436b366831c8f54a7e80b015784da69278c6 |
| LCHIJA | translocators | 2.5.2.81 | Translocators-1.12.2-2.5.2.81-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | traverse | 1.6.0 | Traverse-1.12.2-1.6.0-69.jar | None |
| LCHIJA | universalmodifiers | 1.12.2-1.0.16.1 | valkyrielib-1.12.2-2.0.20.1.jar | None |
| LCHIJA | vanillavistas | 1.0 | Vanilla_Vistas 1.1.jar | None |
| LCHIJA | wanionlib | 1.12.2-2.2 | WanionLib-1.12.2-2.2.jar | None |
| LCHIJA | waystones | 4.0.72 | Waystones_1.12.2-4.0.72.jar | None |
| LCHIJA | wct | 3.12.93 | WirelessCraftingTerminal-1.12.2-3.12.93.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | woot | 1.12.2-1.4.11 | woot-1.12.2-1.4.11.jar | None |
| LCHIJA | worldedit | 6.1.10-SNAPSHOT | worldedit-forge-mc1.12.2-6.1.10-SNAPSHOT-dist.jar | None |
| LCHIJA | wrcbe | 2.3.2 | WR-CBE-1.12.2-2.3.2.33-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | xnet | 1.8.2 | xnet-1.12-1.8.2.jar | None |
| LCHIJA | xtones | 1.12-1.0.8-11 | Xtones-1.12-1.0.8-11.jar | None |
| LCHIJA | yabba | 1.1.2.53 | YABBA-1.1.2.53.jar | None |
| LCHIJA | structurize | 1.12.2-0.10.124-ALPHA | structurize-1.12.2-0.10.124-ALPHA.jar | None |
| LCHIJA | minecolonies | 1.12.2-0.10.348-RELEASE | minecolonies-1.12.2-0.10.348-RELEASE-universal.jar | None |
| LCHIJA | industrialwires | 1.7-36 | IndustrialWires-1.7-36.jar | 7e11c175d1e24007afec7498a1616bef0000027d |
| LCHIJA | orelib | 3.5.2.2 | OreLib-1.12.2-3.5.2.2.jar | 7a2128d395ad96ceb9d9030fbd41d035b435753a |
| LCHIJA | dsurround | 3.5.4.3 | DynamicSurroundings-1.12.2-3.5.4.3.jar | 7a2128d395ad96ceb9d9030fbd41d035b435753a |
| LCHIJA | eleccoreloader | 1.9.452 | ElecCore-1.12.2-1.9.452.jar | None |
| LCHIJA | mysticallib | 1.12.2-1.3.4 | mysticallib-1.12.2-1.3.4.jar | None |
| LCHIJA | teslacorelib_registries | 1.0.15 | tesla-core-lib-1.12.2-1.0.15.jar | None |
| LCHIJA | unidict | 1.12.2-2.9.2 | UniDict-1.12.2-2.9.2.jar | None |
| LCHIJA | wrapup | 1.12-1.1.3 | WrapUp-1.12-1.1.3.jar | None |
Loaded coremods (and transformers):
RandomPatches (randompatches-1.12.2-1.18.1.1.jar)
com.therandomlabs.randompatches.core.RPTransformer
SplashAnimationCoremod (SplashAnimation-0.2.1.jar)
pl.asie.splashanimation.core.SplashProgressTransformer
BedPatch (bedpatch-2.2-1.12.2.jar)
com.mordenkainen.bedpatch.BedPatchASM
Do not report to Forge! (If you haven't disabled the FoamFix coremod, try disabling it in the config! Note that this bit of text will still appear.) (foamfix-0.10.5-1.12.2.jar)
pl.asie.foamfix.coremod.FoamFixTransformer
Quark Plugin (Quark-r1.6-177.jar)
vazkii.quark.base.asm.ClassTransformer
OTGCorePlugin (OTG-Core.jar)
com.pg85.otg.forge.asm.OTGClassTransformer
OpenModsCorePlugin (OpenModsLib-1.12.2-0.12.2.jar)
openmods.core.OpenModsClassTransformer
TransformerLoader (OpenComputers-MC1.12.2-1.7.4.153.jar)
li.cil.oc.common.asm.ClassTransformer
EnderCorePlugin (EnderCore-1.12.2-0.5.65-core.jar)
com.enderio.core.common.transform.EnderCoreTransformer
com.enderio.core.common.transform.SimpleMixinPatcher
CTMCorePlugin (CTM-MC1.12.2-0.3.3.22.jar)
team.chisel.ctm.client.asm.CTMTransformer
p455w0rdc0re (p455w0rdslib-1.12.2-2.2.156.jar)
p455w0rdslib.asm.ClassTransformer
ForgelinPlugin (Forgelin-1.8.3.jar)
Plugin (NotEnoughIDs-1.5.4.4.jar)
ru.fewizz.neid.asm.Transformer
CorePlugin (ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar)
IELoadingPlugin (ImmersiveEngineering-core-0.12-89.jar)
blusunrize.immersiveengineering.common.asm.IEClassTransformer
Inventory Tweaks Coremod (InventoryTweaks-1.63.jar)
invtweaks.forge.asm.ContainerTransformer
ShetiPhian-ASM (ShetiPhian-ASM-1.12.0.jar)
shetiphian.asm.ClassTransformer
MekanismCoremod (Mekanism-1.12.2-9.8.1.383.jar)
mekanism.coremod.KeybindingMigrationHelper
LoadingPlugin (ResourceLoader-MC1.12.1-1.5.3.jar)
lumien.resourceloader.asm.ClassTransformer
CoreMod (Aroma1997Core-1.12.2-2.0.0.2.b165.jar)
DynamicSurroundingsCore (DynamicSurroundings-core-1.12.2-3.5.4.3.jar)
org.orecruncher.dsurround.asm.Transformer
AdvancedRocketryPlugin (AdvancedRocketry-1.12.2-1.6.0-164-universal.jar)
zmaster587.advancedRocketry.asm.ClassTransformer
OpenModsLib class transformers: [llama_null_fix:FINISHED],[horse_base_null_fix:FINISHED],[pre_world_render_hook:ENABLED],[player_render_hook:ENABLED],[horse_null_fix:FINISHED]
AE2 Version: stable rv6-stable-7 for Forge 14.23.5.2768
Pulsar/tconstruct loaded Pulses:
- TinkerCommons (Enabled/Forced)
- TinkerWorld (Enabled/Not Forced)
- TinkerTools (Enabled/Not Forced)
- TinkerHarvestTools (Enabled/Forced)
- TinkerMeleeWeapons (Enabled/Forced)
- TinkerRangedWeapons (Enabled/Forced)
- TinkerModifiers (Enabled/Forced)
- TinkerSmeltery (Enabled/Not Forced)
- TinkerGadgets (Enabled/Not Forced)
- TinkerOredict (Enabled/Forced)
- TinkerIntegration (Enabled/Forced)
- TinkerFluids (Enabled/Forced)
- TinkerMaterials (Enabled/Forced)
- TinkerModelRegister (Enabled/Forced)
- chiselIntegration (Enabled/Not Forced)
- chiselsandbitsIntegration (Enabled/Not Forced)
- theoneprobeIntegration (Enabled/Not Forced)
Pulsar/natura loaded Pulses:
- NaturaCommons (Enabled/Forced)
- NaturaOverworld (Enabled/Not Forced)
- NaturaNether (Enabled/Not Forced)
- NaturaDecorative (Enabled/Not Forced)
- NaturaTools (Enabled/Not Forced)
- NaturaEntities (Enabled/Not Forced)
- NaturaOredict (Enabled/Forced)
- NaturaWorld (Enabled/Not Forced)
List of loaded APIs:
* actuallyadditionsapi (34) from ActuallyAdditions-1.12.2-r149.jar
* ae2wtlib|API (1.0.30) from AE2WTLib-1.12.2-1.0.30.jar
* AgriCraftAPI (1.0) from AgriCraft-2.12.0-1.12.0-a6.jar
* appliedenergistics2|API (rv6) from appliedenergistics2-rv6-stable-7.jar
* Base|API (1.0.0) from base-1.12.2-3.13.0.jar
* Baubles|API (1.4.0.2) from Baubles-1.12-1.5.2.jar
* betteradvancements|API (0.1.0.77) from BetterAdvancements-1.12.2-0.1.0.77.jar
* BetterWithModsAPI (Beta 0.6) from AppleSkin-mc1.12-1.0.9.jar
* bigreactors|API (4.0.1) from ExtremeReactors-1.12.2-0.4.5.67.jar
* BotaniaAPI (79) from AkashicTome-1.2-12.jar
* chesttransporter|api (2) from ChestTransporter-1.12.2-2.8.8.jar
* Chisel-API (0.0.1) from Chisel-MC1.12.2-0.2.1.35.jar
* ChiselAPI|Carving (0.0.1) from Chisel-MC1.12.2-0.2.1.35.jar
* ChiselsAndBitsAPI (14.25.0) from chiselsandbits-14.33.jar
* cofhapi (2.5.0) from CoFHCore-1.12.2-4.6.3.27-universal.jar
* ctm-api (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-events (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-models (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-textures (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-utils (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* DraconicEvolution|API (1.3) from Draconic-Evolution-1.12.2-2.3.24.349-universal.jar
* ElecCoreAPI (1.0.0) from ElecCore-1.12.2-1.9.452.jar
* enderioapi (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|addon (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|capacitor (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|conduits (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|farm (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|redstone (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|teleport (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|tools (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|upgrades (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* fluxapi (1.0) from fluxnetworks-1.12.2-3.0.19-21.jar
* ForgeEndertechAPI (1.0) from ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar
* Guide-API|API (2.0.0) from Guide-API-1.12-2.1.8-63.jar
* iChunUtil API (1.2.0) from iChunUtil-1.12.2-7.2.1.jar
* ImmersiveEngineering|API (1.0) from ImmersiveEngineering-0.12-89.jar
* ImmersiveEngineering|ImmersiveFluxAPI (1.0) from ImmersiveEngineering-0.12-89.jar
* industrialforegoingapi (5) from industrialforegoing-1.12.2-1.12.13-237.jar
* jeresources|API (0.9.2.60) from JustEnoughResources-1.12.2-0.9.2.60.jar
* journeymap|client-api (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-display (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-event (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-model (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-util (1.4) from journeymap-1.12.2-5.5.5.jar
* JustEnoughItemsAPI (4.13.0) from jei_1.12.2-4.15.0.289.jar
* MekanismAPI|core (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|energy (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|gas (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|infuse (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|laser (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|transmitter (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|util (9.0.0) from Mekanism-1.12.2-9.8.1.383.jar
* minecolonies-api (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|achievements (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|decorative (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|huts (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|interfaces (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|types (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render|modeltype (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render|modeltype|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|views (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|workerbuildings (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|guardtype (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|guardtype|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|jobs (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|jobs|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|managers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|managers|interfaces (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|permissions (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|data (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|factory (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|factory|standard (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|location (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|manager (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|request (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requestable (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requestable|crafting (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requester (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver|player (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver|retrying (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|token (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|workorders (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|candb (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|dynamictrees (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|gbook (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|tinkers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|configuration (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|crafting (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|creativetab (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen|builder (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen|guards (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|pathfinding (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|basestatemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|states (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|tickratestatemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|transition (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|citizen (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|citizen|citizenhandlers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|barbarians (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|pirates (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|util (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|pathfinding (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|pathfinding|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|inventory (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|inventory|api (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|items (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|network (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|sounds (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|tileentities (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|util (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|util|constants (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout|controls (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout|views (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* modtogether|api (1.0) from minetogether-1.10.2-2.2.0-RC6.jar
* MouseTweaks|API (1.0) from MouseTweaks-2.10-mc1.12.2.jar
* openblocks|api (1.2) from OpenBlocks-1.12.2-1.8.1.jar
* opencomputersapi|component (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|core (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|driver (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|driver|item (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|event (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|filesystem (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|internal (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|machine (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|manual (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|network (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|prefab (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* projectred|api (2.1) from ProjectRed-1.12.2-4.9.4.120-Base.jar
* QuarkAPI (4) from Quark-r1.6-177.jar
* reborncoreAPI (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Power (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Praescriptum (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Recipe (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Tile (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* redstonefluxapi (2.1.0) from RedstoneFlux-1.12-2.1.0.6-universal.jar
* sonarapi (1.0.1) from sonarcore-1.12.2-5.0.19-20.jar
* StorageDrawersAPI (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|event (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|registry (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|render (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|storage (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|storage-attribute (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* theoneprobe_api (1.4.4) from theoneprobe-1.12-1.4.28.jar
* thut|API (1.0.0) from thutcore-1.12.2-5.21.0.jar
* valkyrielib.api (1.12.2-2.0.10a) from valkyrielib-1.12.2-2.0.20.1.jar
* wct|api (1.1) from WirelessCraftingTerminal-1.12.2-3.12.93.jar
* zerocore|API|multiblock (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|rectangular (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|tier (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|validation (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
RebornCore:
Plugin Engine: 0
RebornCore Version: 3.16.0.469
Runtime Debofucsation 1
Ender IO: No known problems detected.
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
!!!You are looking at the diagnostics information, not at the crash. !!!
!!!Scroll up until you see the line with '---- Minecraft Crash Report ----'!!!
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
AE2 Integration: IC2:OFF, RC:OFF, MFR:OFF, Waila:OFF, Mekanism:ON, OpenComputers:ON, THE_ONE_PROBE:ON, TESLA:OFF, CRAFTTWEAKER:ON
Profiler Position: N/A (disabled)
Player Count: 1 / 1337; [EntityPlayerMP['firemx'/566, l='VanillaVistas', x=115.22, y=85.79, z=-1632.99]]
Is Modded: Definitely; Server brand changed to 'fml,forge'
Type: Dedicated Server (map_server.txt)
|
1.0
|
[MC1.12.2 - 1.18.1.1] TickNextTick error still occurs, world crashing - ---- Minecraft Crash Report ----
WARNING: coremods are present:
RandomPatches (randompatches-1.12.2-1.18.1.1.jar)
SplashAnimationCoremod (SplashAnimation-0.2.1.jar)
BedPatch (bedpatch-2.2-1.12.2.jar)
Do not report to Forge! (If you haven't disabled the FoamFix coremod, try disabling it in the config! Note that this bit of text will still appear.) (foamfix-0.10.5-1.12.2.jar)
Quark Plugin (Quark-r1.6-177.jar)
OTGCorePlugin (OTG-Core.jar)
OpenModsCorePlugin (OpenModsLib-1.12.2-0.12.2.jar)
TransformerLoader (OpenComputers-MC1.12.2-1.7.4.153.jar)
EnderCorePlugin (EnderCore-1.12.2-0.5.65-core.jar)
CTMCorePlugin (CTM-MC1.12.2-0.3.3.22.jar)
p455w0rdc0re (p455w0rdslib-1.12.2-2.2.156.jar)
ForgelinPlugin (Forgelin-1.8.3.jar)
Plugin (NotEnoughIDs-1.5.4.4.jar)
CorePlugin (ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar)
IELoadingPlugin (ImmersiveEngineering-core-0.12-89.jar)
Inventory Tweaks Coremod (InventoryTweaks-1.63.jar)
ShetiPhian-ASM (ShetiPhian-ASM-1.12.0.jar)
MekanismCoremod (Mekanism-1.12.2-9.8.1.383.jar)
LoadingPlugin (ResourceLoader-MC1.12.1-1.5.3.jar)
CoreMod (Aroma1997Core-1.12.2-2.0.0.2.b165.jar)
DynamicSurroundingsCore (DynamicSurroundings-core-1.12.2-3.5.4.3.jar)
AdvancedRocketryPlugin (AdvancedRocketry-1.12.2-1.6.0-164-universal.jar)
Contact their authors BEFORE contacting forge
// Don't do that.
Time: 9/12/19 6:44 PM
Description: Exception ticking world
java.lang.IllegalStateException: TickNextTick list out of synch
at net.minecraft.world.WorldServer.func_72955_a(WorldServer.java:691)
at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:223)
at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:756)
at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397)
at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668)
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526)
at java.lang.Thread.run(Thread.java:748)
A detailed walkthrough of the error, its code path and all known details is as follows:
---------------------------------------------------------------------------------------
-- Head --
Thread: Server thread
Stacktrace:
at net.minecraft.world.WorldServer.func_72955_a(WorldServer.java:691)
at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:223)
-- Affected level --
Details:
Level name: VanillaVistas
All players: 1 total; [EntityPlayerMP['firemx'/566, l='VanillaVistas', x=115.22, y=85.79, z=-1632.99]]
Chunk stats: ServerChunkCache: 1357 Drop: 0
Level seed: -8290667671068293349
Level generator: ID 11 - OTG, ver 0. Features enabled: true
Level generator options: OpenTerrainGenerator
Level spawn location: World: (365,64,-1295), Chunk: (at 13,4,1 in 22,-81; contains blocks 352,0,-1296 to 367,255,-1281), Region: (0,-3; contains chunks 0,-96 to 31,-65, blocks 0,0,-1536 to 511,255,-1025)
Level time: 20033484 game time, 2840813 day time
Level dimension: 0
Level storage version: 0x04ABD - Anvil
Level weather: Rain time: 40550 (now: false), thunder time: 13025 (now: true)
Level game mode: Game mode: survival (ID 0). Hardcore: false. Cheats: false
Stacktrace:
at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:756)
at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397)
at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668)
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526)
at java.lang.Thread.run(Thread.java:748)
-- System Details --
Details:
Minecraft Version: 1.12.2
Operating System: Linux (amd64) version 3.10.0-957.21.3.el7.x86_64
Java Version: 1.8.0_212, Oracle Corporation
Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Oracle Corporation
Memory: 5802951920 bytes (5534 MB) / 7057965056 bytes (6731 MB) up to 7057965056 bytes (6731 MB)
JVM Flags: 3 total; -Xmx7168M -Xms7168M -XX:MaxPermSize=256m
IntCache: cache: 0, tcache: 0, allocated: 0, tallocated: 0
FML: MCP 9.42 Powered by Forge 14.23.5.2838 207 mods loaded, 207 mods active
States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored
| State | ID | Version | Source | Signature |
|:------ |:--------------------------------- |:------------------------ |:--------------------------------------------------- |:---------------------------------------- |
| LCHIJA | minecraft | 1.12.2 | minecraft.jar | None |
| LCHIJA | mcp | 9.42 | minecraft.jar | None |
| LCHIJA | FML | 8.0.99.99 | forge1122.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 |
| LCHIJA | forge | 14.23.5.2838 | forge1122.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 |
| LCHIJA | advancedrocketrycore | 1 | minecraft.jar | None |
| LCHIJA | otgcore | 1.12.2 - v7 | minecraft.jar | None |
| LCHIJA | openmodscore | 0.12.2 | minecraft.jar | None |
| LCHIJA | foamfixcore | 7.7.4 | minecraft.jar | None |
| LCHIJA | opencomputers|core | 1.7.4.153 | minecraft.jar | None |
| LCHIJA | dsurroundcore | 3.5.4.3 | minecraft.jar | None |
| LCHIJA | randompatches | 1.12.2-1.18.1.1 | randompatches-1.12.2-1.18.1.1.jar | None |
| LCHIJA | fastbench | 1.7.2 | FastWorkbench-1.12.2-1.7.2.jar | None |
| LCHIJA | actuallyadditions | 1.12.2-r149 | ActuallyAdditions-1.12.2-r149.jar | None |
| LCHIJA | baubles | 1.5.2 | Baubles-1.12-1.5.2.jar | None |
| LCHIJA | actuallybaubles | 1.1 | ActuallyBaubles-1.12-1.1.jar | None |
| LCHIJA | forgeendertech | 1.12.2-4.4.22.2 | ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar | None |
| LCHIJA | adlods | 1.12.2-1.0.6.2 | AdLods-1.12.2-1.0.6.2-build.0408.jar | None |
| LCHIJA | redstoneflux | 2.1.0 | RedstoneFlux-1.12-2.1.0.6-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | cofhcore | 4.6.3 | CoFHCore-1.12.2-4.6.3.27-universal.jar | None |
| LCHIJA | libvulpes | 0.4.1.-62 | LibVulpes-1.12.2-0.4.1-62-universal.jar | None |
| LCHIJA | advancedrocketry | 1.6.0.-164 | AdvancedRocketry-1.12.2-1.6.0-164-universal.jar | None |
| LCHIJA | appliedenergistics2 | rv6-stable-7 | appliedenergistics2-rv6-stable-7.jar | dfa4d3ac143316c6f32aa1a1beda1e34d42132e5 |
| LCHIJA | bdlib | 1.14.3.12 | bdlib-1.14.3.12-mc1.12.2.jar | None |
| LCHIJA | ae2stuff | 0.7.0.4 | ae2stuff-0.7.0.4-mc1.12.2.jar | None |
| LCHIJA | endercore | 1.12.2-0.5.65 | EnderCore-1.12.2-0.5.65.jar | None |
| LCHIJA | crafttweaker | 4.1.19 | CraftTweaker2-1.12-4.1.19.jar | None |
| LCHIJA | mtlib | 3.0.6 | MTLib-3.0.6.jar | None |
| LCHIJA | modtweaker | 4.0.17 | modtweaker-4.0.17.jar | None |
| LCHIJA | jei | 4.15.0.289 | jei_1.12.2-4.15.0.289.jar | None |
| LCHIJA | codechickenlib | 3.2.3.357 | CodeChickenLib-1.12.2-3.2.3.357-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | brandonscore | 2.4.17 | BrandonsCore-1.12.2-2.4.17.208-universal.jar | None |
| LCHIJA | cofhworld | 1.3.1 | CoFHWorld-1.12.2-1.3.1.7-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermalfoundation | 2.6.3 | ThermalFoundation-1.12.2-2.6.3.27-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | draconicevolution | 2.3.24 | Draconic-Evolution-1.12.2-2.3.24.349-universal.jar | None |
| LCHIJA | thermalexpansion | 5.5.4 | ThermalExpansion-1.12.2-5.5.4.43-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | enderio | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | mantle | 1.12-1.3.3.55 | Mantle-1.12-1.3.3.55.jar | None |
| LCHIJA | chisel | MC1.12.2-0.2.1.35 | Chisel-MC1.12.2-0.2.1.35.jar | None |
| LCHIJA | enderiointegrationtic | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | tconstruct | 1.12.2-2.12.0.157 | TConstruct-1.12.2-2.12.0.157.jar | None |
| LCHIJA | p455w0rdslib | 2.2.156 | p455w0rdslib-1.12.2-2.2.156.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | ae2wtlib | 1.0.30 | AE2WTLib-1.12.2-1.0.30.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | infinitylib | 1.12.2-1.12.0 | infinitylib-1.12.0.jar | None |
| LCHIJA | agricraft | 2.12.0-1.12.0-a6 | AgriCraft-2.12.0-1.12.0-a6.jar | None |
| LCHIJA | akashictome | 1.2-12 | AkashicTome-1.2-12.jar | None |
| LCHIJA | appleskin | 1.0.9 | AppleSkin-mc1.12-1.0.9.jar | None |
| LCHIJA | architecturecraft | @VERSION@ | architecturecraft-1.12-3.98.jar | None |
| LCHIJA | aroma1997core | 2.0.0.2.b165 | Aroma1997Core-1.12.2-2.0.0.2.b165.jar | dfbfe4c473253d8c5652417689848f650b2cbe32 |
| LCHIJA | aroma1997sdimension | 2.0.0.2.b89 | Aroma1997s-Dimensional-World-1.12.2-2.0.0.2.b89.jar | dfbfe4c473253d8c5652417689848f650b2cbe32 |
| LCHIJA | morphtool | 1.2-21 | Morph-o-Tool-1.2-21.jar | None |
| LCHIJA | quark | r1.6-177 | Quark-r1.6-177.jar | None |
| LCHIJA | autoreglib | 1.3-32 | AutoRegLib-1.3-32.jar | None |
| LCHIJA | badwithernocookiereloaded | 1.12.2-3.3.16 | badwithernocookiereloaded-1.12.2-3.3.16.jar | None |
| LCHIJA | base | 3.13.0 | base-1.12.2-3.13.0.jar | None |
| LCHIJA | bedpatch | 2.2 | bedpatch-2.2-1.12.2.jar | 6bf7527e690fb5e8719b9832bce5000a3e87dfe6 |
| LCHIJA | betterbuilderswands | 0.13.2 | BetterBuildersWands-1.12.2-0.13.2.271+5997513.jar | None |
| LCHIJA | blockcraftery | 1.12.2-1.3.1 | blockcraftery-1.12.2-1.3.1.jar | None |
| LCHIJA | bookshelf | 2.3.585 | Bookshelf-1.12.2-2.3.585.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | botania | r1.10-361 | Botania r1.10-361.jar | None |
| LCHIJA | chameleon | 1.12-4.1.3 | Chameleon-1.12-4.1.3.jar | None |
| LCHIJA | chesttransporter | 2.8.8 | ChestTransporter-1.12.2-2.8.8.jar | None |
| LCHIJA | chiselsandbits | 14.33 | chiselsandbits-14.33.jar | None |
| LCHIJA | clumps | 3.1.2 | Clumps-3.1.2.jar | None |
| LCHIJA | compactmachines3 | 3.0.18 | compactmachines3-1.12.2-3.0.18-b278.jar | None |
| LCHIJA | conarm | 1.2.4 | conarm-1.12.2-1.2.4.jar | 5d5b8aee896a4f5ea3f3114784742662a67ad32f |
| LCHIJA | cookingforblockheads | 6.4.71 | CookingForBlockheads_1.12.2-6.4.71.jar | None |
| LCHIJA | crafttweakerjei | 2.0.3 | CraftTweaker2-1.12-4.1.19.jar | None |
| LCHIJA | theoneprobe | 1.4.28 | theoneprobe-1.12-1.4.28.jar | None |
| LCHIJA | danknull | 1.6.77 | DankNull-1.12.2-1.6.77.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | darkutils | 1.8.226 | DarkUtils-1.12.2-1.8.226.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | diethopper | 1.1 | diethopper-1.1.jar | None |
| LCHIJA | dirt2path | 1.7.1 | dirt2path-1.8.0.jar | None |
| LCHIJA | dynamictrees | 1.12.2-0.9.5 | DynamicTrees-1.12.2-0.9.5.jar | None |
| LCHIJA | eleccore | 1.9.452 | ElecCore-1.12.2-1.9.452.jar | None |
| LCHIJA | enderiobase | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduits | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduitsappliedenergistics | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | opencomputers | 1.7.4.153 | OpenComputers-MC1.12.2-1.7.4.153.jar | None |
| LCHIJA | enderioconduitsopencomputers | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderioconduitsrefinedstorage | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiointegrationforestry | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiointegrationticlate | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | ftblib | 5.4.3.132 | FTBLib-5.4.3.132.jar | None |
| LCHIJA | enderiomachines | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderiopowertools | 5.0.50 | EnderIO-1.12.2-5.0.50.jar | None |
| LCHIJA | enderstorage | 2.4.6.137 | EnderStorage-1.12.2-2.4.6.137-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | immersiveengineering | 0.12-89 | ImmersiveEngineering-0.12-89.jar | 4cb49fcde3b43048c9889e0a3d083225da926334 |
| LCHIJA | engineersdoors | 0.8.0 | engineers_doors-1.12.2-0.8.0.jar | None |
| LCHIJA | valkyrielib | 1.12.2-2.0.20.1 | valkyrielib-1.12.2-2.0.20.1.jar | None |
| LCHIJA | environmentaltech | 1.12.2-2.0.20.1 | environmentaltech-1.12.2-2.0.20.1.jar | None |
| LCHIJA | etlunar | 1.12.2-2.0.20.1 | etlunar-1.12.2-2.0.20.1.jar | None |
| LCHIJA | mcmultipart | 2.5.3 | MCMultiPart-2.5.3.jar | None |
| LCHIJA | mekanism | 1.12.2-9.8.1.383 | Mekanism-1.12.2-9.8.1.383.jar | None |
| LCHIJA | exchangers | 1.12.2-2.8 | Exchangers-1.12.2-2.8.jar | 4ffa87db52cf086d00ecc4853a929367b1c39b5c |
| LCHIJA | extracells | 2.6.2 | ExtraCells-1.12.2-2.6.2a.jar | None |
| LCHIJA | shadowmc | 3.8.0 | ShadowMC-1.12-3.8.0.jar | None |
| LCHIJA | extrarails | 1.3.0 | ExtraRails-1.12-1.3.0.jar | None |
| LCHIJA | extrautils2 | 1.0 | extrautils2-1.12-1.9.9.jar | None |
| LCHIJA | zerocore | 1.12.2-0.1.2.8 | zerocore-1.12.2-0.1.2.8.jar | None |
| LCHIJA | bigreactors | 1.12.2-0.4.5.67 | ExtremeReactors-1.12.2-0.4.5.67.jar | None |
| LCHIJA | fencejumper | 1.0.5 | fencejumper-1.12-1.0.5.jar | None |
| LCHIJA | flatcoloredblocks | mc1.12-6.8 | flatcoloredblocks-mc1.12-6.8.jar | None |
| LCHIJA | sonarcore | 5.0.19 | sonarcore-1.12.2-5.0.19-20.jar | None |
| LCHIJA | fluxnetworks | 3.0.19 | fluxnetworks-1.12.2-3.0.19-21.jar | None |
| LCHIJA | foamfix | 0.10.5-1.12.2 | foamfix-0.10.5-1.12.2.jar | None |
| LCHIJA | forgelin | 1.8.3 | Forgelin-1.8.3.jar | None |
| LCHIJA | forgemultipartcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | microblockcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | None |
| LCHIJA | minecraftmultipartcbe | 2.6.2.83 | ForgeMultipart-1.12.2-2.6.2.83-universal.jar | None |
| LCHIJA | ftbguides | 1.0.2.33 | FTBGuides-1.0.2.33.jar | None |
| LCHIJA | ftbutilities | 5.4.0.102 | FTBUtilities-5.4.0.102.jar | None |
| LCHIJA | ftbbackups | 0.0.0.ftbbackups | FTBUtilitiesBackups-1.0.0.5.jar | None |
| LCHIJA | modularrouters | 1.12.2-3.2.2 | modular-routers-1.12.2-3.2.2.jar | None |
| LCHIJA | guideapi | 1.12-2.1.8-63 | Guide-API-1.12-2.1.8-63.jar | None |
| LCHIJA | harvest | 1.12-1.2.7-20 | Harvest-1.12-1.2.7-20.jar | None |
| LCHIJA | ichunutil | 7.2.1 | iChunUtil-1.12.2-7.2.1.jar | 4db5c2bd1b556f252a5b8b54b256d381b2a0a6b8 |
| LCHIJA | immersivepetroleum | 1.1.9 | immersivepetroleum-1.12.2-1.1.9.jar | None |
| LCHIJA | immersivetech | 1.3.10 | immersivetech-1.12-1.3.10.jar | None |
| LCHIJA | teslacorelib | 1.0.15 | tesla-core-lib-1.12.2-1.0.15.jar | d476d1b22b218a10d845928d1665d45fce301b27 |
| LCHIJA | industrialforegoing | 1.12.2-1.12.2 | industrialforegoing-1.12.2-1.12.13-237.jar | None |
| LCHIJA | initialinventory | 2.0.2 | InitialInventory-3.0.0.jar | None |
| LCHIJA | natura | 1.12.2-4.3.2.69 | natura-1.12.2-4.3.2.69.jar | None |
| LCHIJA | openterraingenerator | v6 | OpenTerrainGenerator-1.12.2 - v6.jar | e9f7847a78c5342af5b0a9e04e5abc0b554d69e0 |
| LCHIJA | nuclearcraft | 2.18f | NuclearCraft-2.18f-1.12.2.jar | None |
| LCHIJA | mcjtylib_ng | 3.5.4 | mcjtylib-1.12-3.5.4.jar | None |
| LCHIJA | rftools | 7.72 | rftools-1.12-7.72.jar | None |
| LCHIJA | rustic | 1.1.1 | rustic-1.1.1.jar | None |
| LCHIJA | integrationforegoing | 1.12.2-1.9 | IntegrationForegoing-1.12.2-1.9.jar | 4ffa87db52cf086d00ecc4853a929367b1c39b5c |
| LCHIJA | inventorytweaks | 1.63+release.109.220f184 | InventoryTweaks-1.63.jar | 55d2cd4f5f0961410bf7b91ef6c6bf00a766dcbe |
| LCHIJA | ironbackpacks | 1.12.2-3.0.8-12 | IronBackpacks-1.12.2-3.0.8-12.jar | None |
| LCHIJA | ironchest | 1.12.2-7.0.67.844 | ironchest-1.12.2-7.0.72.847.jar | None |
| LCHIJA | jaopca | 1.12.2-2.2.8.103 | JAOPCA-1.12.2-2.2.8.103.jar | None |
| LCHIJA | oredictinit | 1.12.2-2.2.1.71 | JAOPCA-1.12.2-2.2.8.103.jar | None |
| LCHIJA | journeymap | 1.12.2-5.5.5 | journeymap-1.12.2-5.5.5.jar | None |
| LCHIJA | kleeslabs | 5.4.12 | KleeSlabs_1.12.2-5.4.12.jar | None |
| LCHIJA | laggoggles | FAT-1.12.2-4.8 | LagGoggles-FAT-1.12.2-4.8.jar | None |
| LCHIJA | letsencryptcraft | @VERSION@ | letsencryptcraft-1.10.2-1.2.0.jar | None |
| LCHIJA | longfallboots | 1.2.1a | longfallboots-1.2.1b.jar | None |
| LCHIJA | lostcities | 2.0.21 | lostcities-1.12-2.0.21.jar | None |
| LCHIJA | mekanismgenerators | 1.12.2-9.8.1.383 | MekanismGenerators-1.12.2-9.8.1.383.jar | None |
| LCHIJA | minetogether | unspecified | minetogether-1.10.2-2.2.0-RC6.jar | None |
| LCHIJA | minetogetherserver | unspecified | minetogether-1.10.2-2.2.0-RC6.jar | None |
| LCHIJA | morpheus | 1.12.2-3.5.106 | Morpheus-1.12.2-3.5.106.jar | None |
| LCHIJA | mrtjpcore | 2.1.4.43 | MrTJPCore-1.12.2-2.1.4.43-universal.jar | None |
| LCHIJA | naturescompass | 1.5.1 | NaturesCompass-1.12.2-1.5.1.jar | None |
| LCHIJA | netherportalfix | 5.3.17 | NetherPortalFix_1.12.1-5.3.17.jar | None |
| LCHIJA | nice | 0.4.0 | nice-1.12-0.4.0.jar | None |
| LCHIJA | neid | 1.5.4.4 | NotEnoughIDs-1.5.4.4.jar | None |
| LCHIJA | notenoughwands | 1.8.1 | notenoughwands-1.12-1.8.1.jar | None |
| LCHIJA | openmods | 0.12.2 | OpenModsLib-1.12.2-0.12.2.jar | d2a9a8e8440196e26a268d1f3ddc01b2e9c572a5 |
| LCHIJA | openblocks | 1.8.1 | OpenBlocks-1.12.2-1.8.1.jar | d2a9a8e8440196e26a268d1f3ddc01b2e9c572a5 |
| LCHIJA | placebo | 1.6.0 | Placebo-1.12.2-1.6.0.jar | None |
| LCHIJA | shetiphiancore | 3.5.9 | shetiphiancore-1.12.0-3.5.9.jar | None |
| LCHIJA | platforms | 1.4.6 | platforms-1.12.0-1.4.6.jar | None |
| LCHIJA | portalgun | 7.1.0 | PortalGun-1.12.2-7.1.0.jar | 4db5c2bd1b556f252a5b8b54b256d381b2a0a6b8 |
| LCHIJA | projectred-core | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-Base.jar | None |
| LCHIJA | projectred-integration | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-integration.jar | None |
| LCHIJA | projectred-transmission | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-integration.jar | None |
| LCHIJA | projectred-fabrication | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-fabrication.jar | None |
| LCHIJA | projectred-illumination | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-lighting.jar | None |
| LCHIJA | projectred-expansion | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-relocation | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-transportation | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-mechanical.jar | None |
| LCHIJA | projectred-exploration | 4.9.4.120 | ProjectRed-1.12.2-4.9.4.120-world.jar | None |
| LCHIJA | quickleafdecay | 1.2.4 | QuickLeafDecay-MC1.12.1-1.2.4.jar | None |
| LCHIJA | rangedpumps | 0.5 | rangedpumps-0.5.jar | None |
| LCHIJA | reborncore | 3.16.0.469 | RebornCore-1.12.2-3.16.0.469-universal.jar | 8727a3141c8ec7f173b87aa78b9b9807867c4e6b |
| LCHIJA | redstonearsenal | 2.6.3 | RedstoneArsenal-1.12.2-2.6.3.18-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | rftoolscontrol | 2.0.2 | rftoolsctrl-1.12-2.0.2.jar | None |
| LCHIJA | rftoolspower | 1.2.0 | rftoolspower-1.12-1.2.0.jar | None |
| LCHIJA | signals | 1.4.1-30 | Signals-1.12.2-1.4.1-30-universal.jar | None |
| LCHIJA | simplevoidworld | 1.2.0.9 | Simple-Void-World-1.12-1.2.0.9-universal.jar | None |
| LCHIJA | snad | 1.12.1-1.7.09.16a | Snad-1.12.1-1.7.09.16a.jar | None |
| LCHIJA | storagedrawers | 1.12-5.3.5 | StorageDrawers-1.12.2-5.3.8.jar | None |
| LCHIJA | storagedrawersextra | @VERSION@ | StorageDrawersExtras-1.12-3.1.0.jar | None |
| LCHIJA | thermalcultivation | 0.3.3 | ThermalCultivation-1.12.2-0.3.3.12-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermaldynamics | 2.5.5 | ThermalDynamics-1.12.2-2.5.5.21-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thermalinnovation | 0.3.3 | ThermalInnovation-1.12.2-0.3.3.12-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f |
| LCHIJA | thutcore | 5.21.0 | thutcore-1.12.2-5.21.0.jar | None |
| LCHIJA | thutcore_compat | 1.0 | thutcore-1.12.2-5.21.0.jar | None |
| LCHIJA | thuttech | 6.1.6 | thuttech-1.12.2-6.1.6.jar | None |
| LCHIJA | tinkertoolleveling | 1.12.2-1.1.0.DEV.b23e769 | TinkerToolLeveling-1.12.2-1.1.0.jar | None |
| LCHIJA | topaddons | 1.12.2-1.11.0 | topaddons-1.12.2-1.11.0.jar | None |
| LCHIJA | torchmaster | 1.8.0.79 | torchmaster_1.12.2-1.8.0.79.jar | 5e9a436b366831c8f54a7e80b015784da69278c6 |
| LCHIJA | translocators | 2.5.2.81 | Translocators-1.12.2-2.5.2.81-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | traverse | 1.6.0 | Traverse-1.12.2-1.6.0-69.jar | None |
| LCHIJA | universalmodifiers | 1.12.2-1.0.16.1 | valkyrielib-1.12.2-2.0.20.1.jar | None |
| LCHIJA | vanillavistas | 1.0 | Vanilla_Vistas 1.1.jar | None |
| LCHIJA | wanionlib | 1.12.2-2.2 | WanionLib-1.12.2-2.2.jar | None |
| LCHIJA | waystones | 4.0.72 | Waystones_1.12.2-4.0.72.jar | None |
| LCHIJA | wct | 3.12.93 | WirelessCraftingTerminal-1.12.2-3.12.93.jar | 186bc454cd122c9c2f1aa4f95611254bcc543363 |
| LCHIJA | woot | 1.12.2-1.4.11 | woot-1.12.2-1.4.11.jar | None |
| LCHIJA | worldedit | 6.1.10-SNAPSHOT | worldedit-forge-mc1.12.2-6.1.10-SNAPSHOT-dist.jar | None |
| LCHIJA | wrcbe | 2.3.2 | WR-CBE-1.12.2-2.3.2.33-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 |
| LCHIJA | xnet | 1.8.2 | xnet-1.12-1.8.2.jar | None |
| LCHIJA | xtones | 1.12-1.0.8-11 | Xtones-1.12-1.0.8-11.jar | None |
| LCHIJA | yabba | 1.1.2.53 | YABBA-1.1.2.53.jar | None |
| LCHIJA | structurize | 1.12.2-0.10.124-ALPHA | structurize-1.12.2-0.10.124-ALPHA.jar | None |
| LCHIJA | minecolonies | 1.12.2-0.10.348-RELEASE | minecolonies-1.12.2-0.10.348-RELEASE-universal.jar | None |
| LCHIJA | industrialwires | 1.7-36 | IndustrialWires-1.7-36.jar | 7e11c175d1e24007afec7498a1616bef0000027d |
| LCHIJA | orelib | 3.5.2.2 | OreLib-1.12.2-3.5.2.2.jar | 7a2128d395ad96ceb9d9030fbd41d035b435753a |
| LCHIJA | dsurround | 3.5.4.3 | DynamicSurroundings-1.12.2-3.5.4.3.jar | 7a2128d395ad96ceb9d9030fbd41d035b435753a |
| LCHIJA | eleccoreloader | 1.9.452 | ElecCore-1.12.2-1.9.452.jar | None |
| LCHIJA | mysticallib | 1.12.2-1.3.4 | mysticallib-1.12.2-1.3.4.jar | None |
| LCHIJA | teslacorelib_registries | 1.0.15 | tesla-core-lib-1.12.2-1.0.15.jar | None |
| LCHIJA | unidict | 1.12.2-2.9.2 | UniDict-1.12.2-2.9.2.jar | None |
| LCHIJA | wrapup | 1.12-1.1.3 | WrapUp-1.12-1.1.3.jar | None |
Loaded coremods (and transformers):
RandomPatches (randompatches-1.12.2-1.18.1.1.jar)
com.therandomlabs.randompatches.core.RPTransformer
SplashAnimationCoremod (SplashAnimation-0.2.1.jar)
pl.asie.splashanimation.core.SplashProgressTransformer
BedPatch (bedpatch-2.2-1.12.2.jar)
com.mordenkainen.bedpatch.BedPatchASM
Do not report to Forge! (If you haven't disabled the FoamFix coremod, try disabling it in the config! Note that this bit of text will still appear.) (foamfix-0.10.5-1.12.2.jar)
pl.asie.foamfix.coremod.FoamFixTransformer
Quark Plugin (Quark-r1.6-177.jar)
vazkii.quark.base.asm.ClassTransformer
OTGCorePlugin (OTG-Core.jar)
com.pg85.otg.forge.asm.OTGClassTransformer
OpenModsCorePlugin (OpenModsLib-1.12.2-0.12.2.jar)
openmods.core.OpenModsClassTransformer
TransformerLoader (OpenComputers-MC1.12.2-1.7.4.153.jar)
li.cil.oc.common.asm.ClassTransformer
EnderCorePlugin (EnderCore-1.12.2-0.5.65-core.jar)
com.enderio.core.common.transform.EnderCoreTransformer
com.enderio.core.common.transform.SimpleMixinPatcher
CTMCorePlugin (CTM-MC1.12.2-0.3.3.22.jar)
team.chisel.ctm.client.asm.CTMTransformer
p455w0rdc0re (p455w0rdslib-1.12.2-2.2.156.jar)
p455w0rdslib.asm.ClassTransformer
ForgelinPlugin (Forgelin-1.8.3.jar)
Plugin (NotEnoughIDs-1.5.4.4.jar)
ru.fewizz.neid.asm.Transformer
CorePlugin (ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar)
IELoadingPlugin (ImmersiveEngineering-core-0.12-89.jar)
blusunrize.immersiveengineering.common.asm.IEClassTransformer
Inventory Tweaks Coremod (InventoryTweaks-1.63.jar)
invtweaks.forge.asm.ContainerTransformer
ShetiPhian-ASM (ShetiPhian-ASM-1.12.0.jar)
shetiphian.asm.ClassTransformer
MekanismCoremod (Mekanism-1.12.2-9.8.1.383.jar)
mekanism.coremod.KeybindingMigrationHelper
LoadingPlugin (ResourceLoader-MC1.12.1-1.5.3.jar)
lumien.resourceloader.asm.ClassTransformer
CoreMod (Aroma1997Core-1.12.2-2.0.0.2.b165.jar)
DynamicSurroundingsCore (DynamicSurroundings-core-1.12.2-3.5.4.3.jar)
org.orecruncher.dsurround.asm.Transformer
AdvancedRocketryPlugin (AdvancedRocketry-1.12.2-1.6.0-164-universal.jar)
zmaster587.advancedRocketry.asm.ClassTransformer
OpenModsLib class transformers: [llama_null_fix:FINISHED],[horse_base_null_fix:FINISHED],[pre_world_render_hook:ENABLED],[player_render_hook:ENABLED],[horse_null_fix:FINISHED]
AE2 Version: stable rv6-stable-7 for Forge 14.23.5.2768
Pulsar/tconstruct loaded Pulses:
- TinkerCommons (Enabled/Forced)
- TinkerWorld (Enabled/Not Forced)
- TinkerTools (Enabled/Not Forced)
- TinkerHarvestTools (Enabled/Forced)
- TinkerMeleeWeapons (Enabled/Forced)
- TinkerRangedWeapons (Enabled/Forced)
- TinkerModifiers (Enabled/Forced)
- TinkerSmeltery (Enabled/Not Forced)
- TinkerGadgets (Enabled/Not Forced)
- TinkerOredict (Enabled/Forced)
- TinkerIntegration (Enabled/Forced)
- TinkerFluids (Enabled/Forced)
- TinkerMaterials (Enabled/Forced)
- TinkerModelRegister (Enabled/Forced)
- chiselIntegration (Enabled/Not Forced)
- chiselsandbitsIntegration (Enabled/Not Forced)
- theoneprobeIntegration (Enabled/Not Forced)
Pulsar/natura loaded Pulses:
- NaturaCommons (Enabled/Forced)
- NaturaOverworld (Enabled/Not Forced)
- NaturaNether (Enabled/Not Forced)
- NaturaDecorative (Enabled/Not Forced)
- NaturaTools (Enabled/Not Forced)
- NaturaEntities (Enabled/Not Forced)
- NaturaOredict (Enabled/Forced)
- NaturaWorld (Enabled/Not Forced)
List of loaded APIs:
* actuallyadditionsapi (34) from ActuallyAdditions-1.12.2-r149.jar
* ae2wtlib|API (1.0.30) from AE2WTLib-1.12.2-1.0.30.jar
* AgriCraftAPI (1.0) from AgriCraft-2.12.0-1.12.0-a6.jar
* appliedenergistics2|API (rv6) from appliedenergistics2-rv6-stable-7.jar
* Base|API (1.0.0) from base-1.12.2-3.13.0.jar
* Baubles|API (1.4.0.2) from Baubles-1.12-1.5.2.jar
* betteradvancements|API (0.1.0.77) from BetterAdvancements-1.12.2-0.1.0.77.jar
* BetterWithModsAPI (Beta 0.6) from AppleSkin-mc1.12-1.0.9.jar
* bigreactors|API (4.0.1) from ExtremeReactors-1.12.2-0.4.5.67.jar
* BotaniaAPI (79) from AkashicTome-1.2-12.jar
* chesttransporter|api (2) from ChestTransporter-1.12.2-2.8.8.jar
* Chisel-API (0.0.1) from Chisel-MC1.12.2-0.2.1.35.jar
* ChiselAPI|Carving (0.0.1) from Chisel-MC1.12.2-0.2.1.35.jar
* ChiselsAndBitsAPI (14.25.0) from chiselsandbits-14.33.jar
* cofhapi (2.5.0) from CoFHCore-1.12.2-4.6.3.27-universal.jar
* ctm-api (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-events (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-models (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-textures (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* ctm-api-utils (0.1.0) from CTM-MC1.12.2-0.3.3.22.jar
* DraconicEvolution|API (1.3) from Draconic-Evolution-1.12.2-2.3.24.349-universal.jar
* ElecCoreAPI (1.0.0) from ElecCore-1.12.2-1.9.452.jar
* enderioapi (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|addon (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|capacitor (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|conduits (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|farm (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|redstone (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|teleport (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|tools (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* enderioapi|upgrades (4.0.0) from EnderIO-1.12.2-5.0.50.jar
* fluxapi (1.0) from fluxnetworks-1.12.2-3.0.19-21.jar
* ForgeEndertechAPI (1.0) from ForgeEndertech-1.12.2-4.4.22.2-build.0405.jar
* Guide-API|API (2.0.0) from Guide-API-1.12-2.1.8-63.jar
* iChunUtil API (1.2.0) from iChunUtil-1.12.2-7.2.1.jar
* ImmersiveEngineering|API (1.0) from ImmersiveEngineering-0.12-89.jar
* ImmersiveEngineering|ImmersiveFluxAPI (1.0) from ImmersiveEngineering-0.12-89.jar
* industrialforegoingapi (5) from industrialforegoing-1.12.2-1.12.13-237.jar
* jeresources|API (0.9.2.60) from JustEnoughResources-1.12.2-0.9.2.60.jar
* journeymap|client-api (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-display (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-event (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-model (1.4) from journeymap-1.12.2-5.5.5.jar
* journeymap|client-api-util (1.4) from journeymap-1.12.2-5.5.5.jar
* JustEnoughItemsAPI (4.13.0) from jei_1.12.2-4.15.0.289.jar
* MekanismAPI|core (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|energy (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|gas (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|infuse (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|laser (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|transmitter (9.8.1) from Mekanism-1.12.2-9.8.1.383.jar
* MekanismAPI|util (9.0.0) from Mekanism-1.12.2-9.8.1.383.jar
* minecolonies-api (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|achievements (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|decorative (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|huts (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|interfaces (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|blocks|types (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render|modeltype (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|client|render|modeltype|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|views (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|buildings|workerbuildings (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|guardtype (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|guardtype|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|jobs (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|jobs|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|managers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|managers|interfaces (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|permissions (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|data (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|factory (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|factory|standard (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|location (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|manager (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|request (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requestable (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requestable|crafting (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|requester (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver|player (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|resolver|retrying (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|requestsystem|token (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|colony|workorders (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|candb (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|dynamictrees (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|gbook (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|compatibility|tinkers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|configuration (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|crafting (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|creativetab (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen|builder (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|citizen|guards (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|pathfinding (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|basestatemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|states (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|tickratestatemachine (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|ai|statemachine|transition (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|citizen (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|citizen|citizenhandlers (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|barbarians (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|pirates (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|mobs|util (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|pathfinding (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|entity|pathfinding|registry (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|inventory (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|inventory|api (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|items (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|network (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|sounds (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|tileentities (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|util (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-api|util|constants (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout|controls (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* minecolonies-blockout|views (1.12.2-0.10.348-RELEASE) from minecolonies-1.12.2-0.10.348-RELEASE-universal.jar
* modtogether|api (1.0) from minetogether-1.10.2-2.2.0-RC6.jar
* MouseTweaks|API (1.0) from MouseTweaks-2.10-mc1.12.2.jar
* openblocks|api (1.2) from OpenBlocks-1.12.2-1.8.1.jar
* opencomputersapi|component (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|core (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|driver (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|driver|item (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|event (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|filesystem (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|internal (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|machine (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|manual (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|network (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* opencomputersapi|prefab (7.0.0-alpha) from OpenComputers-MC1.12.2-1.7.4.153.jar
* projectred|api (2.1) from ProjectRed-1.12.2-4.9.4.120-Base.jar
* QuarkAPI (4) from Quark-r1.6-177.jar
* reborncoreAPI (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Power (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Praescriptum (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Recipe (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* reborncoreAPI|Tile (3.16.0.469) from RebornCore-1.12.2-3.16.0.469-universal.jar
* redstonefluxapi (2.1.0) from RedstoneFlux-1.12-2.1.0.6-universal.jar
* sonarapi (1.0.1) from sonarcore-1.12.2-5.0.19-20.jar
* StorageDrawersAPI (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|event (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|registry (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|render (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|storage (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* StorageDrawersAPI|storage-attribute (2.1.0) from StorageDrawers-1.12.2-5.3.8.jar
* theoneprobe_api (1.4.4) from theoneprobe-1.12-1.4.28.jar
* thut|API (1.0.0) from thutcore-1.12.2-5.21.0.jar
* valkyrielib.api (1.12.2-2.0.10a) from valkyrielib-1.12.2-2.0.20.1.jar
* wct|api (1.1) from WirelessCraftingTerminal-1.12.2-3.12.93.jar
* zerocore|API|multiblock (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|rectangular (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|tier (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
* zerocore|API|multiblock|validation (1.10.2-0.0.2) from zerocore-1.12.2-0.1.2.8.jar
RebornCore:
Plugin Engine: 0
RebornCore Version: 3.16.0.469
Runtime Debofucsation 1
Ender IO: No known problems detected.
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
!!!You are looking at the diagnostics information, not at the crash. !!!
!!!Scroll up until you see the line with '---- Minecraft Crash Report ----'!!!
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
AE2 Integration: IC2:OFF, RC:OFF, MFR:OFF, Waila:OFF, Mekanism:ON, OpenComputers:ON, THE_ONE_PROBE:ON, TESLA:OFF, CRAFTTWEAKER:ON
Profiler Position: N/A (disabled)
Player Count: 1 / 1337; [EntityPlayerMP['firemx'/566, l='VanillaVistas', x=115.22, y=85.79, z=-1632.99]]
Is Modded: Definitely; Server brand changed to 'fml,forge'
Type: Dedicated Server (map_server.txt)
|
non_test
|
ticknexttick error still occurs world crashing minecraft crash report warning coremods are present randompatches randompatches jar splashanimationcoremod splashanimation jar bedpatch bedpatch jar do not report to forge if you haven t disabled the foamfix coremod try disabling it in the config note that this bit of text will still appear foamfix jar quark plugin quark jar otgcoreplugin otg core jar openmodscoreplugin openmodslib jar transformerloader opencomputers jar endercoreplugin endercore core jar ctmcoreplugin ctm jar jar forgelinplugin forgelin jar plugin notenoughids jar coreplugin forgeendertech build jar ieloadingplugin immersiveengineering core jar inventory tweaks coremod inventorytweaks jar shetiphian asm shetiphian asm jar mekanismcoremod mekanism jar loadingplugin resourceloader jar coremod jar dynamicsurroundingscore dynamicsurroundings core jar advancedrocketryplugin advancedrocketry universal jar contact their authors before contacting forge don t do that time pm description exception ticking world java lang illegalstateexception ticknexttick list out of synch at net minecraft world worldserver func a worldserver java at net minecraft world worldserver func b worldserver java at net minecraft server minecraftserver func q minecraftserver java at net minecraft server dedicated dedicatedserver func q dedicatedserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java a detailed walkthrough of the error its code path and all known details is as follows head thread server thread stacktrace at net minecraft world worldserver func a worldserver java at net minecraft world worldserver func b worldserver java affected level details level name vanillavistas all players total chunk stats serverchunkcache drop level seed level generator id otg ver features enabled true level generator options openterraingenerator level spawn location world chunk at in contains blocks to region contains chunks to blocks to level time game time day time level dimension level storage version anvil level weather rain time now false thunder time now true level game mode game mode survival id hardcore false cheats false stacktrace at net minecraft server minecraftserver func q minecraftserver java at net minecraft server dedicated dedicatedserver func q dedicatedserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java system details details minecraft version operating system linux version java version oracle corporation java vm version openjdk bit server vm mixed mode oracle corporation memory bytes mb bytes mb up to bytes mb jvm flags total xx maxpermsize intcache cache tcache allocated tallocated fml mcp powered by forge mods loaded mods active states u unloaded l loaded c constructed h pre initialized i initialized j post initialized a available d disabled e errored state id version source signature lchija minecraft minecraft jar none lchija mcp minecraft jar none lchija fml jar lchija forge jar lchija advancedrocketrycore minecraft jar none lchija otgcore minecraft jar none lchija openmodscore minecraft jar none lchija foamfixcore minecraft jar none lchija opencomputers core minecraft jar none lchija dsurroundcore minecraft jar none lchija randompatches randompatches jar none lchija fastbench fastworkbench jar none lchija actuallyadditions actuallyadditions jar none lchija baubles baubles jar none lchija actuallybaubles actuallybaubles jar none lchija forgeendertech forgeendertech build jar none lchija adlods adlods build jar none lchija redstoneflux redstoneflux universal jar lchija cofhcore cofhcore universal jar none lchija libvulpes libvulpes universal jar none lchija advancedrocketry advancedrocketry universal jar none lchija stable stable jar lchija bdlib bdlib jar none lchija jar none lchija endercore endercore jar none lchija crafttweaker jar none lchija mtlib mtlib jar none lchija modtweaker modtweaker jar none lchija jei jei jar none lchija codechickenlib codechickenlib universal jar lchija brandonscore brandonscore universal jar none lchija cofhworld cofhworld universal jar lchija thermalfoundation thermalfoundation universal jar lchija draconicevolution draconic evolution universal jar none lchija thermalexpansion thermalexpansion universal jar lchija enderio enderio jar none lchija mantle mantle jar none lchija chisel chisel jar none lchija enderiointegrationtic enderio jar none lchija tconstruct tconstruct jar none lchija jar lchija jar lchija infinitylib infinitylib jar none lchija agricraft agricraft jar none lchija akashictome akashictome jar none lchija appleskin appleskin jar none lchija architecturecraft version architecturecraft jar none lchija jar lchija dimensional world jar lchija morphtool morph o tool jar none lchija quark quark jar none lchija autoreglib autoreglib jar none lchija badwithernocookiereloaded badwithernocookiereloaded jar none lchija base base jar none lchija bedpatch bedpatch jar lchija betterbuilderswands betterbuilderswands jar none lchija blockcraftery blockcraftery jar none lchija bookshelf bookshelf jar lchija botania botania jar none lchija chameleon chameleon jar none lchija chesttransporter chesttransporter jar none lchija chiselsandbits chiselsandbits jar none lchija clumps clumps jar none lchija jar none lchija conarm conarm jar lchija cookingforblockheads cookingforblockheads jar none lchija crafttweakerjei jar none lchija theoneprobe theoneprobe jar none lchija danknull danknull jar lchija darkutils darkutils jar lchija diethopper diethopper jar none lchija jar none lchija dynamictrees dynamictrees jar none lchija eleccore eleccore jar none lchija enderiobase enderio jar none lchija enderioconduits enderio jar none lchija enderioconduitsappliedenergistics enderio jar none lchija opencomputers opencomputers jar none lchija enderioconduitsopencomputers enderio jar none lchija enderioconduitsrefinedstorage enderio jar none lchija enderiointegrationforestry enderio jar none lchija enderiointegrationticlate enderio jar none lchija ftblib ftblib jar none lchija enderiomachines enderio jar none lchija enderiopowertools enderio jar none lchija enderstorage enderstorage universal jar lchija immersiveengineering immersiveengineering jar lchija engineersdoors engineers doors jar none lchija valkyrielib valkyrielib jar none lchija environmentaltech environmentaltech jar none lchija etlunar etlunar jar none lchija mcmultipart mcmultipart jar none lchija mekanism mekanism jar none lchija exchangers exchangers jar lchija extracells extracells jar none lchija shadowmc shadowmc jar none lchija extrarails extrarails jar none lchija jar none lchija zerocore zerocore jar none lchija bigreactors extremereactors jar none lchija fencejumper fencejumper jar none lchija flatcoloredblocks flatcoloredblocks jar none lchija sonarcore sonarcore jar none lchija fluxnetworks fluxnetworks jar none lchija foamfix foamfix jar none lchija forgelin forgelin jar none lchija forgemultipartcbe forgemultipart universal jar lchija microblockcbe forgemultipart universal jar none lchija minecraftmultipartcbe forgemultipart universal jar none lchija ftbguides ftbguides jar none lchija ftbutilities ftbutilities jar none lchija ftbbackups ftbbackups ftbutilitiesbackups jar none lchija modularrouters modular routers jar none lchija guideapi guide api jar none lchija harvest harvest jar none lchija ichunutil ichunutil jar lchija immersivepetroleum immersivepetroleum jar none lchija immersivetech immersivetech jar none lchija teslacorelib tesla core lib jar lchija industrialforegoing industrialforegoing jar none lchija initialinventory initialinventory jar none lchija natura natura jar none lchija openterraingenerator openterraingenerator jar lchija nuclearcraft nuclearcraft jar none lchija mcjtylib ng mcjtylib jar none lchija rftools rftools jar none lchija rustic rustic jar none lchija integrationforegoing integrationforegoing jar lchija inventorytweaks release inventorytweaks jar lchija ironbackpacks ironbackpacks jar none lchija ironchest ironchest jar none lchija jaopca jaopca jar none lchija oredictinit jaopca jar none lchija journeymap journeymap jar none lchija kleeslabs kleeslabs jar none lchija laggoggles fat laggoggles fat jar none lchija letsencryptcraft version letsencryptcraft jar none lchija longfallboots longfallboots jar none lchija lostcities lostcities jar none lchija mekanismgenerators mekanismgenerators jar none lchija minetogether unspecified minetogether jar none lchija minetogetherserver unspecified minetogether jar none lchija morpheus morpheus jar none lchija mrtjpcore mrtjpcore universal jar none lchija naturescompass naturescompass jar none lchija netherportalfix netherportalfix jar none lchija nice nice jar none lchija neid notenoughids jar none lchija notenoughwands notenoughwands jar none lchija openmods openmodslib jar lchija openblocks openblocks jar lchija placebo placebo jar none lchija shetiphiancore shetiphiancore jar none lchija platforms platforms jar none lchija portalgun portalgun jar lchija projectred core projectred base jar none lchija projectred integration projectred integration jar none lchija projectred transmission projectred integration jar none lchija projectred fabrication projectred fabrication jar none lchija projectred illumination projectred lighting jar none lchija projectred expansion projectred mechanical jar none lchija projectred relocation projectred mechanical jar none lchija projectred transportation projectred mechanical jar none lchija projectred exploration projectred world jar none lchija quickleafdecay quickleafdecay jar none lchija rangedpumps rangedpumps jar none lchija reborncore reborncore universal jar lchija redstonearsenal redstonearsenal universal jar lchija rftoolscontrol rftoolsctrl jar none lchija rftoolspower rftoolspower jar none lchija signals signals universal jar none lchija simplevoidworld simple void world universal jar none lchija snad snad jar none lchija storagedrawers storagedrawers jar none lchija storagedrawersextra version storagedrawersextras jar none lchija thermalcultivation thermalcultivation universal jar lchija thermaldynamics thermaldynamics universal jar lchija thermalinnovation thermalinnovation universal jar lchija thutcore thutcore jar none lchija thutcore compat thutcore jar none lchija thuttech thuttech jar none lchija tinkertoolleveling dev tinkertoolleveling jar none lchija topaddons topaddons jar none lchija torchmaster torchmaster jar lchija translocators translocators universal jar lchija traverse traverse jar none lchija universalmodifiers valkyrielib jar none lchija vanillavistas vanilla vistas jar none lchija wanionlib wanionlib jar none lchija waystones waystones jar none lchija wct wirelesscraftingterminal jar lchija woot woot jar none lchija worldedit snapshot worldedit forge snapshot dist jar none lchija wrcbe wr cbe universal jar lchija xnet xnet jar none lchija xtones xtones jar none lchija yabba yabba jar none lchija structurize alpha structurize alpha jar none lchija minecolonies release minecolonies release universal jar none lchija industrialwires industrialwires jar lchija orelib orelib jar lchija dsurround dynamicsurroundings jar lchija eleccoreloader eleccore jar none lchija mysticallib mysticallib jar none lchija teslacorelib registries tesla core lib jar none lchija unidict unidict jar none lchija wrapup wrapup jar none loaded coremods and transformers randompatches randompatches jar com therandomlabs randompatches core rptransformer splashanimationcoremod splashanimation jar pl asie splashanimation core splashprogresstransformer bedpatch bedpatch jar com mordenkainen bedpatch bedpatchasm do not report to forge if you haven t disabled the foamfix coremod try disabling it in the config note that this bit of text will still appear foamfix jar pl asie foamfix coremod foamfixtransformer quark plugin quark jar vazkii quark base asm classtransformer otgcoreplugin otg core jar com otg forge asm otgclasstransformer openmodscoreplugin openmodslib jar openmods core openmodsclasstransformer transformerloader opencomputers jar li cil oc common asm classtransformer endercoreplugin endercore core jar com enderio core common transform endercoretransformer com enderio core common transform simplemixinpatcher ctmcoreplugin ctm jar team chisel ctm client asm ctmtransformer jar asm classtransformer forgelinplugin forgelin jar plugin notenoughids jar ru fewizz neid asm transformer coreplugin forgeendertech build jar ieloadingplugin immersiveengineering core jar blusunrize immersiveengineering common asm ieclasstransformer inventory tweaks coremod inventorytweaks jar invtweaks forge asm containertransformer shetiphian asm shetiphian asm jar shetiphian asm classtransformer mekanismcoremod mekanism jar mekanism coremod keybindingmigrationhelper loadingplugin resourceloader jar lumien resourceloader asm classtransformer coremod jar dynamicsurroundingscore dynamicsurroundings core jar org orecruncher dsurround asm transformer advancedrocketryplugin advancedrocketry universal jar advancedrocketry asm classtransformer openmodslib class transformers version stable stable for forge pulsar tconstruct loaded pulses tinkercommons enabled forced tinkerworld enabled not forced tinkertools enabled not forced tinkerharvesttools enabled forced tinkermeleeweapons enabled forced tinkerrangedweapons enabled forced tinkermodifiers enabled forced tinkersmeltery enabled not forced tinkergadgets enabled not forced tinkeroredict enabled forced tinkerintegration enabled forced tinkerfluids enabled forced tinkermaterials enabled forced tinkermodelregister enabled forced chiselintegration enabled not forced chiselsandbitsintegration enabled not forced theoneprobeintegration enabled not forced pulsar natura loaded pulses naturacommons enabled forced naturaoverworld enabled not forced naturanether enabled not forced naturadecorative enabled not forced naturatools enabled not forced naturaentities enabled not forced naturaoredict enabled forced naturaworld enabled not forced list of loaded apis actuallyadditionsapi from actuallyadditions jar api from jar agricraftapi from agricraft jar api from stable jar base api from base jar baubles api from baubles jar betteradvancements api from betteradvancements jar betterwithmodsapi beta from appleskin jar bigreactors api from extremereactors jar botaniaapi from akashictome jar chesttransporter api from chesttransporter jar chisel api from chisel jar chiselapi carving from chisel jar chiselsandbitsapi from chiselsandbits jar cofhapi from cofhcore universal jar ctm api from ctm jar ctm api events from ctm jar ctm api models from ctm jar ctm api textures from ctm jar ctm api utils from ctm jar draconicevolution api from draconic evolution universal jar eleccoreapi from eleccore jar enderioapi from enderio jar enderioapi addon from enderio jar enderioapi capacitor from enderio jar enderioapi conduits from enderio jar enderioapi farm from enderio jar enderioapi redstone from enderio jar enderioapi teleport from enderio jar enderioapi tools from enderio jar enderioapi upgrades from enderio jar fluxapi from fluxnetworks jar forgeendertechapi from forgeendertech build jar guide api api from guide api jar ichunutil api from ichunutil jar immersiveengineering api from immersiveengineering jar immersiveengineering immersivefluxapi from immersiveengineering jar industrialforegoingapi from industrialforegoing jar jeresources api from justenoughresources jar journeymap client api from journeymap jar journeymap client api display from journeymap jar journeymap client api event from journeymap jar journeymap client api model from journeymap jar journeymap client api util from journeymap jar justenoughitemsapi from jei jar mekanismapi core from mekanism jar mekanismapi energy from mekanism jar mekanismapi gas from mekanism jar mekanismapi infuse from mekanism jar mekanismapi laser from mekanism jar mekanismapi transmitter from mekanism jar mekanismapi util from mekanism jar minecolonies api release from minecolonies release universal jar minecolonies api achievements release from minecolonies release universal jar minecolonies api blocks release from minecolonies release universal jar minecolonies api blocks decorative release from minecolonies release universal jar minecolonies api blocks huts release from minecolonies release universal jar minecolonies api blocks interfaces release from minecolonies release universal jar minecolonies api blocks types release from minecolonies release universal jar minecolonies api client release from minecolonies release universal jar minecolonies api client render release from minecolonies release universal jar minecolonies api client render modeltype release from minecolonies release universal jar minecolonies api client render modeltype registry release from minecolonies release universal jar minecolonies api colony release from minecolonies release universal jar minecolonies api colony buildings release from minecolonies release universal jar minecolonies api colony buildings registry release from minecolonies release universal jar minecolonies api colony buildings views release from minecolonies release universal jar minecolonies api colony buildings workerbuildings release from minecolonies release universal jar minecolonies api colony guardtype release from minecolonies release universal jar minecolonies api colony guardtype registry release from minecolonies release universal jar minecolonies api colony jobs release from minecolonies release universal jar minecolonies api colony jobs registry release from minecolonies release universal jar minecolonies api colony managers release from minecolonies release universal jar minecolonies api colony managers interfaces release from minecolonies release universal jar minecolonies api colony permissions release from minecolonies release universal jar minecolonies api colony requestsystem release from minecolonies release universal jar minecolonies api colony requestsystem data release from minecolonies release universal jar minecolonies api colony requestsystem factory release from minecolonies release universal jar minecolonies api colony requestsystem factory standard release from minecolonies release universal jar minecolonies api colony requestsystem location release from minecolonies release universal jar minecolonies api colony requestsystem manager release from minecolonies release universal jar minecolonies api colony requestsystem request release from minecolonies release universal jar minecolonies api colony requestsystem requestable release from minecolonies release universal jar minecolonies api colony requestsystem requestable crafting release from minecolonies release universal jar minecolonies api colony requestsystem requester release from minecolonies release universal jar minecolonies api colony requestsystem resolver release from minecolonies release universal jar minecolonies api colony requestsystem resolver player release from minecolonies release universal jar minecolonies api colony requestsystem resolver retrying release from minecolonies release universal jar minecolonies api colony requestsystem token release from minecolonies release universal jar minecolonies api colony workorders release from minecolonies release universal jar minecolonies api compatibility release from minecolonies release universal jar minecolonies api compatibility candb release from minecolonies release universal jar minecolonies api compatibility dynamictrees release from minecolonies release universal jar minecolonies api compatibility gbook release from minecolonies release universal jar minecolonies api compatibility tinkers release from minecolonies release universal jar minecolonies api configuration release from minecolonies release universal jar minecolonies api crafting release from minecolonies release universal jar minecolonies api creativetab release from minecolonies release universal jar minecolonies api entity release from minecolonies release universal jar minecolonies api entity ai release from minecolonies release universal jar minecolonies api entity ai citizen release from minecolonies release universal jar minecolonies api entity ai citizen builder release from minecolonies release universal jar minecolonies api entity ai citizen guards release from minecolonies release universal jar minecolonies api entity ai pathfinding release from minecolonies release universal jar minecolonies api entity ai registry release from minecolonies release universal jar minecolonies api entity ai statemachine release from minecolonies release universal jar minecolonies api entity ai statemachine basestatemachine release from minecolonies release universal jar minecolonies api entity ai statemachine states release from minecolonies release universal jar minecolonies api entity ai statemachine tickratestatemachine release from minecolonies release universal jar minecolonies api entity ai statemachine transition release from minecolonies release universal jar minecolonies api entity citizen release from minecolonies release universal jar minecolonies api entity citizen citizenhandlers release from minecolonies release universal jar minecolonies api entity mobs release from minecolonies release universal jar minecolonies api entity mobs barbarians release from minecolonies release universal jar minecolonies api entity mobs pirates release from minecolonies release universal jar minecolonies api entity mobs util release from minecolonies release universal jar minecolonies api entity pathfinding release from minecolonies release universal jar minecolonies api entity pathfinding registry release from minecolonies release universal jar minecolonies api inventory release from minecolonies release universal jar minecolonies api inventory api release from minecolonies release universal jar minecolonies api items release from minecolonies release universal jar minecolonies api network release from minecolonies release universal jar minecolonies api sounds release from minecolonies release universal jar minecolonies api tileentities release from minecolonies release universal jar minecolonies api util release from minecolonies release universal jar minecolonies api util constants release from minecolonies release universal jar minecolonies blockout release from minecolonies release universal jar minecolonies blockout controls release from minecolonies release universal jar minecolonies blockout views release from minecolonies release universal jar modtogether api from minetogether jar mousetweaks api from mousetweaks jar openblocks api from openblocks jar opencomputersapi component alpha from opencomputers jar opencomputersapi core alpha from opencomputers jar opencomputersapi driver alpha from opencomputers jar opencomputersapi driver item alpha from opencomputers jar opencomputersapi event alpha from opencomputers jar opencomputersapi filesystem alpha from opencomputers jar opencomputersapi internal alpha from opencomputers jar opencomputersapi machine alpha from opencomputers jar opencomputersapi manual alpha from opencomputers jar opencomputersapi network alpha from opencomputers jar opencomputersapi prefab alpha from opencomputers jar projectred api from projectred base jar quarkapi from quark jar reborncoreapi from reborncore universal jar reborncoreapi power from reborncore universal jar reborncoreapi praescriptum from reborncore universal jar reborncoreapi recipe from reborncore universal jar reborncoreapi tile from reborncore universal jar redstonefluxapi from redstoneflux universal jar sonarapi from sonarcore jar storagedrawersapi from storagedrawers jar storagedrawersapi event from storagedrawers jar storagedrawersapi registry from storagedrawers jar storagedrawersapi render from storagedrawers jar storagedrawersapi storage from storagedrawers jar storagedrawersapi storage attribute from storagedrawers jar theoneprobe api from theoneprobe jar thut api from thutcore jar valkyrielib api from valkyrielib jar wct api from wirelesscraftingterminal jar zerocore api multiblock from zerocore jar zerocore api multiblock rectangular from zerocore jar zerocore api multiblock tier from zerocore jar zerocore api multiblock validation from zerocore jar reborncore plugin engine reborncore version runtime debofucsation ender io no known problems detected you are looking at the diagnostics information not at the crash scroll up until you see the line with minecraft crash report integration off rc off mfr off waila off mekanism on opencomputers on the one probe on tesla off crafttweaker on profiler position n a disabled player count is modded definitely server brand changed to fml forge type dedicated server map server txt
| 0
|
91,338
| 8,303,331,136
|
IssuesEvent
|
2018-09-21 17:11:33
|
udacity/lesson_feedback_nd113
|
https://api.github.com/repos/udacity/lesson_feedback_nd113
|
closed
|
[Negative]2018-08-15
|
1.0.0 19. Matrices and Transformation of State test
|
No Solutions to coding parts\, this just makes it much harder for people with basic programming skills to understand the work that is going on.
|
1.0
|
[Negative]2018-08-15 - No Solutions to coding parts\, this just makes it much harder for people with basic programming skills to understand the work that is going on.
|
test
|
no solutions to coding parts this just makes it much harder for people with basic programming skills to understand the work that is going on
| 1
|
289,388
| 24,984,971,069
|
IssuesEvent
|
2022-11-02 14:28:20
|
MetaMask/metamask-extension
|
https://api.github.com/repos/MetaMask/metamask-extension
|
closed
|
e2e tests for editing and deleting contacts
|
area-testSuite stedmap team extension client
|
We need to expand the functionality covered by our `test/e2e/tests/address-book.spec.js`
Currently, there are tests for adding a contact and sending to a contact.
We need to add tests for editing an already added contact and deleting an already added contact.
We should be able to use the `'address-entry'` fixture. Add a separate `it` block, description string and `withFixtures` call for each of these cases. Appropriate assertions should be made against rendered text before and after editing and deleting.
|
1.0
|
e2e tests for editing and deleting contacts - We need to expand the functionality covered by our `test/e2e/tests/address-book.spec.js`
Currently, there are tests for adding a contact and sending to a contact.
We need to add tests for editing an already added contact and deleting an already added contact.
We should be able to use the `'address-entry'` fixture. Add a separate `it` block, description string and `withFixtures` call for each of these cases. Appropriate assertions should be made against rendered text before and after editing and deleting.
|
test
|
tests for editing and deleting contacts we need to expand the functionality covered by our test tests address book spec js currently there are tests for adding a contact and sending to a contact we need to add tests for editing an already added contact and deleting an already added contact we should be able to use the address entry fixture add a separate it block description string and withfixtures call for each of these cases appropriate assertions should be made against rendered text before and after editing and deleting
| 1
|
710,498
| 24,420,506,484
|
IssuesEvent
|
2022-10-05 19:51:06
|
chaotic-aur/packages
|
https://api.github.com/repos/chaotic-aur/packages
|
closed
|
[Request] corepaint
|
request:new-pkg priority:low
|
### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/corepaint
### Utility this package has for you
A paint app from the C Suite
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a great amount.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_
|
1.0
|
[Request] corepaint - ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/corepaint
### Utility this package has for you
A paint app from the C Suite
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a great amount.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_
|
non_test
|
corepaint link to the package s in the aur utility this package has for you a paint app from the c suite do you consider the package s to be useful for every chaotic aur user no but for a great amount do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information no response
| 0
|
322,264
| 27,592,482,815
|
IssuesEvent
|
2023-03-09 02:07:12
|
dotnet/machinelearning-modelbuilder
|
https://api.github.com/repos/dotnet/machinelearning-modelbuilder
|
opened
|
All scenarios: "Next step" button on the Environment page is disabled, hence we cannot continue the next steps.
|
Priority:0 Reported by: Test
|
**System Information (please complete the following information):**
Windows OS: Windows-11-Enterprise-22H2
ML.Net Model Builder 2022: 17.14.4.2315802 (Main Build)
Microsoft Visual Studio Enterprise: 2022(17.4.5)
.Net: 6.0
**Describe the bug**
- On which step of the process did you run into an issue:
"Next step" button on the Environment page is disabled, hence we cannot continue the next steps.
**TestMatrix**
https://testpass.blob.core.windows.net/test-pass-data/wikipedia-detox-250-line-data.tsv
**To Reproduce**
Steps to reproduce the behavior:
1. Select Create a new project from the Visual Studio start window.
2. Choose the C# Console App (.NET Core) project template.
3. Add model builder by right click on the project.
4. Select any scenario.
5. Then go to the Environment page, you will see that "Next step" button is disabled.
**Expected behavior**
"Next step" button on the Environment page should be enabled.
**Screenshot**

**Additional context**
For **Image classification & Object detection**, if you choose the **Azure** environment, VS would be crashed.


|
1.0
|
All scenarios: "Next step" button on the Environment page is disabled, hence we cannot continue the next steps. - **System Information (please complete the following information):**
Windows OS: Windows-11-Enterprise-22H2
ML.Net Model Builder 2022: 17.14.4.2315802 (Main Build)
Microsoft Visual Studio Enterprise: 2022(17.4.5)
.Net: 6.0
**Describe the bug**
- On which step of the process did you run into an issue:
"Next step" button on the Environment page is disabled, hence we cannot continue the next steps.
**TestMatrix**
https://testpass.blob.core.windows.net/test-pass-data/wikipedia-detox-250-line-data.tsv
**To Reproduce**
Steps to reproduce the behavior:
1. Select Create a new project from the Visual Studio start window.
2. Choose the C# Console App (.NET Core) project template.
3. Add model builder by right click on the project.
4. Select any scenario.
5. Then go to the Environment page, you will see that "Next step" button is disabled.
**Expected behavior**
"Next step" button on the Environment page should be enabled.
**Screenshot**

**Additional context**
For **Image classification & Object detection**, if you choose the **Azure** environment, VS would be crashed.


|
test
|
all scenarios next step button on the environment page is disabled hence we cannot continue the next steps system information please complete the following information windows os windows enterprise ml net model builder main build microsoft visual studio enterprise net describe the bug on which step of the process did you run into an issue next step button on the environment page is disabled hence we cannot continue the next steps testmatrix to reproduce steps to reproduce the behavior select create a new project from the visual studio start window choose the c console app net core project template add model builder by right click on the project select any scenario then go to the environment page you will see that next step button is disabled expected behavior next step button on the environment page should be enabled screenshot additional context for image classification object detection if you choose the azure environment vs would be crashed
| 1
|
226,016
| 17,934,973,441
|
IssuesEvent
|
2021-09-10 14:16:06
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
opened
|
Crash vmState=0x000509ff
|
comp:jit test failure segfault
|
https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_x86-64_mac_Nightly_testList_0/90/
testDDRExt_Class_0 (NoOptions)
vmState [0x509ff]: {J9VMSTATE_JIT} {localValuePropagation}
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk11_j9_extended.functional_x86-64_mac_Nightly_testList_0/90/functional_test_output.tar.gz
```
TCK.run.tests.ddrext:
[echo] Running the DDR Extension Test
[java] Unhandled exception
[java] Type=Segmentation error vmState=0x000509ff
[java] J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000000
[java] Handler1=0000000006237770 Handler2=000000000650F420
[java] RDI=00000000079C6F3B RSI=0000700000255644 RAX=0000000000000000 RBX=00000000079C6F3B
[java] RCX=0000000000000000 RDX=00007FABF1D010B0 R8=00000000FFFFFFE9 R9=0000000000000017
[java] R10=00007FABF1D010B0 R11=E000000000000000 R12=00000000300F1D10 R13=0000000000008000
[java] R14=0000000000000000 R15=0000000000000000
[java] RIP=00007FFF873F97D4 GS=0000 FS=0000 RSP=0000700000255610
[java] RFlags=0000000000010206 CS=002B RBP=0000700000255630 ERR=079C6F3B00000000
[java] TRAPNO=000000000000000D CPU=6F3B000000000000 FAULTVADDR=00000000079C6F3B
[java] XMM0 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM1 6974752f6176616a (f: 1635148160.000000, d: 9.787124e+199)
[java] XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM3 2f6f692f6176616a (f: 1635148160.000000, d: 3.311397e-80)
[java] XMM4 706e497473614624 (f: 1935754752.000000, d: 3.761681e+233)
[java] XMM5 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM6 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM7 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM8 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM9 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM10 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM11 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM12 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] Module=/usr/lib/system/libsystem_c.dylib
[java] Module_base_address=00007FFF8739A000 Symbol=__findenv
[java] Symbol_address=00007FFF873F9780
[java]
[java] Method_being_compiled=java/util/jar/JarFile.checkForSpecialAttributes()V
[java] Target=2_90_20210910_96 (Mac OS X 10.11.6)
[java] CPU=amd64 (4 logical CPUs) (0x200000000 RAM)
[java] ----------- Stack Backtrace -----------
[java] ---------------------------------------
```
It's on jdk11 but I've assigned it to the 0.28 milestone just in case, we can move it out if it's not a priority.
|
1.0
|
Crash vmState=0x000509ff - https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_x86-64_mac_Nightly_testList_0/90/
testDDRExt_Class_0 (NoOptions)
vmState [0x509ff]: {J9VMSTATE_JIT} {localValuePropagation}
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk11_j9_extended.functional_x86-64_mac_Nightly_testList_0/90/functional_test_output.tar.gz
```
TCK.run.tests.ddrext:
[echo] Running the DDR Extension Test
[java] Unhandled exception
[java] Type=Segmentation error vmState=0x000509ff
[java] J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000000
[java] Handler1=0000000006237770 Handler2=000000000650F420
[java] RDI=00000000079C6F3B RSI=0000700000255644 RAX=0000000000000000 RBX=00000000079C6F3B
[java] RCX=0000000000000000 RDX=00007FABF1D010B0 R8=00000000FFFFFFE9 R9=0000000000000017
[java] R10=00007FABF1D010B0 R11=E000000000000000 R12=00000000300F1D10 R13=0000000000008000
[java] R14=0000000000000000 R15=0000000000000000
[java] RIP=00007FFF873F97D4 GS=0000 FS=0000 RSP=0000700000255610
[java] RFlags=0000000000010206 CS=002B RBP=0000700000255630 ERR=079C6F3B00000000
[java] TRAPNO=000000000000000D CPU=6F3B000000000000 FAULTVADDR=00000000079C6F3B
[java] XMM0 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM1 6974752f6176616a (f: 1635148160.000000, d: 9.787124e+199)
[java] XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM3 2f6f692f6176616a (f: 1635148160.000000, d: 3.311397e-80)
[java] XMM4 706e497473614624 (f: 1935754752.000000, d: 3.761681e+233)
[java] XMM5 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM6 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM7 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM8 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM9 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM10 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM11 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM12 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00)
[java] Module=/usr/lib/system/libsystem_c.dylib
[java] Module_base_address=00007FFF8739A000 Symbol=__findenv
[java] Symbol_address=00007FFF873F9780
[java]
[java] Method_being_compiled=java/util/jar/JarFile.checkForSpecialAttributes()V
[java] Target=2_90_20210910_96 (Mac OS X 10.11.6)
[java] CPU=amd64 (4 logical CPUs) (0x200000000 RAM)
[java] ----------- Stack Backtrace -----------
[java] ---------------------------------------
```
It's on jdk11 but I've assigned it to the 0.28 milestone just in case, we can move it out if it's not a priority.
|
test
|
crash vmstate testddrext class nooptions vmstate jit localvaluepropagation tck run tests ddrext running the ddr extension test unhandled exception type segmentation error vmstate signal number signal number error value signal code rdi rsi rax rbx rcx rdx rip gs fs rsp rflags cs rbp err trapno cpu faultvaddr f d f d f d f d f d f d f d f d f d f d f d f d f d f d f d f d module usr lib system libsystem c dylib module base address symbol findenv symbol address method being compiled java util jar jarfile checkforspecialattributes v target mac os x cpu logical cpus ram stack backtrace it s on but i ve assigned it to the milestone just in case we can move it out if it s not a priority
| 1
|
122,844
| 26,174,200,028
|
IssuesEvent
|
2023-01-02 07:17:56
|
arduino/arduino-language-server
|
https://api.github.com/repos/arduino/arduino-language-server
|
opened
|
Temporary files are not cleaned up
|
topic: code type: imperfection
|
### Describe the problem
Arduino Language Server and [**clangd**](https://clangd.llvm.org/) create some temporary files:
- Name format: `arduino-language-server2131811926/`
- Arduino CLI sketch build folder
- Name format: `system-includes-0f3fe3.clangd`
- Created by **clangd**
- Name format: `preamble-4df37f.pch`
- Created by **clangd**
The first two of these persist even after the user exits the IDE.
The files accumulate over time, consuming significant amounts of disk space and may eventually cause a performance impact.
### To reproduce
1. Take note of the contents of your temporary folder.
- Windows:
```text
C:\Users\<user name>\AppData\Local\Temp
```
- Linux:
```text
/tmp
```
- macOS
```text
$TMPDIR
```
1. Start Arduino IDE 2.x.
1. Select **Tools > Board > Arduino AVR Boards > Arduino Uno** from the Arduino IDE menus.
**ⓘ** This is only an arbitrary choice of a board that is supported by the language server.
1. Wait for the "**Building sketch**" and/or "**indexing ...**" operations indicated on the left side of the Arduino IDE status bar to finish.
1. Select **File > Quit** from the Arduino IDE menus.
1. Examine the contents of your temporary folder.
🐛 Multiple additional `arduino-language-servernnnnnnnnn` folders are present.
🐛 (If using Windows) Multiple additional `system-includes-nnnnn.clangd` files are present.
🐛 The files produced by this single minimal demo consume a significant amount of disk space (on the order of > 0.5 MB on Windows).
### Expected behavior
Temporary files are cleaned up on exit.
### Arduino Language Server version
bdef5c27830b1d4a33908068ed87514757874052
### Operating system
- Windows
- Linux
- macOS
### Operating system version
- Windows 11
- Ubuntu 22.04
- macOS Ventura
### Additional context
I don't see the `system-includes-nnnnn.clangd` files in the temporary folder on Linux and macOS even while the IDE is running, so I am not sure whether they are never created at all, or whether they are stored in some other location.
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-language-server/issues?q=)
- [X] I verified the problem still occurs when using the latest version
- [X] My report contains all necessary details
|
1.0
|
Temporary files are not cleaned up - ### Describe the problem
Arduino Language Server and [**clangd**](https://clangd.llvm.org/) create some temporary files:
- Name format: `arduino-language-server2131811926/`
- Arduino CLI sketch build folder
- Name format: `system-includes-0f3fe3.clangd`
- Created by **clangd**
- Name format: `preamble-4df37f.pch`
- Created by **clangd**
The first two of these persist even after the user exits the IDE.
The files accumulate over time, consuming significant amounts of disk space and may eventually cause a performance impact.
### To reproduce
1. Take note of the contents of your temporary folder.
- Windows:
```text
C:\Users\<user name>\AppData\Local\Temp
```
- Linux:
```text
/tmp
```
- macOS
```text
$TMPDIR
```
1. Start Arduino IDE 2.x.
1. Select **Tools > Board > Arduino AVR Boards > Arduino Uno** from the Arduino IDE menus.
**ⓘ** This is only an arbitrary choice of a board that is supported by the language server.
1. Wait for the "**Building sketch**" and/or "**indexing ...**" operations indicated on the left side of the Arduino IDE status bar to finish.
1. Select **File > Quit** from the Arduino IDE menus.
1. Examine the contents of your temporary folder.
🐛 Multiple additional `arduino-language-servernnnnnnnnn` folders are present.
🐛 (If using Windows) Multiple additional `system-includes-nnnnn.clangd` files are present.
🐛 The files produced by this single minimal demo consume a significant amount of disk space (on the order of > 0.5 MB on Windows).
### Expected behavior
Temporary files are cleaned up on exit.
### Arduino Language Server version
bdef5c27830b1d4a33908068ed87514757874052
### Operating system
- Windows
- Linux
- macOS
### Operating system version
- Windows 11
- Ubuntu 22.04
- macOS Ventura
### Additional context
I don't see the `system-includes-nnnnn.clangd` files in the temporary folder on Linux and macOS even while the IDE is running, so I am not sure whether they are never created at all, or whether they are stored in some other location.
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-language-server/issues?q=)
- [X] I verified the problem still occurs when using the latest version
- [X] My report contains all necessary details
|
non_test
|
temporary files are not cleaned up describe the problem arduino language server and create some temporary files name format arduino language arduino cli sketch build folder name format system includes clangd created by clangd name format preamble pch created by clangd the first two of these persist even after the user exits the ide the files accumulate over time consuming significant amounts of disk space and may eventually cause a performance impact to reproduce take note of the contents of your temporary folder windows text c users appdata local temp linux text tmp macos text tmpdir start arduino ide x select tools board arduino avr boards arduino uno from the arduino ide menus ⓘ this is only an arbitrary choice of a board that is supported by the language server wait for the building sketch and or indexing operations indicated on the left side of the arduino ide status bar to finish select file quit from the arduino ide menus examine the contents of your temporary folder 🐛 multiple additional arduino language servernnnnnnnnn folders are present 🐛 if using windows multiple additional system includes nnnnn clangd files are present 🐛 the files produced by this single minimal demo consume a significant amount of disk space on the order of mb on windows expected behavior temporary files are cleaned up on exit arduino language server version operating system windows linux macos operating system version windows ubuntu macos ventura additional context i don t see the system includes nnnnn clangd files in the temporary folder on linux and macos even while the ide is running so i am not sure whether they are never created at all or whether they are stored in some other location issue checklist i searched for previous reports in i verified the problem still occurs when using the latest version my report contains all necessary details
| 0
|
165,517
| 26,183,988,518
|
IssuesEvent
|
2023-01-02 20:09:28
|
flutter/website
|
https://api.github.com/repos/flutter/website
|
opened
|
Migrate to Bootstrap 5
|
infrastructure design p3-low blocked e2-days e3-weeks
|
### Describe the problem
Bootstrap 5 is the current release Bootstrap, replacing Bootstrap 4. We use it heavily across the site and we want to make sure we stay up to date. This will also allow us to eventually drop Jquery since Bootstrap 5 no longer uses it. Beyond that, this will also make a dark mode slightly easier with its (not yet released) [color mode functionality](https://getbootstrap.com/docs/5.3/customize/color-modes/).
### Expected fix
We should migrate away from Bootstrap 4 and to 5.
### Additional context
There may be some incompatibilities due to Bootstrap now using the Dart sass compiler which supports some newer features. We may have to configure Jekyll to use [sass-embedded](https://rubygems.org/gems/sass-embedded) instead somehow, which is implemented with Dart sass :)
|
1.0
|
Migrate to Bootstrap 5 - ### Describe the problem
Bootstrap 5 is the current release Bootstrap, replacing Bootstrap 4. We use it heavily across the site and we want to make sure we stay up to date. This will also allow us to eventually drop Jquery since Bootstrap 5 no longer uses it. Beyond that, this will also make a dark mode slightly easier with its (not yet released) [color mode functionality](https://getbootstrap.com/docs/5.3/customize/color-modes/).
### Expected fix
We should migrate away from Bootstrap 4 and to 5.
### Additional context
There may be some incompatibilities due to Bootstrap now using the Dart sass compiler which supports some newer features. We may have to configure Jekyll to use [sass-embedded](https://rubygems.org/gems/sass-embedded) instead somehow, which is implemented with Dart sass :)
|
non_test
|
migrate to bootstrap describe the problem bootstrap is the current release bootstrap replacing bootstrap we use it heavily across the site and we want to make sure we stay up to date this will also allow us to eventually drop jquery since bootstrap no longer uses it beyond that this will also make a dark mode slightly easier with its not yet released expected fix we should migrate away from bootstrap and to additional context there may be some incompatibilities due to bootstrap now using the dart sass compiler which supports some newer features we may have to configure jekyll to use instead somehow which is implemented with dart sass
| 0
|
25,343
| 4,154,841,619
|
IssuesEvent
|
2016-06-16 13:10:25
|
WormBase/website
|
https://api.github.com/repos/WormBase/website
|
closed
|
Searching for a peptide ID returns irrelevant results
|
Under testing Webteam
|
E.g. searching for BM38054 (Bm2, isoform b) yields a selection of other results but not the protein we are interested in: http://www.wormbase.org/search/protein/BM38054
In some cases, e.g. CN06574, results are displayed for a completely different species.
These identifiers are found elsewhere on the web, e.g. Ensembl, so it could be possible users will be using these to identify proteins and will want to search for them on WormBase.
|
1.0
|
Searching for a peptide ID returns irrelevant results - E.g. searching for BM38054 (Bm2, isoform b) yields a selection of other results but not the protein we are interested in: http://www.wormbase.org/search/protein/BM38054
In some cases, e.g. CN06574, results are displayed for a completely different species.
These identifiers are found elsewhere on the web, e.g. Ensembl, so it could be possible users will be using these to identify proteins and will want to search for them on WormBase.
|
test
|
searching for a peptide id returns irrelevant results e g searching for isoform b yields a selection of other results but not the protein we are interested in in some cases e g results are displayed for a completely different species these identifiers are found elsewhere on the web e g ensembl so it could be possible users will be using these to identify proteins and will want to search for them on wormbase
| 1
|
38,385
| 5,184,451,948
|
IssuesEvent
|
2017-01-20 06:14:39
|
Automattic/jetpack
|
https://api.github.com/repos/Automattic/jetpack
|
closed
|
Unit Tests: add PHP 7.1
|
Unit Tests [Type] Enhancement [Type] Good First Bug
|
It would be nice to add PHP 7.1 to our list of Unit Tests, as it includes a few breaking changes from PHP7, and is getting more and more popular with site owners.
|
1.0
|
Unit Tests: add PHP 7.1 - It would be nice to add PHP 7.1 to our list of Unit Tests, as it includes a few breaking changes from PHP7, and is getting more and more popular with site owners.
|
test
|
unit tests add php it would be nice to add php to our list of unit tests as it includes a few breaking changes from and is getting more and more popular with site owners
| 1
|
645
| 2,577,795,466
|
IssuesEvent
|
2015-02-12 19:10:56
|
chrsmith/google-styleguide
|
https://api.github.com/repos/chrsmith/google-styleguide
|
opened
|
No word about standard date format in lispguide.xml, how about http://www.w3.org/TR/NOTE-datetime
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.
https://google-styleguide.googlecode.com/svn/trunk/lispguide.xml?showone=Attenti
on_Required#Attention_Required
2.
3.
What is the expected output? What do you see instead?
I would expect the "Google Common Lisp Style Guide" to specify the format dates
in comments should be in. My strong preference would be towards ISO 8601, to
spare readers the bafflement which is 1/2/3 or 03.02.01.
The only mention of date is in the example below.
While "November, 2012" is unambiguous enough, it leaves the reader with no
recommendation for representing specific dates or date-time.
Here is the excerpt:
Be specific when indicating times or software releases in a TODO comment:
;;--- TODO(brown): Remove this code after release 1.7 or before November, 2012.
What version of the product are you using? On what operating system?
Please provide any additional information below.
Using a standard date-time format would make it a lot easier for humans and for
tooling to extract useful temporal information from program comments.
```
-----
Original issue reported on code.google.com by adrian.a...@gmail.com on 11 Oct 2012 at 11:31
|
1.0
|
No word about standard date format in lispguide.xml, how about http://www.w3.org/TR/NOTE-datetime - ```
What steps will reproduce the problem?
1.
https://google-styleguide.googlecode.com/svn/trunk/lispguide.xml?showone=Attenti
on_Required#Attention_Required
2.
3.
What is the expected output? What do you see instead?
I would expect the "Google Common Lisp Style Guide" to specify the format dates
in comments should be in. My strong preference would be towards ISO 8601, to
spare readers the bafflement which is 1/2/3 or 03.02.01.
The only mention of date is in the example below.
While "November, 2012" is unambiguous enough, it leaves the reader with no
recommendation for representing specific dates or date-time.
Here is the excerpt:
Be specific when indicating times or software releases in a TODO comment:
;;--- TODO(brown): Remove this code after release 1.7 or before November, 2012.
What version of the product are you using? On what operating system?
Please provide any additional information below.
Using a standard date-time format would make it a lot easier for humans and for
tooling to extract useful temporal information from program comments.
```
-----
Original issue reported on code.google.com by adrian.a...@gmail.com on 11 Oct 2012 at 11:31
|
non_test
|
no word about standard date format in lispguide xml how about what steps will reproduce the problem on required attention required what is the expected output what do you see instead i would expect the google common lisp style guide to specify the format dates in comments should be in my strong preference would be towards iso to spare readers the bafflement which is or the only mention of date is in the example below while november is unambiguous enough it leaves the reader with no recommendation for representing specific dates or date time here is the excerpt be specific when indicating times or software releases in a todo comment todo brown remove this code after release or before november what version of the product are you using on what operating system please provide any additional information below using a standard date time format would make it a lot easier for humans and for tooling to extract useful temporal information from program comments original issue reported on code google com by adrian a gmail com on oct at
| 0
|
225,373
| 17,856,339,225
|
IssuesEvent
|
2021-09-05 04:59:34
|
ObliqueNET/Server
|
https://api.github.com/repos/ObliqueNET/Server
|
closed
|
Plot issues
|
needs testing landlord discussion needed
|
I and the ppl in my village are all mayors so we all have the same ability. Last night we had somone join our town and we have been made aware of this issue. Ppl in my town/village are experiencing this same issue. We cannot create, claim, delete, or edit plots that we had set up for new joins. I tried to figure it out just now and still having this issue
|
1.0
|
Plot issues - I and the ppl in my village are all mayors so we all have the same ability. Last night we had somone join our town and we have been made aware of this issue. Ppl in my town/village are experiencing this same issue. We cannot create, claim, delete, or edit plots that we had set up for new joins. I tried to figure it out just now and still having this issue
|
test
|
plot issues i and the ppl in my village are all mayors so we all have the same ability last night we had somone join our town and we have been made aware of this issue ppl in my town village are experiencing this same issue we cannot create claim delete or edit plots that we had set up for new joins i tried to figure it out just now and still having this issue
| 1
|
37,381
| 5,114,751,386
|
IssuesEvent
|
2017-01-06 19:31:34
|
tomMoral/loky
|
https://api.github.com/repos/tomMoral/loky
|
closed
|
Check crash happening after the executor has been GC'ed
|
testing
|
Add a variant of `test_processes_terminate_on_executor_gc` with all possible crashes.
|
1.0
|
Check crash happening after the executor has been GC'ed - Add a variant of `test_processes_terminate_on_executor_gc` with all possible crashes.
|
test
|
check crash happening after the executor has been gc ed add a variant of test processes terminate on executor gc with all possible crashes
| 1
|
3,283
| 2,666,552,562
|
IssuesEvent
|
2015-03-21 17:50:00
|
contao-community-alliance/dc-general
|
https://api.github.com/repos/contao-community-alliance/dc-general
|
closed
|
[develop] TinyMCE funktioniert nicht
|
bug testing
|
Ich habe bei Contao 3.3.5 das Problem das der TinyMCE nicht angezeigt wurde. Ich habe dann an folgender Stelle https://github.com/contao-community-alliance/dc-general/blob/f0bd490f0d24b6a34a1140ab81b67b93e98606f3/src/ContaoCommunityAlliance/DcGeneral/Contao/View/Contao2BackendView/ContaoWidgetManager.php#L343
dies eingefügt
```
$selector = $propertyId;
```
danach geht es.
|
1.0
|
[develop] TinyMCE funktioniert nicht - Ich habe bei Contao 3.3.5 das Problem das der TinyMCE nicht angezeigt wurde. Ich habe dann an folgender Stelle https://github.com/contao-community-alliance/dc-general/blob/f0bd490f0d24b6a34a1140ab81b67b93e98606f3/src/ContaoCommunityAlliance/DcGeneral/Contao/View/Contao2BackendView/ContaoWidgetManager.php#L343
dies eingefügt
```
$selector = $propertyId;
```
danach geht es.
|
test
|
tinymce funktioniert nicht ich habe bei contao das problem das der tinymce nicht angezeigt wurde ich habe dann an folgender stelle dies eingefügt selector propertyid danach geht es
| 1
|
212,377
| 23,882,526,099
|
IssuesEvent
|
2022-09-08 03:32:02
|
Azure/AKS
|
https://api.github.com/repos/Azure/AKS
|
opened
|
Node Access (SSH Refinement) - Phase 3 update SSH key
|
security feature-request
|
With this feature, we allow users to update the SSH key permanently for AKS all existing nodepools.
|
True
|
Node Access (SSH Refinement) - Phase 3 update SSH key - With this feature, we allow users to update the SSH key permanently for AKS all existing nodepools.
|
non_test
|
node access ssh refinement phase update ssh key with this feature we allow users to update the ssh key permanently for aks all existing nodepools
| 0
|
10,618
| 3,131,030,604
|
IssuesEvent
|
2015-09-09 12:57:00
|
nmaguirre/eiffel-subtitle-converter
|
https://api.github.com/repos/nmaguirre/eiffel-subtitle-converter
|
opened
|
Missing unit tests for routine adjust_stop_frame.MICRODVD_SUBTITLE_ITEM
|
enhancement testing
|
Routine adjust_stop_frame from class MICRODVD_SUBTITLE_ITEM has no unit tests to assess its behaviour. At least one unit "positive" and one "negative" unit test must be provided for this routine. In this context, 'positive' is a test that evaluates the correct behaviour of the routine when right arguments are passed, while 'negative' refers to tests that check that the routine correctly controls ill-formed direct or indirect inputs.
Please use a fine grained granularity for the tests. If you want to assert several things on the same situation, please provide different tests for each assert. See tests for make.MICRODVD_SUBTITLE_ITEM for an example.
|
1.0
|
Missing unit tests for routine adjust_stop_frame.MICRODVD_SUBTITLE_ITEM - Routine adjust_stop_frame from class MICRODVD_SUBTITLE_ITEM has no unit tests to assess its behaviour. At least one unit "positive" and one "negative" unit test must be provided for this routine. In this context, 'positive' is a test that evaluates the correct behaviour of the routine when right arguments are passed, while 'negative' refers to tests that check that the routine correctly controls ill-formed direct or indirect inputs.
Please use a fine grained granularity for the tests. If you want to assert several things on the same situation, please provide different tests for each assert. See tests for make.MICRODVD_SUBTITLE_ITEM for an example.
|
test
|
missing unit tests for routine adjust stop frame microdvd subtitle item routine adjust stop frame from class microdvd subtitle item has no unit tests to assess its behaviour at least one unit positive and one negative unit test must be provided for this routine in this context positive is a test that evaluates the correct behaviour of the routine when right arguments are passed while negative refers to tests that check that the routine correctly controls ill formed direct or indirect inputs please use a fine grained granularity for the tests if you want to assert several things on the same situation please provide different tests for each assert see tests for make microdvd subtitle item for an example
| 1
|
201,027
| 22,946,647,439
|
IssuesEvent
|
2022-07-19 01:06:16
|
liorzilberg/swagger-parser
|
https://api.github.com/repos/liorzilberg/swagger-parser
|
opened
|
CVE-2020-10650 (Medium) detected in jackson-databind-2.9.5.jar
|
security vulnerability
|
## CVE-2020-10650 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind library before versions 2.9.10.4 is vulnerable to an Unsafe Deserialization vulnerability when handling interactions related to the class ignite-jta
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10650>CVE-2020-10650</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-rpr3-cw39-3pxh">https://github.com/advisories/GHSA-rpr3-cw39-3pxh</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-10650 (Medium) detected in jackson-databind-2.9.5.jar - ## CVE-2020-10650 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind library before versions 2.9.10.4 is vulnerable to an Unsafe Deserialization vulnerability when handling interactions related to the class ignite-jta
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10650>CVE-2020-10650</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-rpr3-cw39-3pxh">https://github.com/advisories/GHSA-rpr3-cw39-3pxh</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar target lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details jackson databind library before versions is vulnerable to an unsafe deserialization vulnerability when handling interactions related to the class ignite jta publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
257,591
| 22,196,845,830
|
IssuesEvent
|
2022-06-07 07:45:03
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
sql/catalog/lease: TestLeaseTxnDeadlineExtension failed
|
C-test-failure O-robot branch-master
|
sql/catalog/lease.TestLeaseTxnDeadlineExtension [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/5396090?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/5396090?buildTab=artifacts#/) on master @ [8344e69de417470b7dfbf7b29d490977854c5ddd](https://github.com/cockroachdb/cockroach/commits/8344e69de417470b7dfbf7b29d490977854c5ddd):
```
=== RUN TestLeaseTxnDeadlineExtension/validate-lease-txn-deadline-ext
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTxnDeadlineExtension.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
sql/catalog/lease: TestLeaseTxnDeadlineExtension failed - sql/catalog/lease.TestLeaseTxnDeadlineExtension [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/5396090?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/5396090?buildTab=artifacts#/) on master @ [8344e69de417470b7dfbf7b29d490977854c5ddd](https://github.com/cockroachdb/cockroach/commits/8344e69de417470b7dfbf7b29d490977854c5ddd):
```
=== RUN TestLeaseTxnDeadlineExtension/validate-lease-txn-deadline-ext
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTxnDeadlineExtension.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
sql catalog lease testleasetxndeadlineextension failed sql catalog lease testleasetxndeadlineextension with on master run testleasetxndeadlineextension validate lease txn deadline ext parameters tags bazel gss help see also cc cockroachdb sql schema
| 1
|
51,199
| 6,150,831,797
|
IssuesEvent
|
2017-06-27 23:57:22
|
mozilla/activity-stream
|
https://api.github.com/repos/mozilla/activity-stream
|
closed
|
The "active" property in experiments has an unclear meaning
|
Chore P2 Test Pilot
|
Right now it's unclear what we are using the `"active"` field on experiments for -- this came in a with a discussion with @emtwo. In practice we use it for "archiving" old experiments that may not even be referenced in the code, whereas for experiments that are not yet running, we tend to just set the `"threshold"` value to `0`.
We should clarify this distinction.
|
1.0
|
The "active" property in experiments has an unclear meaning - Right now it's unclear what we are using the `"active"` field on experiments for -- this came in a with a discussion with @emtwo. In practice we use it for "archiving" old experiments that may not even be referenced in the code, whereas for experiments that are not yet running, we tend to just set the `"threshold"` value to `0`.
We should clarify this distinction.
|
test
|
the active property in experiments has an unclear meaning right now it s unclear what we are using the active field on experiments for this came in a with a discussion with emtwo in practice we use it for archiving old experiments that may not even be referenced in the code whereas for experiments that are not yet running we tend to just set the threshold value to we should clarify this distinction
| 1
|
116,700
| 24,969,120,743
|
IssuesEvent
|
2022-11-01 22:28:19
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
Add a reference to System.Linq automatically when LINQ is used
|
Area-IDE Feature Request IDE-CodeStyle
|
_This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/content/idea/1035070/add-a-reference-to-systemlinq-automatically-when-l.html)._
---
VS2019 has brought in an extremely helpful feature; if you use an item (for example an attribute) that's in a non-referenced namespace, the approprtiate 'using' statement is automatically added.
It would be great if VS could automatically add 'using System.Linq' In a similar fashion, if you start to write a Linq statement.
---
### Original Comments
#### Feedback Bot on 5/18/2020, 03:46 AM:
Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further. We'll provide an update once the issue has been triaged by the product team.
|
1.0
|
Add a reference to System.Linq automatically when LINQ is used - _This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/content/idea/1035070/add-a-reference-to-systemlinq-automatically-when-l.html)._
---
VS2019 has brought in an extremely helpful feature; if you use an item (for example an attribute) that's in a non-referenced namespace, the approprtiate 'using' statement is automatically added.
It would be great if VS could automatically add 'using System.Linq' In a similar fashion, if you start to write a Linq statement.
---
### Original Comments
#### Feedback Bot on 5/18/2020, 03:46 AM:
Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further. We'll provide an update once the issue has been triaged by the product team.
|
non_test
|
add a reference to system linq automatically when linq is used this issue has been moved from has brought in an extremely helpful feature if you use an item for example an attribute that s in a non referenced namespace the approprtiate using statement is automatically added it would be great if vs could automatically add using system linq in a similar fashion if you start to write a linq statement original comments feedback bot on am thank you for taking the time to provide your suggestion we will do some preliminary checks to make sure we can proceed further we ll provide an update once the issue has been triaged by the product team
| 0
|
341,247
| 30,575,803,242
|
IssuesEvent
|
2023-07-21 05:15:52
|
ChainSafe/gossamer
|
https://api.github.com/repos/ChainSafe/gossamer
|
closed
|
Add `go test -race` workflows
|
tests Type: Feature
|
## Task summary
<!-- A clear and concise description of what the task is. -->
- We don't run tests with -race automatically in CI.
- There is only `test-state-race` which tests only for dot/state. We have data race tests in `dot/sync` and `lib/blocktree`, but they are not running with -race at the moment.
So, add CI workflows to run data race tests.
## Specification reference
<!-- Provide a reference to the specification as to what is being implemented. -->
-
## Other information and links
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
Based on discussion happened in
- https://github.com/ChainSafe/gossamer/pull/2923#discussion_r1021327464
- https://github.com/ChainSafe/gossamer/pull/2923#discussion_r1021490441
<!-- Thank you 💪 -->
|
1.0
|
Add `go test -race` workflows - ## Task summary
<!-- A clear and concise description of what the task is. -->
- We don't run tests with -race automatically in CI.
- There is only `test-state-race` which tests only for dot/state. We have data race tests in `dot/sync` and `lib/blocktree`, but they are not running with -race at the moment.
So, add CI workflows to run data race tests.
## Specification reference
<!-- Provide a reference to the specification as to what is being implemented. -->
-
## Other information and links
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
Based on discussion happened in
- https://github.com/ChainSafe/gossamer/pull/2923#discussion_r1021327464
- https://github.com/ChainSafe/gossamer/pull/2923#discussion_r1021490441
<!-- Thank you 💪 -->
|
test
|
add go test race workflows task summary we don t run tests with race automatically in ci there is only test state race which tests only for dot state we have data race tests in dot sync and lib blocktree but they are not running with race at the moment so add ci workflows to run data race tests specification reference other information and links based on discussion happened in
| 1
|
305
| 2,736,749,196
|
IssuesEvent
|
2015-04-19 18:49:31
|
ChelseaStats/issues
|
https://api.github.com/repos/ChelseaStats/issues
|
closed
|
football_league April 17 2015 at 11:05AM
|
to process tweet
|
<blockquote class="twitter-tweet">
<p>So which of the three shortlisted players do you think should win the prize at the <a href="http://u.thechels.uk/1b9vjTj">#FLAwards</a> on Sunday evening? <a href="http://u.thechels.uk/1IloIiN">http://u.thechels.uk/1b9vjTk</a></p>
— The Football League (@football_league) <a href="http://u.thechels.uk/1IloCI7">April 17, 2015</a>
</blockquote>
<br><br>
April 17, 2015 at 11:05AM<br>
via Twitter<br><hr><br><br>
http://u.thechels.uk/1b9vj5z
|
1.0
|
football_league April 17 2015 at 11:05AM - <blockquote class="twitter-tweet">
<p>So which of the three shortlisted players do you think should win the prize at the <a href="http://u.thechels.uk/1b9vjTj">#FLAwards</a> on Sunday evening? <a href="http://u.thechels.uk/1IloIiN">http://u.thechels.uk/1b9vjTk</a></p>
— The Football League (@football_league) <a href="http://u.thechels.uk/1IloCI7">April 17, 2015</a>
</blockquote>
<br><br>
April 17, 2015 at 11:05AM<br>
via Twitter<br><hr><br><br>
http://u.thechels.uk/1b9vj5z
|
non_test
|
football league april at so which of the three shortlisted players do you think should win the prize at the a href on sunday evening a href mdash the football league football league april at via twitter
| 0
|
311,190
| 26,774,544,825
|
IssuesEvent
|
2023-01-31 16:14:38
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
Latest Posts and More Block
|
[Type] Enhancement [Status] Needs More Info [Status] Stale [Block] Latest Posts
|
In my organization, it has been the policy to always include basic information about any partner supplying the original information contained in a news posting. To that end, we have always included a "More..." block in our posts. On our home page, we would like the Latest Posts block to display information in complete chunks with no character named an "ellipsis" in sight.
On a home page with a nested loop (either Post List or Latest Posts block) of five additional posts, I'm having to decide between showing the full post content, including the information about our partners, or the excerpt stopping at the very ugly three-fanged ellipsis. What I actually want is to display the post content as it would be displayed on the default WordPress homepage. Is there anybody else who can see the benefit of refactoring this piece of code of the WordPress homepage into a function leveraged by both the homepage and a Gutenberg block?
|
1.0
|
Latest Posts and More Block - In my organization, it has been the policy to always include basic information about any partner supplying the original information contained in a news posting. To that end, we have always included a "More..." block in our posts. On our home page, we would like the Latest Posts block to display information in complete chunks with no character named an "ellipsis" in sight.
On a home page with a nested loop (either Post List or Latest Posts block) of five additional posts, I'm having to decide between showing the full post content, including the information about our partners, or the excerpt stopping at the very ugly three-fanged ellipsis. What I actually want is to display the post content as it would be displayed on the default WordPress homepage. Is there anybody else who can see the benefit of refactoring this piece of code of the WordPress homepage into a function leveraged by both the homepage and a Gutenberg block?
|
test
|
latest posts and more block in my organization it has been the policy to always include basic information about any partner supplying the original information contained in a news posting to that end we have always included a more block in our posts on our home page we would like the latest posts block to display information in complete chunks with no character named an ellipsis in sight on a home page with a nested loop either post list or latest posts block of five additional posts i m having to decide between showing the full post content including the information about our partners or the excerpt stopping at the very ugly three fanged ellipsis what i actually want is to display the post content as it would be displayed on the default wordpress homepage is there anybody else who can see the benefit of refactoring this piece of code of the wordpress homepage into a function leveraged by both the homepage and a gutenberg block
| 1
|
177,724
| 13,744,802,381
|
IssuesEvent
|
2020-10-06 01:00:39
|
codewars/codewars-runner-cli
|
https://api.github.com/repos/codewars/codewars-runner-cli
|
closed
|
Add test support for SASS/SCSS
|
request/language request/test-framework
|
I've been recently experimenting with a test framework for SASS called true. It would be nice to have it included in the Codewars environment.
Characteristics:
* Open Source ( BSD-3 License )
* node module
* can be called from mocha as `var sassTrue = require('sass-true');`
|
1.0
|
Add test support for SASS/SCSS - I've been recently experimenting with a test framework for SASS called true. It would be nice to have it included in the Codewars environment.
Characteristics:
* Open Source ( BSD-3 License )
* node module
* can be called from mocha as `var sassTrue = require('sass-true');`
|
test
|
add test support for sass scss i ve been recently experimenting with a test framework for sass called true it would be nice to have it included in the codewars environment characteristics open source bsd license node module can be called from mocha as var sasstrue require sass true
| 1
|
490,862
| 14,141,082,710
|
IssuesEvent
|
2020-11-10 12:10:41
|
Puzzlepart/prosjektportalen365
|
https://api.github.com/repos/Puzzlepart/prosjektportalen365
|
closed
|
Enable project extensions by default when creating new project
|
Complexity: medium Priority: low enhancement
|
**Describe the solution you'd like**
Boolean to enable/disable project extensions by default. Identical to PP2.
**Additional context**

|
1.0
|
Enable project extensions by default when creating new project - **Describe the solution you'd like**
Boolean to enable/disable project extensions by default. Identical to PP2.
**Additional context**

|
non_test
|
enable project extensions by default when creating new project describe the solution you d like boolean to enable disable project extensions by default identical to additional context
| 0
|
147,613
| 5,642,827,396
|
IssuesEvent
|
2017-04-06 22:08:53
|
Azure/acs-engine
|
https://api.github.com/repos/Azure/acs-engine
|
closed
|
acstgen: kubernetes: consider switching to `kubeadm`
|
orchestrator/k8s priority/P2
|
Filing this for ongoing and future discussion/consideration.
This could dramatically reduce the amount of code we maintain to deploy Kubernetes.
Advantages:
- removing 90+% of the yaml we have in the repo for Kubernetes (cloud-config, certs, etc)
- not having to worry about drift between upstream addons and our addon yamls
Disadvantages:
- no HA option in kubeadm (if it even matters)
- no easy way to get a `kubeconfig` out (yet)
- unclear which addons are deployed
|
1.0
|
acstgen: kubernetes: consider switching to `kubeadm` - Filing this for ongoing and future discussion/consideration.
This could dramatically reduce the amount of code we maintain to deploy Kubernetes.
Advantages:
- removing 90+% of the yaml we have in the repo for Kubernetes (cloud-config, certs, etc)
- not having to worry about drift between upstream addons and our addon yamls
Disadvantages:
- no HA option in kubeadm (if it even matters)
- no easy way to get a `kubeconfig` out (yet)
- unclear which addons are deployed
|
non_test
|
acstgen kubernetes consider switching to kubeadm filing this for ongoing and future discussion consideration this could dramatically reduce the amount of code we maintain to deploy kubernetes advantages removing of the yaml we have in the repo for kubernetes cloud config certs etc not having to worry about drift between upstream addons and our addon yamls disadvantages no ha option in kubeadm if it even matters no easy way to get a kubeconfig out yet unclear which addons are deployed
| 0
|
248,267
| 18,858,055,516
|
IssuesEvent
|
2021-11-12 09:20:06
|
frederickpek/pe
|
https://api.github.com/repos/frederickpek/pe
|
opened
|
Developer Guide: Use Cases not numbered
|
type.DocumentationBug severity.Medium
|
This can be very confusing for developer trying to follow the use cases, especially when inclusions are involved.<br>


<!--session: 1636703094898-26dfe01e-bab1-430f-b90f-6e347ccd8d84-->
<!--Version: Web v3.4.1-->
|
1.0
|
Developer Guide: Use Cases not numbered - This can be very confusing for developer trying to follow the use cases, especially when inclusions are involved.<br>


<!--session: 1636703094898-26dfe01e-bab1-430f-b90f-6e347ccd8d84-->
<!--Version: Web v3.4.1-->
|
non_test
|
developer guide use cases not numbered this can be very confusing for developer trying to follow the use cases especially when inclusions are involved
| 0
|
731,311
| 25,209,572,304
|
IssuesEvent
|
2022-11-14 01:41:09
|
OpenMined/PySyft
|
https://api.github.com/repos/OpenMined/PySyft
|
opened
|
Add DataSubjectRegistry
|
Type: Improvement :chart_with_upwards_trend: Priority: 1 - Immediate :fire: 0.7
|
## Description
The current data structure for `Data Subjects` is not going to scale well and is already causing some issues in the internals of the `DataSubjectLedger`.
We should revert to using unique sequential `int` values from 0 onwards for the code backed by a lookup in the database with a unique int id primary key.
`Table`: `data_subject_registry`:
`id` int64 primary key UNIQUE (int8 or bigint)
`data_subject` str UNIQUE 255
add index for the data_subject column so we can query it
`created_at` timestamp
Private server side methods:
`get_all_ids` returns the entire table as a `np.ndarray[str]` array of strings from 0 onwards
this method needs to be paginateable so select from id 0 -> 1 million
`get_all_data_subjects` returns the entire table as a `Dict[str, int]` mapping, this method needs to be paginateable so like select from id 0 -> 1 million
`get_all_ids_and_data_subjects` returns both as a Tuple
`register_data_subjects` takes `UnindexedDataSubjects` and does all the work to deduplicate and then insert them into the db returning `IndexedDataSubjects` object.
needs to be paginateable so like select from id 0 -> 1 million
During `load_dataset` we need to register these data subjects by sending the unique ones to the service. Anything not already in the table (perhaps do a `get_all_data_subjects`) can be passed to the `insert` function. Exceptions for duplicate rejection should be handled.
This means the DPTensors need to have two types of `DataSubject` objects.
`IndexedDataSubjects` and `UnindexedDataSubjects`.
When users submit tensors they can have either but if they have `UnindexedDataSubjects` they need to be swapped out after Registration.
Question: @iamtrask Do we need to prevent non `DO`s from registering DataSubjects even though the internal ids are always hidden. In the case of some kind of leak of information, the ability to upload names or ids to see what they are converted to could be abused.
## Acceptance Criteria
We need to have some tests which verify that these methods work and the `UNIQUE` attributes of the `id` and `str` are maintained.
- ids should never be reused or recycled (so no row deletion)
- data_subject strings need to be unique otherwise we will get duplicates
|
1.0
|
Add DataSubjectRegistry - ## Description
The current data structure for `Data Subjects` is not going to scale well and is already causing some issues in the internals of the `DataSubjectLedger`.
We should revert to using unique sequential `int` values from 0 onwards for the code backed by a lookup in the database with a unique int id primary key.
`Table`: `data_subject_registry`:
`id` int64 primary key UNIQUE (int8 or bigint)
`data_subject` str UNIQUE 255
add index for the data_subject column so we can query it
`created_at` timestamp
Private server side methods:
`get_all_ids` returns the entire table as a `np.ndarray[str]` array of strings from 0 onwards
this method needs to be paginateable so select from id 0 -> 1 million
`get_all_data_subjects` returns the entire table as a `Dict[str, int]` mapping, this method needs to be paginateable so like select from id 0 -> 1 million
`get_all_ids_and_data_subjects` returns both as a Tuple
`register_data_subjects` takes `UnindexedDataSubjects` and does all the work to deduplicate and then insert them into the db returning `IndexedDataSubjects` object.
needs to be paginateable so like select from id 0 -> 1 million
During `load_dataset` we need to register these data subjects by sending the unique ones to the service. Anything not already in the table (perhaps do a `get_all_data_subjects`) can be passed to the `insert` function. Exceptions for duplicate rejection should be handled.
This means the DPTensors need to have two types of `DataSubject` objects.
`IndexedDataSubjects` and `UnindexedDataSubjects`.
When users submit tensors they can have either but if they have `UnindexedDataSubjects` they need to be swapped out after Registration.
Question: @iamtrask Do we need to prevent non `DO`s from registering DataSubjects even though the internal ids are always hidden. In the case of some kind of leak of information, the ability to upload names or ids to see what they are converted to could be abused.
## Acceptance Criteria
We need to have some tests which verify that these methods work and the `UNIQUE` attributes of the `id` and `str` are maintained.
- ids should never be reused or recycled (so no row deletion)
- data_subject strings need to be unique otherwise we will get duplicates
|
non_test
|
add datasubjectregistry description the current data structure for data subjects is not going to scale well and is already causing some issues in the internals of the datasubjectledger we should revert to using unique sequential int values from onwards for the code backed by a lookup in the database with a unique int id primary key table data subject registry id primary key unique or bigint data subject str unique add index for the data subject column so we can query it created at timestamp private server side methods get all ids returns the entire table as a np ndarray array of strings from onwards this method needs to be paginateable so select from id million get all data subjects returns the entire table as a dict mapping this method needs to be paginateable so like select from id million get all ids and data subjects returns both as a tuple register data subjects takes unindexeddatasubjects and does all the work to deduplicate and then insert them into the db returning indexeddatasubjects object needs to be paginateable so like select from id million during load dataset we need to register these data subjects by sending the unique ones to the service anything not already in the table perhaps do a get all data subjects can be passed to the insert function exceptions for duplicate rejection should be handled this means the dptensors need to have two types of datasubject objects indexeddatasubjects and unindexeddatasubjects when users submit tensors they can have either but if they have unindexeddatasubjects they need to be swapped out after registration question iamtrask do we need to prevent non do s from registering datasubjects even though the internal ids are always hidden in the case of some kind of leak of information the ability to upload names or ids to see what they are converted to could be abused acceptance criteria we need to have some tests which verify that these methods work and the unique attributes of the id and str are maintained ids should never be reused or recycled so no row deletion data subject strings need to be unique otherwise we will get duplicates
| 0
|
86,633
| 8,042,425,910
|
IssuesEvent
|
2018-07-31 08:06:41
|
brave/browser-laptop
|
https://api.github.com/repos/brave/browser-laptop
|
closed
|
Intermittent failure on travis with fingerprinting 2 not 3
|
OS/unix-like/linux automated-tests bug stale upstream
|
There's some kind of problem with travis and fingerprinting test with a permfail on the `blocks fingerprinting` test. This value should be 3 but it returns 2 always on travis.
For now I check for either of them but we should fix this or adjust the test to account for what's wrong. I think it's the latter.
#2626
|
1.0
|
Intermittent failure on travis with fingerprinting 2 not 3 - There's some kind of problem with travis and fingerprinting test with a permfail on the `blocks fingerprinting` test. This value should be 3 but it returns 2 always on travis.
For now I check for either of them but we should fix this or adjust the test to account for what's wrong. I think it's the latter.
#2626
|
test
|
intermittent failure on travis with fingerprinting not there s some kind of problem with travis and fingerprinting test with a permfail on the blocks fingerprinting test this value should be but it returns always on travis for now i check for either of them but we should fix this or adjust the test to account for what s wrong i think it s the latter
| 1
|
122,367
| 10,220,730,139
|
IssuesEvent
|
2019-08-15 22:19:35
|
sylabs/singularity
|
https://api.github.com/repos/sylabs/singularity
|
closed
|
[E2E] Add tests for key importing/exporting
|
Release 3.4 Testing e2e
|
### Expected behavior
CCI/Travis is testing the key importing/exporting.
### Actual behavior
CCI/Travis does not test the key import/export.
|
1.0
|
[E2E] Add tests for key importing/exporting - ### Expected behavior
CCI/Travis is testing the key importing/exporting.
### Actual behavior
CCI/Travis does not test the key import/export.
|
test
|
add tests for key importing exporting expected behavior cci travis is testing the key importing exporting actual behavior cci travis does not test the key import export
| 1
|
630,699
| 20,116,675,614
|
IssuesEvent
|
2022-02-07 20:18:53
|
minio/minio
|
https://api.github.com/repos/minio/minio
|
closed
|
AWS S3 Gateway and Combination of Cache generates I/O errors on Ranged Request
|
community priority: medium
|
Hi, we are experiencing something similar to https://github.com/minio/minio/issues/8504 but using `docker-compose` with AWS S3 Gateway in Combination with Caching.
Never before without caching. Errors only happen when a RANGED request is issued and are all of `read/write on closed pipe` type
Setting `MINIO_CACHE_RANGE: "off"` eliminates the errors.
e.g
```Shell
API: GetObject(bucket=archipelago-hosting-1c090abe-2a7e-4d2a-9409-5d05e67b93fe, object=media/19e/application-altered-states-8d716848-3691-4aa8-9e77-eeb5f9d9c1da.wacz)
Time: 14:50:30 UTC 01/28/2022
DeploymentID: 3dab5c9d-eca5-448b-821b-81165e87d576
RequestID: 16CE76B449977CA3
RemoteHost: 172.18.0.4
Host: esmero-minio:9000
UserAgent: aws-sdk-php/3.204.2 OS/Linux/5.10.82-83.359.amzn2.aarch64 lang/php/7.4.20 GuzzleHttp/6.5.5 curl/7.77.0 PHP/7.4.20
Error: io: read/write on closed pipe (*errors.errorString)
5: cmd/api-errors.go:2103:cmd.toAPIErrorCode()
4: cmd/api-errors.go:2128:cmd.toAPIError()
3: cmd/object-handlers.go:517:cmd.objectAPIHandlers.getObjectHandler()
2: cmd/object-handlers.go:575:cmd.objectAPIHandlers.GetObjectHandler()
1: net/http/server.go:2047:http.HandlerFunc.ServeHTTP()
```
## Expected Behavior
No errors on ranged requests on cached Objects.
## Current Behavior
Error: io: read/write on closed pipe when issuing a Ranged request. This errors are permeating into our Client APP (using AWS S3 Client PHP library)
## Possible Solution
## Steps to Reproduce (for bugs)
1. Upload an OBJECT to an AWS S3 Bucket
2. Setup MIN.IO to use S3 AWS Gateway in combination with default cache (Using e.g a standard IAM inline policy that ALLOW access to Get*, List*, Abort* to that bucket and bucket/* + general applicable to any resources with s3:ListAllMyBuckets
3. ISSUE ranged request to that OBJECT with offsets
4. Look at your minio logs (docker logs -f)
## Context
We have ranged requests to S3 to allow very large `WACZ` files (Zipped Webarchive with a TOC that allows us identify resource offsets and then request those via normal RANGED http request) to be fetched in real time. Without Caching enabled these ranged requests work perfectly and gateway (and also when using local MIN.IO storage) work well.
## Regression
No (but not sure, first time we test with Caching in production)
## Your Environment
* Version used: minio version RELEASE.2022-01-08T03-11-54Z
* Server setup and configuration: Via docker-compose, AWS S3 Gateway, Basic Cache one (including ranged requests) with a dedicated partition (correctly formatted/mounted) for this.
* Operating System and version (`uname -a`): aarch64 aarch64 aarch64 GNU/Linux
|
1.0
|
AWS S3 Gateway and Combination of Cache generates I/O errors on Ranged Request - Hi, we are experiencing something similar to https://github.com/minio/minio/issues/8504 but using `docker-compose` with AWS S3 Gateway in Combination with Caching.
Never before without caching. Errors only happen when a RANGED request is issued and are all of `read/write on closed pipe` type
Setting `MINIO_CACHE_RANGE: "off"` eliminates the errors.
e.g
```Shell
API: GetObject(bucket=archipelago-hosting-1c090abe-2a7e-4d2a-9409-5d05e67b93fe, object=media/19e/application-altered-states-8d716848-3691-4aa8-9e77-eeb5f9d9c1da.wacz)
Time: 14:50:30 UTC 01/28/2022
DeploymentID: 3dab5c9d-eca5-448b-821b-81165e87d576
RequestID: 16CE76B449977CA3
RemoteHost: 172.18.0.4
Host: esmero-minio:9000
UserAgent: aws-sdk-php/3.204.2 OS/Linux/5.10.82-83.359.amzn2.aarch64 lang/php/7.4.20 GuzzleHttp/6.5.5 curl/7.77.0 PHP/7.4.20
Error: io: read/write on closed pipe (*errors.errorString)
5: cmd/api-errors.go:2103:cmd.toAPIErrorCode()
4: cmd/api-errors.go:2128:cmd.toAPIError()
3: cmd/object-handlers.go:517:cmd.objectAPIHandlers.getObjectHandler()
2: cmd/object-handlers.go:575:cmd.objectAPIHandlers.GetObjectHandler()
1: net/http/server.go:2047:http.HandlerFunc.ServeHTTP()
```
## Expected Behavior
No errors on ranged requests on cached Objects.
## Current Behavior
Error: io: read/write on closed pipe when issuing a Ranged request. This errors are permeating into our Client APP (using AWS S3 Client PHP library)
## Possible Solution
## Steps to Reproduce (for bugs)
1. Upload an OBJECT to an AWS S3 Bucket
2. Setup MIN.IO to use S3 AWS Gateway in combination with default cache (Using e.g a standard IAM inline policy that ALLOW access to Get*, List*, Abort* to that bucket and bucket/* + general applicable to any resources with s3:ListAllMyBuckets
3. ISSUE ranged request to that OBJECT with offsets
4. Look at your minio logs (docker logs -f)
## Context
We have ranged requests to S3 to allow very large `WACZ` files (Zipped Webarchive with a TOC that allows us identify resource offsets and then request those via normal RANGED http request) to be fetched in real time. Without Caching enabled these ranged requests work perfectly and gateway (and also when using local MIN.IO storage) work well.
## Regression
No (but not sure, first time we test with Caching in production)
## Your Environment
* Version used: minio version RELEASE.2022-01-08T03-11-54Z
* Server setup and configuration: Via docker-compose, AWS S3 Gateway, Basic Cache one (including ranged requests) with a dedicated partition (correctly formatted/mounted) for this.
* Operating System and version (`uname -a`): aarch64 aarch64 aarch64 GNU/Linux
|
non_test
|
aws gateway and combination of cache generates i o errors on ranged request hi we are experiencing something similar to but using docker compose with aws gateway in combination with caching never before without caching errors only happen when a ranged request is issued and are all of read write on closed pipe type setting minio cache range off eliminates the errors e g shell api getobject bucket archipelago hosting object media application altered states wacz time utc deploymentid requestid remotehost host esmero minio useragent aws sdk php os linux lang php guzzlehttp curl php error io read write on closed pipe errors errorstring cmd api errors go cmd toapierrorcode cmd api errors go cmd toapierror cmd object handlers go cmd objectapihandlers getobjecthandler cmd object handlers go cmd objectapihandlers getobjecthandler net http server go http handlerfunc servehttp expected behavior no errors on ranged requests on cached objects current behavior error io read write on closed pipe when issuing a ranged request this errors are permeating into our client app using aws client php library possible solution steps to reproduce for bugs upload an object to an aws bucket setup min io to use aws gateway in combination with default cache using e g a standard iam inline policy that allow access to get list abort to that bucket and bucket general applicable to any resources with listallmybuckets issue ranged request to that object with offsets look at your minio logs docker logs f context we have ranged requests to to allow very large wacz files zipped webarchive with a toc that allows us identify resource offsets and then request those via normal ranged http request to be fetched in real time without caching enabled these ranged requests work perfectly and gateway and also when using local min io storage work well regression no but not sure first time we test with caching in production your environment version used minio version release server setup and configuration via docker compose aws gateway basic cache one including ranged requests with a dedicated partition correctly formatted mounted for this operating system and version uname a gnu linux
| 0
|
196,192
| 14,839,610,100
|
IssuesEvent
|
2021-01-16 01:28:20
|
USGS-Astrogeology/PDS-Pipelines
|
https://api.github.com/repos/USGS-Astrogeology/PDS-Pipelines
|
opened
|
Write test for the pvl_utils module
|
Tests
|
Current the pvl utils module functions are untested and it's likely that the pvl_load function is unnecessary with the update to pvl 1.0.0. If this function is removed that leaves two other functions to be tested. If not, then all three need to be tested.
|
1.0
|
Write test for the pvl_utils module - Current the pvl utils module functions are untested and it's likely that the pvl_load function is unnecessary with the update to pvl 1.0.0. If this function is removed that leaves two other functions to be tested. If not, then all three need to be tested.
|
test
|
write test for the pvl utils module current the pvl utils module functions are untested and it s likely that the pvl load function is unnecessary with the update to pvl if this function is removed that leaves two other functions to be tested if not then all three need to be tested
| 1
|
297,581
| 25,744,260,635
|
IssuesEvent
|
2022-12-08 08:46:49
|
tshipenchko/gdsc-tgb-contest
|
https://api.github.com/repos/tshipenchko/gdsc-tgb-contest
|
opened
|
Boring /top
|
enhancement contest-task
|
It's kind of boring. The leaderboard consists simply of user IDs. It's boring and generally very strange. Why would an ordinary bot user see some numbers?
|
1.0
|
Boring /top - It's kind of boring. The leaderboard consists simply of user IDs. It's boring and generally very strange. Why would an ordinary bot user see some numbers?
|
test
|
boring top it s kind of boring the leaderboard consists simply of user ids it s boring and generally very strange why would an ordinary bot user see some numbers
| 1
|
269,119
| 23,420,964,389
|
IssuesEvent
|
2022-08-13 17:39:08
|
benoitkugler/maths-online
|
https://api.github.com/repos/benoitkugler/maths-online
|
closed
|
[prof] Variable sur préfixe fonctions
|
enhancement A tester
|
Le nom de la fonction dans le préfixe d'un champ de réponse expression n'est pas reconnue (&f& ne lit pas la variable f)
<img width="684" alt="isyro" src="https://user-images.githubusercontent.com/105806953/180877945-ebd97228-68d9-4b86-a9cc-f826d3d6a2c0.PNG">
|
1.0
|
[prof] Variable sur préfixe fonctions - Le nom de la fonction dans le préfixe d'un champ de réponse expression n'est pas reconnue (&f& ne lit pas la variable f)
<img width="684" alt="isyro" src="https://user-images.githubusercontent.com/105806953/180877945-ebd97228-68d9-4b86-a9cc-f826d3d6a2c0.PNG">
|
test
|
variable sur préfixe fonctions le nom de la fonction dans le préfixe d un champ de réponse expression n est pas reconnue f ne lit pas la variable f img width alt isyro src
| 1
|
168,013
| 13,055,646,267
|
IssuesEvent
|
2020-07-30 02:18:58
|
vmware-tanzu/antrea
|
https://api.github.com/repos/vmware-tanzu/antrea
|
opened
|
[e2e] t.Error for cleanup functions
|
area/test/e2e good first issue kind/feature
|
**Describe the problem/challenge you have**
Suggested by @antoninbas , it is better to use `t.Error` instead of `t.Fatal` in cleanup functions.
**Describe the solution you'd like**
Reconsider the usage of `t.Error` and `t.Fatal` in e2e tests.
|
1.0
|
[e2e] t.Error for cleanup functions - **Describe the problem/challenge you have**
Suggested by @antoninbas , it is better to use `t.Error` instead of `t.Fatal` in cleanup functions.
**Describe the solution you'd like**
Reconsider the usage of `t.Error` and `t.Fatal` in e2e tests.
|
test
|
t error for cleanup functions describe the problem challenge you have suggested by antoninbas it is better to use t error instead of t fatal in cleanup functions describe the solution you d like reconsider the usage of t error and t fatal in tests
| 1
|
102,974
| 8,872,401,074
|
IssuesEvent
|
2019-01-11 15:20:14
|
paritytech/substrate
|
https://api.github.com/repos/paritytech/substrate
|
opened
|
Restore integration tests
|
F4-tests Q3-medium
|
Integration tests introduced in #805 were lost somewhere when moving to aura. Would be nice to re-introduce them.
|
1.0
|
Restore integration tests - Integration tests introduced in #805 were lost somewhere when moving to aura. Would be nice to re-introduce them.
|
test
|
restore integration tests integration tests introduced in were lost somewhere when moving to aura would be nice to re introduce them
| 1
|
290,725
| 25,090,126,260
|
IssuesEvent
|
2022-11-08 05:09:50
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
[Bug]: Report Generation Failure for ballerinax/postgresql with the New Testerina Framework
|
Type/Bug Team/DevTools Area/TestFramework Reason/EngineeringMistake
|
### Description
With the new testerina framework, the postgres module test report generation fails with the following output.
```
[fail] testInOutParameterArray:
Timestamp timezone range array does not match.
expected: <postgresql:TimestamptzRange[]> '[{"upper":"2010-01-01 14:00:00+00","lower":"2010-01-01 12:00:00+00","upperboundI
nclusive":true,"lowerboundInclusive":false}]'
actual : <(postgresql:TimestamptzRange & readonly)[] & readonly> '[{"upper":"2010-01-01 19:30:00+05:30","lower":"2010-01-01 17:30:00+05:30","upper
boundInclusive":true,"lowerboundInclusive":false}]'
382 passing
1 failing
0 skipped
ballerina: Oh no, something really went wrong. Bad. Sad.
We appreciate it if you can report the code that broke Ballerina in
https://github.com/ballerina-platform/ballerina-lang/issues with the
log you get below and your sample code.
We thank you for helping make us better.
[2022-11-01 10:17:35,087] SEVERE {b7a.log.crash} - com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
com.google.gson.JsonSyntaxException: com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
at com.google.gson.Gson.fromJson(Gson.java:978)
at com.google.gson.Gson.fromJson(Gson.java:901)
at io.ballerina.cli.utils.TestUtils.loadModuleStatusFromFile(TestUtils.java:222)
at io.ballerina.cli.task.RunTestsTask.execute(RunTestsTask.java:219)
at io.ballerina.cli.TaskExecutor.executeTasks(TaskExecutor.java:40)
at io.ballerina.cli.cmd.TestCommand.execute(TestCommand.java:303)
at java.base/java.util.Optional.ifPresent(Optional.java:183)
at io.ballerina.cli.launcher.Main.main(Main.java:51)
Caused by: com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
at com.google.gson.stream.JsonReader.syntaxError(JsonReader.java:1562)
at com.google.gson.stream.JsonReader.doPeek(JsonReader.java:490)
at com.google.gson.stream.JsonReader.hasNext(JsonReader.java:413)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$Adapter.read(ReflectiveTypeAdapterFactory.java:216)
at com.google.gson.internal.bind.TypeAdapterRuntimeTypeWrapper.read(TypeAdapterRuntimeTypeWrapper.java:41)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:82)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$1.read(ReflectiveTypeAdapterFactory.java:131)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$Adapter.read(ReflectiveTypeAdapterFactory.java:222)
at com.google.gson.Gson.fromJson(Gson.java:963)
... 7 more
```
### Steps to Reproduce
- Build the ballerina-lang repo and publish to the maven local repo.
- Use that lang version to build the ballerinax/postgresql module with tests.
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Test Framework
### Related issue(s) (optional)
https://github.com/ballerina-platform/ballerina-lang/issues/38092
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
1.0
|
[Bug]: Report Generation Failure for ballerinax/postgresql with the New Testerina Framework - ### Description
With the new testerina framework, the postgres module test report generation fails with the following output.
```
[fail] testInOutParameterArray:
Timestamp timezone range array does not match.
expected: <postgresql:TimestamptzRange[]> '[{"upper":"2010-01-01 14:00:00+00","lower":"2010-01-01 12:00:00+00","upperboundI
nclusive":true,"lowerboundInclusive":false}]'
actual : <(postgresql:TimestamptzRange & readonly)[] & readonly> '[{"upper":"2010-01-01 19:30:00+05:30","lower":"2010-01-01 17:30:00+05:30","upper
boundInclusive":true,"lowerboundInclusive":false}]'
382 passing
1 failing
0 skipped
ballerina: Oh no, something really went wrong. Bad. Sad.
We appreciate it if you can report the code that broke Ballerina in
https://github.com/ballerina-platform/ballerina-lang/issues with the
log you get below and your sample code.
We thank you for helping make us better.
[2022-11-01 10:17:35,087] SEVERE {b7a.log.crash} - com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
com.google.gson.JsonSyntaxException: com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
at com.google.gson.Gson.fromJson(Gson.java:978)
at com.google.gson.Gson.fromJson(Gson.java:901)
at io.ballerina.cli.utils.TestUtils.loadModuleStatusFromFile(TestUtils.java:222)
at io.ballerina.cli.task.RunTestsTask.execute(RunTestsTask.java:219)
at io.ballerina.cli.TaskExecutor.executeTasks(TaskExecutor.java:40)
at io.ballerina.cli.cmd.TestCommand.execute(TestCommand.java:303)
at java.base/java.util.Optional.ifPresent(Optional.java:183)
at io.ballerina.cli.launcher.Main.main(Main.java:51)
Caused by: com.google.gson.stream.MalformedJsonException: Unterminated object at line 1 column 21818 path $.tests[382].failureMessage
at com.google.gson.stream.JsonReader.syntaxError(JsonReader.java:1562)
at com.google.gson.stream.JsonReader.doPeek(JsonReader.java:490)
at com.google.gson.stream.JsonReader.hasNext(JsonReader.java:413)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$Adapter.read(ReflectiveTypeAdapterFactory.java:216)
at com.google.gson.internal.bind.TypeAdapterRuntimeTypeWrapper.read(TypeAdapterRuntimeTypeWrapper.java:41)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:82)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$1.read(ReflectiveTypeAdapterFactory.java:131)
at com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$Adapter.read(ReflectiveTypeAdapterFactory.java:222)
at com.google.gson.Gson.fromJson(Gson.java:963)
... 7 more
```
### Steps to Reproduce
- Build the ballerina-lang repo and publish to the maven local repo.
- Use that lang version to build the ballerinax/postgresql module with tests.
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Test Framework
### Related issue(s) (optional)
https://github.com/ballerina-platform/ballerina-lang/issues/38092
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
test
|
report generation failure for ballerinax postgresql with the new testerina framework description with the new testerina framework the postgres module test report generation fails with the following output testinoutparameterarray timestamp timezone range array does not match expected upper lower upperboundi nclusive true lowerboundinclusive false actual upper lower upper boundinclusive true lowerboundinclusive false passing failing skipped ballerina oh no something really went wrong bad sad we appreciate it if you can report the code that broke ballerina in with the log you get below and your sample code we thank you for helping make us better severe log crash com google gson stream malformedjsonexception unterminated object at line column path tests failuremessage com google gson jsonsyntaxexception com google gson stream malformedjsonexception unterminated object at line column path tests failuremessage at com google gson gson fromjson gson java at com google gson gson fromjson gson java at io ballerina cli utils testutils loadmodulestatusfromfile testutils java at io ballerina cli task runteststask execute runteststask java at io ballerina cli taskexecutor executetasks taskexecutor java at io ballerina cli cmd testcommand execute testcommand java at java base java util optional ifpresent optional java at io ballerina cli launcher main main main java caused by com google gson stream malformedjsonexception unterminated object at line column path tests failuremessage at com google gson stream jsonreader syntaxerror jsonreader java at com google gson stream jsonreader dopeek jsonreader java at com google gson stream jsonreader hasnext jsonreader java at com google gson internal bind reflectivetypeadapterfactory adapter read reflectivetypeadapterfactory java at com google gson internal bind typeadapterruntimetypewrapper read typeadapterruntimetypewrapper java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson internal bind reflectivetypeadapterfactory read reflectivetypeadapterfactory java at com google gson internal bind reflectivetypeadapterfactory adapter read reflectivetypeadapterfactory java at com google gson gson fromjson gson java more steps to reproduce build the ballerina lang repo and publish to the maven local repo use that lang version to build the ballerinax postgresql module with tests affected version s no response os db other environment details and versions no response related area test framework related issue s optional suggested label s optional no response suggested assignee s optional no response
| 1
|
77,323
| 14,784,856,191
|
IssuesEvent
|
2021-01-12 01:15:09
|
mangonaise/braincache
|
https://api.github.com/repos/mangonaise/braincache
|
opened
|
Gameplay logic is unnecessarily bundled together with screen-switching logic.
|
nasty code
|
This is an issue with the implementation and not a bug.
Currently the gameplay logic is in App, alongside the more general app logic such as switching screens (e.g. home screen, game over screen) and saving high scores.
The gameplay logic should be separated into its own component.
|
1.0
|
Gameplay logic is unnecessarily bundled together with screen-switching logic. - This is an issue with the implementation and not a bug.
Currently the gameplay logic is in App, alongside the more general app logic such as switching screens (e.g. home screen, game over screen) and saving high scores.
The gameplay logic should be separated into its own component.
|
non_test
|
gameplay logic is unnecessarily bundled together with screen switching logic this is an issue with the implementation and not a bug currently the gameplay logic is in app alongside the more general app logic such as switching screens e g home screen game over screen and saving high scores the gameplay logic should be separated into its own component
| 0
|
80,465
| 7,748,559,145
|
IssuesEvent
|
2018-05-30 08:42:42
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: largerange/splits/size=10GiB,nodes=3 failed on release-2.0
|
C-test-failure O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/32b7aa635af34c5b150abba9df1cd51a5fafe804
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=686347&tab=buildLog
```
cluster.go:678,large_range.go:74,large_range.go:45: /home/agent/work/.go/bin/roachprod start teamcity-686347-largerange-splits-size-10gib-nodes-3:1-3 --encrypt: exit status 1
```
|
1.0
|
roachtest: largerange/splits/size=10GiB,nodes=3 failed on release-2.0 - SHA: https://github.com/cockroachdb/cockroach/commits/32b7aa635af34c5b150abba9df1cd51a5fafe804
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=686347&tab=buildLog
```
cluster.go:678,large_range.go:74,large_range.go:45: /home/agent/work/.go/bin/roachprod start teamcity-686347-largerange-splits-size-10gib-nodes-3:1-3 --encrypt: exit status 1
```
|
test
|
roachtest largerange splits size nodes failed on release sha parameters failed test cluster go large range go large range go home agent work go bin roachprod start teamcity largerange splits size nodes encrypt exit status
| 1
|
75,602
| 14,495,891,053
|
IssuesEvent
|
2020-12-11 11:54:48
|
IgniteUI/igniteui-angular-samples
|
https://api.github.com/repos/IgniteUI/igniteui-angular-samples
|
closed
|
Calendar angular samples are not loaded in code view
|
code-view status: in-review status: resolved
|
- [x] Both samples under Views section are not loaded in Code View:
1. Open https://www.infragistics.com/products/ignite-ui-angular/angular/components/calendar#views
2. See the result:

The "View in full screen button" is working fine.
- [x] Toggle sample is also not loaded -
https://www.infragistics.com/products/ignite-ui-angular/angular/components/toggle
|
1.0
|
Calendar angular samples are not loaded in code view - - [x] Both samples under Views section are not loaded in Code View:
1. Open https://www.infragistics.com/products/ignite-ui-angular/angular/components/calendar#views
2. See the result:

The "View in full screen button" is working fine.
- [x] Toggle sample is also not loaded -
https://www.infragistics.com/products/ignite-ui-angular/angular/components/toggle
|
non_test
|
calendar angular samples are not loaded in code view both samples under views section are not loaded in code view open see the result the view in full screen button is working fine toggle sample is also not loaded
| 0
|
82,139
| 7,818,240,341
|
IssuesEvent
|
2018-06-13 11:39:52
|
hazelcast/hazelcast-jet
|
https://api.github.com/repos/hazelcast/hazelcast-jet
|
opened
|
com.hazelcast.jet.impl.connector.StreamJmsPTest.when_topic
|
test-failure
|
```Error Message
expected:<2db03d3c-ee91-45a4-bac2-ee4d13e79731> but was:<null>
Stacktrace
java.lang.AssertionError: expected:<2db03d3c-ee91-45a4-bac2-ee4d13e79731> but was:<null>
at com.hazelcast.jet.impl.connector.StreamJmsPTest.when_topic(StreamJmsPTest.java:93)```
|
1.0
|
com.hazelcast.jet.impl.connector.StreamJmsPTest.when_topic - ```Error Message
expected:<2db03d3c-ee91-45a4-bac2-ee4d13e79731> but was:<null>
Stacktrace
java.lang.AssertionError: expected:<2db03d3c-ee91-45a4-bac2-ee4d13e79731> but was:<null>
at com.hazelcast.jet.impl.connector.StreamJmsPTest.when_topic(StreamJmsPTest.java:93)```
|
test
|
com hazelcast jet impl connector streamjmsptest when topic error message expected but was stacktrace java lang assertionerror expected but was at com hazelcast jet impl connector streamjmsptest when topic streamjmsptest java
| 1
|
240,904
| 26,256,533,649
|
IssuesEvent
|
2023-01-06 01:34:48
|
farooqmir/React-Redux-Demonstration-with-api
|
https://api.github.com/repos/farooqmir/React-Redux-Demonstration-with-api
|
opened
|
CVE-2021-3803 (High) detected in nth-check-1.0.1.tgz
|
security vulnerability
|
## CVE-2021-3803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.1.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.1.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.1.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- enzyme-3.9.0.tgz (Root Library)
- cheerio-1.0.0-rc.3.tgz
- css-select-1.2.0.tgz
- :x: **nth-check-1.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/farooqmir/React-Redux-Demonstration-with-api/commit/1992b3197c035a77997ce719ef2a2d9c3f4bb0ba">1992b3197c035a77997ce719ef2a2d9c3f4bb0ba</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution (nth-check): 2.0.1</p>
<p>Direct dependency fix Resolution (enzyme): 3.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-3803 (High) detected in nth-check-1.0.1.tgz - ## CVE-2021-3803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.1.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.1.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.1.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- enzyme-3.9.0.tgz (Root Library)
- cheerio-1.0.0-rc.3.tgz
- css-select-1.2.0.tgz
- :x: **nth-check-1.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/farooqmir/React-Redux-Demonstration-with-api/commit/1992b3197c035a77997ce719ef2a2d9c3f4bb0ba">1992b3197c035a77997ce719ef2a2d9c3f4bb0ba</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution (nth-check): 2.0.1</p>
<p>Direct dependency fix Resolution (enzyme): 3.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in nth check tgz cve high severity vulnerability vulnerable library nth check tgz performant nth check parser compiler library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules nth check package json dependency hierarchy enzyme tgz root library cheerio rc tgz css select tgz x nth check tgz vulnerable library found in head commit a href vulnerability details nth check is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution nth check direct dependency fix resolution enzyme step up your open source security game with mend
| 0
|
278,712
| 24,169,612,972
|
IssuesEvent
|
2022-09-22 17:57:15
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
[BUG] Local cluster v2prov non-functional when feature flag multi-cluster-management=false
|
kind/bug [zube]: To Test area/harvester team/area2
|
**Rancher Server Setup**
- Rancher version: v2.6.7
- Installation option (Docker install/Helm Chart): Helm
- If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc): v1.22.12+rke2r1
**User Information**
- What is the role of the user logged in? Admin
**Describe the bug**
When starting rke2 with rancher, helm-operation-xxx pods can't be completed. The `proxy` container is not terminated.


In the rancher pod, it keeps showing:
```
2022/08/24 03:54:04 [ERROR] error syncing 'local': handler rke-control-plane-trigger: Index with name by-cluster does not exist, requeuing
```
rke-control-plan has an error message in condition.
```
configuring bootstrap node(s) custom-51ee9b389b9c: waiting for cluster agent to connect
```
rancher-system-agent also has some error messages.
```
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error while appending ca cert to pool for probe kube-controller-manager"
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error loading CA cert for probe (kube-scheduler) /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: open /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: no such file or directory"
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error while appending ca cert to pool for probe kube-scheduler"
Aug 24 05:44:00 node1 rancher-system-agent[12871]: time="2022-08-24T05:44:00Z" level=error msg="error loading CA cert for probe (kube-scheduler) /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: open /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: no such file or directory"
```
**To Reproduce**
1. Update `settings.yaml` in https://github.com/bk201/vagrant-rancherd.
```
kubernetes_version: v1.22.12+rke2r1
rancher_version: v2.6.7
```
2. Run `vagrant up node1`.
3. Run `./kube-env.sh`.
**Result**
Check pods `KUBECONFIG=./kubeconfig kubectl get pods -n cattle-system`. Some pods can't be completed.
**Expected Result**
All helm-operation-xxx pods can be completed.
|
1.0
|
[BUG] Local cluster v2prov non-functional when feature flag multi-cluster-management=false - **Rancher Server Setup**
- Rancher version: v2.6.7
- Installation option (Docker install/Helm Chart): Helm
- If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc): v1.22.12+rke2r1
**User Information**
- What is the role of the user logged in? Admin
**Describe the bug**
When starting rke2 with rancher, helm-operation-xxx pods can't be completed. The `proxy` container is not terminated.


In the rancher pod, it keeps showing:
```
2022/08/24 03:54:04 [ERROR] error syncing 'local': handler rke-control-plane-trigger: Index with name by-cluster does not exist, requeuing
```
rke-control-plan has an error message in condition.
```
configuring bootstrap node(s) custom-51ee9b389b9c: waiting for cluster agent to connect
```
rancher-system-agent also has some error messages.
```
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error while appending ca cert to pool for probe kube-controller-manager"
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error loading CA cert for probe (kube-scheduler) /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: open /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: no such file or directory"
Aug 24 05:43:55 node1 rancher-system-agent[12871]: time="2022-08-24T05:43:55Z" level=error msg="error while appending ca cert to pool for probe kube-scheduler"
Aug 24 05:44:00 node1 rancher-system-agent[12871]: time="2022-08-24T05:44:00Z" level=error msg="error loading CA cert for probe (kube-scheduler) /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: open /var/lib/rancher/rke2/server/tls/kube-scheduler/kube-scheduler.crt: no such file or directory"
```
**To Reproduce**
1. Update `settings.yaml` in https://github.com/bk201/vagrant-rancherd.
```
kubernetes_version: v1.22.12+rke2r1
rancher_version: v2.6.7
```
2. Run `vagrant up node1`.
3. Run `./kube-env.sh`.
**Result**
Check pods `KUBECONFIG=./kubeconfig kubectl get pods -n cattle-system`. Some pods can't be completed.
**Expected Result**
All helm-operation-xxx pods can be completed.
|
test
|
local cluster non functional when feature flag multi cluster management false rancher server setup rancher version installation option docker install helm chart helm if helm chart kubernetes cluster and version eks etc user information what is the role of the user logged in admin describe the bug when starting with rancher helm operation xxx pods can t be completed the proxy container is not terminated in the rancher pod it keeps showing error syncing local handler rke control plane trigger index with name by cluster does not exist requeuing rke control plan has an error message in condition configuring bootstrap node s custom waiting for cluster agent to connect rancher system agent also has some error messages aug rancher system agent time level error msg error while appending ca cert to pool for probe kube controller manager aug rancher system agent time level error msg error loading ca cert for probe kube scheduler var lib rancher server tls kube scheduler kube scheduler crt open var lib rancher server tls kube scheduler kube scheduler crt no such file or directory aug rancher system agent time level error msg error while appending ca cert to pool for probe kube scheduler aug rancher system agent time level error msg error loading ca cert for probe kube scheduler var lib rancher server tls kube scheduler kube scheduler crt open var lib rancher server tls kube scheduler kube scheduler crt no such file or directory to reproduce update settings yaml in kubernetes version rancher version run vagrant up run kube env sh result check pods kubeconfig kubeconfig kubectl get pods n cattle system some pods can t be completed expected result all helm operation xxx pods can be completed
| 1
|
190,095
| 6,808,665,191
|
IssuesEvent
|
2017-11-04 06:28:38
|
ballerinalang/plugin-intellij
|
https://api.github.com/repos/ballerinalang/plugin-intellij
|
closed
|
Quick doc support not working
|
Priority/Highest Type/Improvement
|
Quick doc support is not working currently because the doc package was move to built-in package.
|
1.0
|
Quick doc support not working - Quick doc support is not working currently because the doc package was move to built-in package.
|
non_test
|
quick doc support not working quick doc support is not working currently because the doc package was move to built in package
| 0
|
66,460
| 7,001,065,264
|
IssuesEvent
|
2017-12-18 08:48:04
|
NativeScript/nativescript-cli
|
https://api.github.com/repos/NativeScript/nativescript-cli
|
closed
|
tns debug ios: nativescript inspector doesn't open on High Sierra and Xcode 9.2
|
bug debug iOS Ready For Test
|
### Tell us about the problem
When run `tns debug ios` it tries to open Nativescript Inspector and immediately closes it and throws an error.
### Which platform(s) does your issue occur on?
iOS
### Please provide the following version numbers that your issue occurs with:
- CLI: 3.3.1
- Cross-platform modules: 3.1.1
- Runtime(s): tns-ios: 3.3.0
### Please tell us how to recreate the issue in as much detail as possible.
I'm trying to debug nativescript tutorial app nativescript-template-ng-tutorial, and when run `tns debug ios` it has problems.
### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem.
https://youtu.be/XgSjqBD3z0M.
On 17th second it opens and closes the nativescript inspector at the same time.
|
1.0
|
tns debug ios: nativescript inspector doesn't open on High Sierra and Xcode 9.2 - ### Tell us about the problem
When run `tns debug ios` it tries to open Nativescript Inspector and immediately closes it and throws an error.
### Which platform(s) does your issue occur on?
iOS
### Please provide the following version numbers that your issue occurs with:
- CLI: 3.3.1
- Cross-platform modules: 3.1.1
- Runtime(s): tns-ios: 3.3.0
### Please tell us how to recreate the issue in as much detail as possible.
I'm trying to debug nativescript tutorial app nativescript-template-ng-tutorial, and when run `tns debug ios` it has problems.
### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem.
https://youtu.be/XgSjqBD3z0M.
On 17th second it opens and closes the nativescript inspector at the same time.
|
test
|
tns debug ios nativescript inspector doesn t open on high sierra and xcode tell us about the problem when run tns debug ios it tries to open nativescript inspector and immediately closes it and throws an error which platform s does your issue occur on ios please provide the following version numbers that your issue occurs with cli cross platform modules runtime s tns ios please tell us how to recreate the issue in as much detail as possible i m trying to debug nativescript tutorial app nativescript template ng tutorial and when run tns debug ios it has problems is there code involved if so please share the minimal amount of code needed to recreate the problem on second it opens and closes the nativescript inspector at the same time
| 1
|
19,374
| 10,349,849,701
|
IssuesEvent
|
2019-09-05 00:11:12
|
uniquelyparticular/sync-moltin-to-algolia
|
https://api.github.com/repos/uniquelyparticular/sync-moltin-to-algolia
|
opened
|
CVE-2018-20834 (High) detected in tar-2.2.1.tgz
|
security vulnerability
|
## CVE-2018-20834 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /sync-moltin-to-algolia/package.json</p>
<p>Path to vulnerable library: /tmp/git/sync-moltin-to-algolia/node_modules/npm/node_modules/node-gyp/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.13.14.tgz (Root Library)
- npm-5.1.7.tgz
- npm-6.9.0.tgz
- node-gyp-3.8.0.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/sync-moltin-to-algolia/commit/63db5619b58f0f651b47ee8c0cf1d64e1c439dff">63db5619b58f0f651b47ee8c0cf1d64e1c439dff</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20834>CVE-2018-20834</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/344595">https://hackerone.com/reports/344595</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: v4.4.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20834 (High) detected in tar-2.2.1.tgz - ## CVE-2018-20834 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /sync-moltin-to-algolia/package.json</p>
<p>Path to vulnerable library: /tmp/git/sync-moltin-to-algolia/node_modules/npm/node_modules/node-gyp/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.13.14.tgz (Root Library)
- npm-5.1.7.tgz
- npm-6.9.0.tgz
- node-gyp-3.8.0.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/sync-moltin-to-algolia/commit/63db5619b58f0f651b47ee8c0cf1d64e1c439dff">63db5619b58f0f651b47ee8c0cf1d64e1c439dff</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20834>CVE-2018-20834</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/344595">https://hackerone.com/reports/344595</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: v4.4.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file sync moltin to algolia package json path to vulnerable library tmp git sync moltin to algolia node modules npm node modules node gyp node modules tar package json dependency hierarchy semantic release tgz root library npm tgz npm tgz node gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details a vulnerability was found in node tar before version an arbitrary file overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system in conjunction with a later plain file with the same name as the hardlink this plain file content replaces the existing file content publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
237,657
| 19,663,719,931
|
IssuesEvent
|
2022-01-10 19:52:45
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
closed
|
Intermittent UI test failure - <Classname.testName>
|
needs:triage eng:ui-test
|
Firefox 95.2.0 android 7.1.2
The logic for working with inserts and bookmarks is terribly inconvenient. In version 68, I opened a bookmark in the current tab if a new tab was not needed. Now the bookmark is forced to open in a new tab and I have to go to the tab bar to remove the unnecessary one 😖 are you kidding me? 😤 make a setting so that you can open a bookmark in the current tab! 2- remove keyboard autostart when opening a new tab! 😤 3 - twenty years I have been use the Internet and have never pressed the button of the house! Make the setting for replacing the house next to the address bar with the refresh / stop button! And 4- add the translate web pages extension already!
Ужасно неудобно сделана логика работы с вкдадками и закладками. В версии 68 я открывал закладку в текущей вкладке, если новая вкладка была не нужна. Теперь же закладка принудительно открывается в новой вкладке и мне приходится идти в панель вкладок, чтобы убрать ненужную 😖 вы издеваетесь? 😤 сделайте настройку, чтобы можно было открывать закладку в текущей вкладке! 2- уберите автозапуск клавиатуры при открытии новой вкладки! 😤 3- двадцать лет пользуюсь интернетом и ни разу не нажимал на кнопку домика! Сделайте настройку замены домика рядом с адресной строкой на кнопку обновить/остановить! И 4- добавьте уже расширение translate web pages!
|
1.0
|
Intermittent UI test failure - <Classname.testName> - Firefox 95.2.0 android 7.1.2
The logic for working with inserts and bookmarks is terribly inconvenient. In version 68, I opened a bookmark in the current tab if a new tab was not needed. Now the bookmark is forced to open in a new tab and I have to go to the tab bar to remove the unnecessary one 😖 are you kidding me? 😤 make a setting so that you can open a bookmark in the current tab! 2- remove keyboard autostart when opening a new tab! 😤 3 - twenty years I have been use the Internet and have never pressed the button of the house! Make the setting for replacing the house next to the address bar with the refresh / stop button! And 4- add the translate web pages extension already!
Ужасно неудобно сделана логика работы с вкдадками и закладками. В версии 68 я открывал закладку в текущей вкладке, если новая вкладка была не нужна. Теперь же закладка принудительно открывается в новой вкладке и мне приходится идти в панель вкладок, чтобы убрать ненужную 😖 вы издеваетесь? 😤 сделайте настройку, чтобы можно было открывать закладку в текущей вкладке! 2- уберите автозапуск клавиатуры при открытии новой вкладки! 😤 3- двадцать лет пользуюсь интернетом и ни разу не нажимал на кнопку домика! Сделайте настройку замены домика рядом с адресной строкой на кнопку обновить/остановить! И 4- добавьте уже расширение translate web pages!
|
test
|
intermittent ui test failure firefox android the logic for working with inserts and bookmarks is terribly inconvenient in version i opened a bookmark in the current tab if a new tab was not needed now the bookmark is forced to open in a new tab and i have to go to the tab bar to remove the unnecessary one 😖 are you kidding me 😤 make a setting so that you can open a bookmark in the current tab remove keyboard autostart when opening a new tab 😤 twenty years i have been use the internet and have never pressed the button of the house make the setting for replacing the house next to the address bar with the refresh stop button and add the translate web pages extension already ужасно неудобно сделана логика работы с вкдадками и закладками в версии я открывал закладку в текущей вкладке если новая вкладка была не нужна теперь же закладка принудительно открывается в новой вкладке и мне приходится идти в панель вкладок чтобы убрать ненужную 😖 вы издеваетесь 😤 сделайте настройку чтобы можно было открывать закладку в текущей вкладке уберите автозапуск клавиатуры при открытии новой вкладки 😤 двадцать лет пользуюсь интернетом и ни разу не нажимал на кнопку домика сделайте настройку замены домика рядом с адресной строкой на кнопку обновить остановить и добавьте уже расширение translate web pages
| 1
|
364,790
| 25,502,573,017
|
IssuesEvent
|
2022-11-28 06:19:50
|
oleksandrblazhko/ai203-tokarev
|
https://api.github.com/repos/oleksandrblazhko/ai203-tokarev
|
closed
|
CW5
|
documentation
|
**Завдання 1**
[Посилання](https://github.com/oleksandrblazhko/ai203-tokarev/blob/main/2-SoftwareDesign/2.7-PlantUML/UML-Activity.md)
|TC id|Опис кроків тестового сценарію|Опис очікуваних результатів|
|-|-|-|
|TC1.1|Початкові умови: платіж ще не підтверджено</br>Кроки: </br>Користувач підтверджує оплату, платіж успішний|Зберегається інформація що платіж успішно проведено|
|TC1.2|Початкові умови: платіж ще не підтверджено</br>Кроки: </br>Користувач підтверджує оплату, виникає помилка|Зберегається інформація що платіж провести не вдалося|
|
1.0
|
CW5 - **Завдання 1**
[Посилання](https://github.com/oleksandrblazhko/ai203-tokarev/blob/main/2-SoftwareDesign/2.7-PlantUML/UML-Activity.md)
|TC id|Опис кроків тестового сценарію|Опис очікуваних результатів|
|-|-|-|
|TC1.1|Початкові умови: платіж ще не підтверджено</br>Кроки: </br>Користувач підтверджує оплату, платіж успішний|Зберегається інформація що платіж успішно проведено|
|TC1.2|Початкові умови: платіж ще не підтверджено</br>Кроки: </br>Користувач підтверджує оплату, виникає помилка|Зберегається інформація що платіж провести не вдалося|
|
non_test
|
завдання tc id опис кроків тестового сценарію опис очікуваних результатів початкові умови платіж ще не підтверджено кроки користувач підтверджує оплату платіж успішний зберегається інформація що платіж успішно проведено початкові умови платіж ще не підтверджено кроки користувач підтверджує оплату виникає помилка зберегається інформація що платіж провести не вдалося
| 0
|
108,490
| 9,309,115,791
|
IssuesEvent
|
2019-03-25 15:50:06
|
containership/cluster-manager
|
https://api.github.com/repos/containership/cluster-manager
|
closed
|
Travis should use a Docker cache for builds
|
component/test type/optimization
|
### Description
#### Why should this feature be added?
Travis needs to use a Docker cache for builds. Currently builds are painfully slow.
#### How would you like the feature to work?
I think we should just push `latest` on each successful build and then use `--cache-from`. [Here's a more thorough explanation](http://atodorov.org/blog/2017/08/07/faster-travis-ci-tests-with-docker-cache/) of this approach.
We probably want to only have it push `latest` for nighly `master` builds, though.
#### What alternatives have been considered?
None yet, open to suggestions!
#### Additional Context
N/A
|
1.0
|
Travis should use a Docker cache for builds - ### Description
#### Why should this feature be added?
Travis needs to use a Docker cache for builds. Currently builds are painfully slow.
#### How would you like the feature to work?
I think we should just push `latest` on each successful build and then use `--cache-from`. [Here's a more thorough explanation](http://atodorov.org/blog/2017/08/07/faster-travis-ci-tests-with-docker-cache/) of this approach.
We probably want to only have it push `latest` for nighly `master` builds, though.
#### What alternatives have been considered?
None yet, open to suggestions!
#### Additional Context
N/A
|
test
|
travis should use a docker cache for builds description why should this feature be added travis needs to use a docker cache for builds currently builds are painfully slow how would you like the feature to work i think we should just push latest on each successful build and then use cache from of this approach we probably want to only have it push latest for nighly master builds though what alternatives have been considered none yet open to suggestions additional context n a
| 1
|
363,458
| 10,741,404,638
|
IssuesEvent
|
2019-10-29 20:10:25
|
rubrikinc/rubrik-sdk-for-powershell
|
https://api.github.com/repos/rubrikinc/rubrik-sdk-for-powershell
|
closed
|
Force PrimaryClusterId to lowercase
|
area-rcdm exp-intermediate kind-enhancement kind-feature priority-p3
|
When trying to filter on PrimaryClusterId using local, the text must be entered in lowercase. Entering any uppercase text (IE Local) causes the API to return nothing
**Describe the solution you'd like**
Would like to see this support uppercase to fall more inline with how PowerShell handles case.
Can be done by adding code similar to the following to the Test-QueryObject private function
```
if ($location -eq 'primary_cluster_id' -and $object -in ('local','me')) {
$object = $object.toLower()
}
```
|
1.0
|
Force PrimaryClusterId to lowercase - When trying to filter on PrimaryClusterId using local, the text must be entered in lowercase. Entering any uppercase text (IE Local) causes the API to return nothing
**Describe the solution you'd like**
Would like to see this support uppercase to fall more inline with how PowerShell handles case.
Can be done by adding code similar to the following to the Test-QueryObject private function
```
if ($location -eq 'primary_cluster_id' -and $object -in ('local','me')) {
$object = $object.toLower()
}
```
|
non_test
|
force primaryclusterid to lowercase when trying to filter on primaryclusterid using local the text must be entered in lowercase entering any uppercase text ie local causes the api to return nothing describe the solution you d like would like to see this support uppercase to fall more inline with how powershell handles case can be done by adding code similar to the following to the test queryobject private function if location eq primary cluster id and object in local me object object tolower
| 0
|
404,775
| 11,863,017,367
|
IssuesEvent
|
2020-03-25 18:56:33
|
teamforus/forus
|
https://api.github.com/repos/teamforus/forus
|
closed
|
checking if IBAN is companies IBAN
|
Difficulty: Medium Priority: Could have Scope: Medium Status: Refinement Needed enhancement project-31
|
## Main asssignee: @
## Context/goal:
With this tool we can verify that an IBAN number is the IBAN of the company that signed up.
https://www.cm.com/nl-nl/producten/toegang/iban-verificatie/
|
1.0
|
checking if IBAN is companies IBAN - ## Main asssignee: @
## Context/goal:
With this tool we can verify that an IBAN number is the IBAN of the company that signed up.
https://www.cm.com/nl-nl/producten/toegang/iban-verificatie/
|
non_test
|
checking if iban is companies iban main asssignee context goal with this tool we can verify that an iban number is the iban of the company that signed up
| 0
|
141,137
| 11,395,376,581
|
IssuesEvent
|
2020-01-30 11:17:06
|
microsoft/ptvsd
|
https://api.github.com/repos/microsoft/ptvsd
|
closed
|
attach_by_socket tests fail sporadically on MacOS + Python 3.6
|
Test-issue
|
When it happens, it's usually several dozen tests failing in the same run for a given Python version. The only common theme is the attach method - the tests themselves are completely unrelated.
|
1.0
|
attach_by_socket tests fail sporadically on MacOS + Python 3.6 - When it happens, it's usually several dozen tests failing in the same run for a given Python version. The only common theme is the attach method - the tests themselves are completely unrelated.
|
test
|
attach by socket tests fail sporadically on macos python when it happens it s usually several dozen tests failing in the same run for a given python version the only common theme is the attach method the tests themselves are completely unrelated
| 1
|
156,369
| 12,307,684,666
|
IssuesEvent
|
2020-05-12 05:23:02
|
kubernetes/test-infra
|
https://api.github.com/repos/kubernetes/test-infra
|
closed
|
Using kubetest : unable to make the pods run
|
area/kubetest kind/bug triage/needs-information
|
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**When I use kubetest to take e2e test . It occurs something wrong!**
```bash
Apr 24 17:45:31.061: INFO: Density Pods: 219 out of 1320 created, 0 running, 219 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:45:41.061: INFO: Density Pods: 418 out of 1320 created, 0 running, 418 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:45:51.061: INFO: Density Pods: 609 out of 1320 created, 0 running, 609 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:01.061: INFO: Density Pods: 807 out of 1320 created, 0 running, 807 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:11.061: INFO: Density Pods: 1003 out of 1320 created, 0 running, 1003 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:21.011: INFO: Failed to gather kube-scheduler CPU profile: Profile gathering pre-requisite failed: Can't gather profiles as --report-dir is empty
#############################################################
Apr 24 18:04:31.544: INFO: Unexpected error occurred: Only 0 pods started out of 1320
Apr 24 18:04:31.545: INFO: E2E startup time for 1320 pods: 19m10.534149751s
```
I want to test the start up time of the pods in the local cluster. But as I see, pods are not running successfully! How can I solve the problem?
|
1.0
|
Using kubetest : unable to make the pods run - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**When I use kubetest to take e2e test . It occurs something wrong!**
```bash
Apr 24 17:45:31.061: INFO: Density Pods: 219 out of 1320 created, 0 running, 219 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:45:41.061: INFO: Density Pods: 418 out of 1320 created, 0 running, 418 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:45:51.061: INFO: Density Pods: 609 out of 1320 created, 0 running, 609 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:01.061: INFO: Density Pods: 807 out of 1320 created, 0 running, 807 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:11.061: INFO: Density Pods: 1003 out of 1320 created, 0 running, 1003 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady
Apr 24 17:46:21.011: INFO: Failed to gather kube-scheduler CPU profile: Profile gathering pre-requisite failed: Can't gather profiles as --report-dir is empty
#############################################################
Apr 24 18:04:31.544: INFO: Unexpected error occurred: Only 0 pods started out of 1320
Apr 24 18:04:31.545: INFO: E2E startup time for 1320 pods: 19m10.534149751s
```
I want to test the start up time of the pods in the local cluster. But as I see, pods are not running successfully! How can I solve the problem?
|
test
|
using kubetest unable to make the pods run when i use kubetest to take test it occurs something wrong bash apr info density pods out of created running pending waiting inactive terminating unknown runningbutnotready apr info density pods out of created running pending waiting inactive terminating unknown runningbutnotready apr info density pods out of created running pending waiting inactive terminating unknown runningbutnotready apr info density pods out of created running pending waiting inactive terminating unknown runningbutnotready apr info density pods out of created running pending waiting inactive terminating unknown runningbutnotready apr info failed to gather kube scheduler cpu profile profile gathering pre requisite failed can t gather profiles as report dir is empty apr info unexpected error occurred only pods started out of apr info startup time for pods i want to test the start up time of the pods in the local cluster but as i see pods are not running successfully how can i solve the problem
| 1
|
223,824
| 17,633,505,102
|
IssuesEvent
|
2021-08-19 10:58:33
|
icyphy/lingua-franca
|
https://api.github.com/repos/icyphy/lingua-franca
|
opened
|
Errors reported via ErrorReporter during validation do not cause test failure
|
bug compiler testing
|
There is a strange bug where the C++ test [WidthGivenByCode.lf](https://github.com/icyphy/lingua-franca/blob/master/test/Cpp/src/multiport/WidthGivenByCode.lf) passes in our test runs, but does not compile when invoking lfc manually. lfc reports the following error:
```
lfc: error: Cannot infer width.
--> multiport/WidthGivenByCode.lf:23:5
|
22 | foo3 = new Foo(a=9, b=9);
23 | foo_bank = new[{=42=}] Foo();
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Cannot infer width.
|
24 |
lfc: fatal error: Aborting due to previous error
```
The error actually should not appear in the first place, but this is a separate issue (see #465). This issue is concerned with the fact that there is an error reported, that does not cause a test failure. As far as I was able to investigate this, the reason is that the validator has its own injected instance of DefaultErrorReporter, which is used to report errors during validation (in particular in checkModel()). However, this is a different reporter instance than the one used by the code generator. Thus, the code generator does not notice that there is an error and invokes the target compiler. Since in this particular example, the generated C++ code is valid despite the error reported from the model, the code is compiled and the binary correctly executed, leading to the test being passed despite the error reported during validation.
I am not sure what we can do about this. What I don't really understand is why lfc fails, but the tests do not. Apparently lfc has a different way of collecting errors from the validator. This only has become an issue after merging #456, as before also lfc ignored those errors. @oowekyala do you know what is happening here?
|
1.0
|
Errors reported via ErrorReporter during validation do not cause test failure - There is a strange bug where the C++ test [WidthGivenByCode.lf](https://github.com/icyphy/lingua-franca/blob/master/test/Cpp/src/multiport/WidthGivenByCode.lf) passes in our test runs, but does not compile when invoking lfc manually. lfc reports the following error:
```
lfc: error: Cannot infer width.
--> multiport/WidthGivenByCode.lf:23:5
|
22 | foo3 = new Foo(a=9, b=9);
23 | foo_bank = new[{=42=}] Foo();
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Cannot infer width.
|
24 |
lfc: fatal error: Aborting due to previous error
```
The error actually should not appear in the first place, but this is a separate issue (see #465). This issue is concerned with the fact that there is an error reported, that does not cause a test failure. As far as I was able to investigate this, the reason is that the validator has its own injected instance of DefaultErrorReporter, which is used to report errors during validation (in particular in checkModel()). However, this is a different reporter instance than the one used by the code generator. Thus, the code generator does not notice that there is an error and invokes the target compiler. Since in this particular example, the generated C++ code is valid despite the error reported from the model, the code is compiled and the binary correctly executed, leading to the test being passed despite the error reported during validation.
I am not sure what we can do about this. What I don't really understand is why lfc fails, but the tests do not. Apparently lfc has a different way of collecting errors from the validator. This only has become an issue after merging #456, as before also lfc ignored those errors. @oowekyala do you know what is happening here?
|
test
|
errors reported via errorreporter during validation do not cause test failure there is a strange bug where the c test passes in our test runs but does not compile when invoking lfc manually lfc reports the following error lfc error cannot infer width multiport widthgivenbycode lf new foo a b foo bank new foo cannot infer width lfc fatal error aborting due to previous error the error actually should not appear in the first place but this is a separate issue see this issue is concerned with the fact that there is an error reported that does not cause a test failure as far as i was able to investigate this the reason is that the validator has its own injected instance of defaulterrorreporter which is used to report errors during validation in particular in checkmodel however this is a different reporter instance than the one used by the code generator thus the code generator does not notice that there is an error and invokes the target compiler since in this particular example the generated c code is valid despite the error reported from the model the code is compiled and the binary correctly executed leading to the test being passed despite the error reported during validation i am not sure what we can do about this what i don t really understand is why lfc fails but the tests do not apparently lfc has a different way of collecting errors from the validator this only has become an issue after merging as before also lfc ignored those errors oowekyala do you know what is happening here
| 1
|
15,823
| 5,188,133,655
|
IssuesEvent
|
2017-01-20 18:59:38
|
phetsims/unit-rates
|
https://api.github.com/repos/phetsims/unit-rates
|
closed
|
items on shelf and scale don't have a pointer cursor
|
dev:code-review type:bug type:wontfix
|
Code review #52.
Three's no indication that the items on the shelf or scale are draggable, because they don't show the pointer cursor when you mouse over them. It's not until you actually start dragging that the cursor changes.
@arouinfar Is this a bug or intended behavior?
|
1.0
|
items on shelf and scale don't have a pointer cursor - Code review #52.
Three's no indication that the items on the shelf or scale are draggable, because they don't show the pointer cursor when you mouse over them. It's not until you actually start dragging that the cursor changes.
@arouinfar Is this a bug or intended behavior?
|
non_test
|
items on shelf and scale don t have a pointer cursor code review three s no indication that the items on the shelf or scale are draggable because they don t show the pointer cursor when you mouse over them it s not until you actually start dragging that the cursor changes arouinfar is this a bug or intended behavior
| 0
|
268,614
| 23,383,998,002
|
IssuesEvent
|
2022-08-11 12:14:37
|
MTES-MCT/histologe
|
https://api.github.com/repos/MTES-MCT/histologe
|
closed
|
[FO - Création du signalement] - info propriétaire
|
A tester Contenu Manquant
|
Sur version de test 447
Lors de la création du signalement, si déclarant occupant, le champ nom du propriétaire doit être obligatoire mais s'il n'est pas renseigné, il n'y a pas de message d'information liè à l'erreur (sous Chrome) :

Proposition : Ajouter la mention "Merci de renseigner le nom ou la raison sociale du propriétaire du logement."
|
1.0
|
[FO - Création du signalement] - info propriétaire - Sur version de test 447
Lors de la création du signalement, si déclarant occupant, le champ nom du propriétaire doit être obligatoire mais s'il n'est pas renseigné, il n'y a pas de message d'information liè à l'erreur (sous Chrome) :

Proposition : Ajouter la mention "Merci de renseigner le nom ou la raison sociale du propriétaire du logement."
|
test
|
info propriétaire sur version de test lors de la création du signalement si déclarant occupant le champ nom du propriétaire doit être obligatoire mais s il n est pas renseigné il n y a pas de message d information liè à l erreur sous chrome proposition ajouter la mention merci de renseigner le nom ou la raison sociale du propriétaire du logement
| 1
|
224,992
| 17,788,456,442
|
IssuesEvent
|
2021-08-31 13:47:01
|
rpgp/rpgp
|
https://api.github.com/repos/rpgp/rpgp
|
closed
|
Bring back cross based testing
|
help wanted testing
|
Need to figure out how to enable `cross` with docker on circle ci and reenable all the targets
|
1.0
|
Bring back cross based testing - Need to figure out how to enable `cross` with docker on circle ci and reenable all the targets
|
test
|
bring back cross based testing need to figure out how to enable cross with docker on circle ci and reenable all the targets
| 1
|
85,178
| 24,532,245,847
|
IssuesEvent
|
2022-10-11 17:27:25
|
angular/angular
|
https://api.github.com/repos/angular/angular
|
closed
|
Guide/Universal: Describe how to pre-generate HTML files
|
feature comp: docs effort2: days freq3: high comp: server state: needs eng input subtype: docs-completeness docsarea: build & deploy P4 under consideration doc-topic: understanding
|
<!--
PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION.
ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION.
-->
## I'm submitting a...
<pre>
[x] Documentation issue or request
</pre>
## Current behavior
<!-- Describe how the issue manifests. -->
The [current Angular Universal guide page](https://github.com/angular/angular/blob/505ae752b61a44ca41396838a49f83cfe7badb3b/aio/content/guide/universal.md) (located [here on the web](https://angular.io/guide/universal)) says of Angular Universal:
> It can also pre-generate pages as HTML files that you serve later.
However, no direction is given to accomplish that.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
The guide page should either:
* Describe how to accomplish pre-generated HTML files.
* Or not mention the possibility of pre-generating HTML files.
Clearly, if pre-generating HTML files is a possibility, then the preferred solution would be to explain how to do that.
## What is the motivation / use case for changing the behavior?
<!-- Describe the motivation or the concrete use case. -->
Statically-hosting an Angular Universal application would provide all the [benefits of Angular Universal](https://angular.io/guide/universal#why-universal) and also give greater hosting flexibility.
|
1.0
|
Guide/Universal: Describe how to pre-generate HTML files - <!--
PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION.
ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION.
-->
## I'm submitting a...
<pre>
[x] Documentation issue or request
</pre>
## Current behavior
<!-- Describe how the issue manifests. -->
The [current Angular Universal guide page](https://github.com/angular/angular/blob/505ae752b61a44ca41396838a49f83cfe7badb3b/aio/content/guide/universal.md) (located [here on the web](https://angular.io/guide/universal)) says of Angular Universal:
> It can also pre-generate pages as HTML files that you serve later.
However, no direction is given to accomplish that.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
The guide page should either:
* Describe how to accomplish pre-generated HTML files.
* Or not mention the possibility of pre-generating HTML files.
Clearly, if pre-generating HTML files is a possibility, then the preferred solution would be to explain how to do that.
## What is the motivation / use case for changing the behavior?
<!-- Describe the motivation or the concrete use case. -->
Statically-hosting an Angular Universal application would provide all the [benefits of Angular Universal](https://angular.io/guide/universal#why-universal) and also give greater hosting flexibility.
|
non_test
|
guide universal describe how to pre generate html files please help us process github issues faster by providing the following information issues missing important information may be closed without investigation i m submitting a documentation issue or request current behavior the located says of angular universal it can also pre generate pages as html files that you serve later however no direction is given to accomplish that expected behavior the guide page should either describe how to accomplish pre generated html files or not mention the possibility of pre generating html files clearly if pre generating html files is a possibility then the preferred solution would be to explain how to do that what is the motivation use case for changing the behavior statically hosting an angular universal application would provide all the and also give greater hosting flexibility
| 0
|
21,258
| 3,876,437,116
|
IssuesEvent
|
2016-04-12 07:55:13
|
RestComm/restcomm-web-sdk
|
https://api.github.com/repos/RestComm/restcomm-web-sdk
|
closed
|
Improve logging for easier troubleshooting during load tests
|
in progress load testing framework
|
For now I have added logging in webrtc-client.html that we use for load test, but we need to enhance RestCommWebClient.js so that all WebRTComm events that are forwarded to RCDevice/RCConnection objects are logged together with their paramaters
|
1.0
|
Improve logging for easier troubleshooting during load tests - For now I have added logging in webrtc-client.html that we use for load test, but we need to enhance RestCommWebClient.js so that all WebRTComm events that are forwarded to RCDevice/RCConnection objects are logged together with their paramaters
|
test
|
improve logging for easier troubleshooting during load tests for now i have added logging in webrtc client html that we use for load test but we need to enhance restcommwebclient js so that all webrtcomm events that are forwarded to rcdevice rcconnection objects are logged together with their paramaters
| 1
|
695,389
| 23,855,035,026
|
IssuesEvent
|
2022-09-06 22:08:56
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
usa.embassy.gov.au - Link arrows are displayed misaligned
|
browser-firefox priority-normal severity-important engine-gecko diagnosis-priority-p3
|
<!-- @browser: Firefox 64bit 103.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/104.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108904 -->
**URL**: https://usa.embassy.gov.au/passport-locations
**Browser / Version**: Firefox 64bit 103.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
the design to make an appointment online is overlapped
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/7d0182c7-5af2-447b-8856-cc70bee2a9c4.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
2.0
|
usa.embassy.gov.au - Link arrows are displayed misaligned - <!-- @browser: Firefox 64bit 103.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/104.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108904 -->
**URL**: https://usa.embassy.gov.au/passport-locations
**Browser / Version**: Firefox 64bit 103.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
the design to make an appointment online is overlapped
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/7d0182c7-5af2-447b-8856-cc70bee2a9c4.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
usa embassy gov au link arrows are displayed misaligned url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description items are overlapped steps to reproduce the design to make an appointment online is overlapped view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
312,827
| 26,880,054,653
|
IssuesEvent
|
2023-02-05 14:24:55
|
tomjn/vagrant-tart-ruby
|
https://api.github.com/repos/tomjn/vagrant-tart-ruby
|
opened
|
VM stop script
|
Task Needs Testing
|
Tart has a direct equivalent to `vagrant halt`:
```sh
tart stop VMNAME
```
There's already files and code for this:
- shellscript: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/scripts/stop_vm.sh
- action: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/action/stop_instance.rb
- action builder: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/action.rb#L65-L70
- driver: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/driver.rb#L65-L67
|
1.0
|
VM stop script - Tart has a direct equivalent to `vagrant halt`:
```sh
tart stop VMNAME
```
There's already files and code for this:
- shellscript: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/scripts/stop_vm.sh
- action: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/action/stop_instance.rb
- action builder: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/action.rb#L65-L70
- driver: https://github.com/tomjn/vagrant-tart-ruby/blob/main/lib/vagrant/tart/driver.rb#L65-L67
|
test
|
vm stop script tart has a direct equivalent to vagrant halt sh tart stop vmname there s already files and code for this shellscript action action builder driver
| 1
|
82,831
| 7,853,795,918
|
IssuesEvent
|
2018-06-20 18:35:25
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Make splunk HEC index configurable
|
area/tools area/ui kind/enhancement priority/-1 status/resolved status/to-test version/2.0
|
This is currently not configurable and defaults to `main`: https://github.com/brycied00d/fluent-plugin-splunk-http-eventcollector/blob/master/lib/fluent/plugin/out_splunk-http-eventcollector.rb#L49
|
1.0
|
Make splunk HEC index configurable - This is currently not configurable and defaults to `main`: https://github.com/brycied00d/fluent-plugin-splunk-http-eventcollector/blob/master/lib/fluent/plugin/out_splunk-http-eventcollector.rb#L49
|
test
|
make splunk hec index configurable this is currently not configurable and defaults to main
| 1
|
179,534
| 13,885,837,111
|
IssuesEvent
|
2020-10-18 21:46:57
|
ericberglund117/Rancid-Tomatillos
|
https://api.github.com/repos/ericberglund117/Rancid-Tomatillos
|
closed
|
Add Testing for Login and Movies
|
testing
|
- [x] Tested the display
- [x] Tested the button click
- [x] Test the API
- [x] Test the clear inputs
- [x] Test the submit
|
1.0
|
Add Testing for Login and Movies - - [x] Tested the display
- [x] Tested the button click
- [x] Test the API
- [x] Test the clear inputs
- [x] Test the submit
|
test
|
add testing for login and movies tested the display tested the button click test the api test the clear inputs test the submit
| 1
|
25,524
| 4,160,740,526
|
IssuesEvent
|
2016-06-17 14:20:33
|
sasstools/sass-lint
|
https://api.github.com/repos/sasstools/sass-lint
|
closed
|
Sass Syntax - Tests progress
|
tests
|
Having had a few Sass Syntax related issues popping up, I think it would be a good move to create tests for that syntax too.
I've started going through the current tests and have created quite a few .sass equivalents. So far there has been a few tests that have failed but the majority are ok.
By testing both the SCSS and Sass syntax I feel we will then be able to confidently say that we have cross syntax support.
I'll update the list below as I go.
### The rules
- [x] border-zero
- [x] clean-import-paths - (Issue #179, Fixed #205)
- [x] empty-args
- [x] empty-line-between-blocks - (:bug: Issue #216)
- [x] extends-before-declarations
- [x] extends-before-mixins - (Fixed issue #203)
- [x] final-newline - (Issue #207, Fixed #208)
- [x] hex-length
- [x] hex-notation
- [x] indentation
- [x] leading-zero
- [x] mixing-before-declarations
- [x] nesting-depth
- [x] no-color-keywords
- [x] no-color-literals (Issue #186, Fixed in #187)
- [x] no-css-comments
- [x] no-debug
- [x] no-duplicate-properties
- [x] no-extends - (Fixed issue - #173)
- [x] no-ids
- [x] no-important
- [x] no-invalid-hex
- [x] no-qualifying-elements
- [x] no-trailing-zeros
- [x] no-url-protocols
- [x] no-vendor-prefix (Issue #182, Fixed in #183)
- [x] no-warn
- [x] placeholder-in-extend - (Fixed in #204)
- [x] property-sort-order
- [x] quotes
- [x] shorthand values
- [x] single-line-per-selector - (Fixed issue - #165)
- [x] space-after-bang
- [x] space-after-colon
- [x] space-after-comma
- [x] space-before bang
- [x] space-before-colon
- [x] space-before-parens
- [x] url-quotes
- [x] variable-for-property
- [x] zero-unit
### Rules that are N/A
- trailing-semicolon
- space-before-brace
- no-empty-rulesets
- brace-style
- one-declaration-per-line
|
1.0
|
Sass Syntax - Tests progress - Having had a few Sass Syntax related issues popping up, I think it would be a good move to create tests for that syntax too.
I've started going through the current tests and have created quite a few .sass equivalents. So far there has been a few tests that have failed but the majority are ok.
By testing both the SCSS and Sass syntax I feel we will then be able to confidently say that we have cross syntax support.
I'll update the list below as I go.
### The rules
- [x] border-zero
- [x] clean-import-paths - (Issue #179, Fixed #205)
- [x] empty-args
- [x] empty-line-between-blocks - (:bug: Issue #216)
- [x] extends-before-declarations
- [x] extends-before-mixins - (Fixed issue #203)
- [x] final-newline - (Issue #207, Fixed #208)
- [x] hex-length
- [x] hex-notation
- [x] indentation
- [x] leading-zero
- [x] mixing-before-declarations
- [x] nesting-depth
- [x] no-color-keywords
- [x] no-color-literals (Issue #186, Fixed in #187)
- [x] no-css-comments
- [x] no-debug
- [x] no-duplicate-properties
- [x] no-extends - (Fixed issue - #173)
- [x] no-ids
- [x] no-important
- [x] no-invalid-hex
- [x] no-qualifying-elements
- [x] no-trailing-zeros
- [x] no-url-protocols
- [x] no-vendor-prefix (Issue #182, Fixed in #183)
- [x] no-warn
- [x] placeholder-in-extend - (Fixed in #204)
- [x] property-sort-order
- [x] quotes
- [x] shorthand values
- [x] single-line-per-selector - (Fixed issue - #165)
- [x] space-after-bang
- [x] space-after-colon
- [x] space-after-comma
- [x] space-before bang
- [x] space-before-colon
- [x] space-before-parens
- [x] url-quotes
- [x] variable-for-property
- [x] zero-unit
### Rules that are N/A
- trailing-semicolon
- space-before-brace
- no-empty-rulesets
- brace-style
- one-declaration-per-line
|
test
|
sass syntax tests progress having had a few sass syntax related issues popping up i think it would be a good move to create tests for that syntax too i ve started going through the current tests and have created quite a few sass equivalents so far there has been a few tests that have failed but the majority are ok by testing both the scss and sass syntax i feel we will then be able to confidently say that we have cross syntax support i ll update the list below as i go the rules border zero clean import paths issue fixed empty args empty line between blocks bug issue extends before declarations extends before mixins fixed issue final newline issue fixed hex length hex notation indentation leading zero mixing before declarations nesting depth no color keywords no color literals issue fixed in no css comments no debug no duplicate properties no extends fixed issue no ids no important no invalid hex no qualifying elements no trailing zeros no url protocols no vendor prefix issue fixed in no warn placeholder in extend fixed in property sort order quotes shorthand values single line per selector fixed issue space after bang space after colon space after comma space before bang space before colon space before parens url quotes variable for property zero unit rules that are n a trailing semicolon space before brace no empty rulesets brace style one declaration per line
| 1
|
240,772
| 20,073,516,542
|
IssuesEvent
|
2022-02-04 10:04:42
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: jepsen/g2/split failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-20.2
|
[(roachtest).jepsen/g2/split failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=buildLog) on [release-20.2@09707aabb12e50f6e7345b5c9664c0745bb7d742](https://github.com/cockroachdb/cockroach/commits/09707aabb12e50f6e7345b5c9664c0745bb7d742):
```
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2291
| main.runJepsen.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:160
| main.runJepsen.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:195
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1581
Wraps: (2) output in run_100400.477_n6_bash
Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-4299094-1643958549-61-n6cpu4:6 -- bash -e -c "\
| cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
| ~/lein run test \
| --tarball file://${PWD}/cockroach.tgz \
| --username ${USER} \
| --ssh-private-key ~/.ssh/id_rsa \
| --os ubuntu \
| --time-limit 300 \
| --concurrency 30 \
| --recovery-time 25 \
| --test-count 1 \
| -n 10.128.0.11 -n 10.128.15.193 -n 10.128.0.189 -n 10.128.0.184 -n 10.128.0.9 \
| --test g2 --nemesis split \
| > invoke.log 2>&1 \
| " returned
| stderr:
| Error: SSH_PROBLEM: exit status 255
| (1) SSH_PROBLEM
| Wraps: (2) Node 6. Command with error:
| | ```
| | bash -e -c "\
| | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
| | ~/lein run test \
| | --tarball file://${PWD}/cockroach.tgz \
| | --username ${USER} \
| | --ssh-private-key ~/.ssh/id_rsa \
| | --os ubuntu \
| | --time-limit 300 \
| | --concurrency 30 \
| | --recovery-time 25 \
| | --test-count 1 \
| | -n 10.128.0.11 -n 10.128.15.193 -n 10.128.0.189 -n 10.128.0.184 -n 10.128.0.9 \
| | --test g2 --nemesis split \
| | > invoke.log 2>&1 \
| | "
| | ```
| Wraps: (3) exit status 255
| Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (4) exit status 10
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/jepsen/g2/split](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=artifacts#/jepsen/g2/split)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fg2%2Fsplit.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: jepsen/g2/split failed - [(roachtest).jepsen/g2/split failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=buildLog) on [release-20.2@09707aabb12e50f6e7345b5c9664c0745bb7d742](https://github.com/cockroachdb/cockroach/commits/09707aabb12e50f6e7345b5c9664c0745bb7d742):
```
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2291
| main.runJepsen.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:160
| main.runJepsen.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:195
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1581
Wraps: (2) output in run_100400.477_n6_bash
Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-4299094-1643958549-61-n6cpu4:6 -- bash -e -c "\
| cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
| ~/lein run test \
| --tarball file://${PWD}/cockroach.tgz \
| --username ${USER} \
| --ssh-private-key ~/.ssh/id_rsa \
| --os ubuntu \
| --time-limit 300 \
| --concurrency 30 \
| --recovery-time 25 \
| --test-count 1 \
| -n 10.128.0.11 -n 10.128.15.193 -n 10.128.0.189 -n 10.128.0.184 -n 10.128.0.9 \
| --test g2 --nemesis split \
| > invoke.log 2>&1 \
| " returned
| stderr:
| Error: SSH_PROBLEM: exit status 255
| (1) SSH_PROBLEM
| Wraps: (2) Node 6. Command with error:
| | ```
| | bash -e -c "\
| | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
| | ~/lein run test \
| | --tarball file://${PWD}/cockroach.tgz \
| | --username ${USER} \
| | --ssh-private-key ~/.ssh/id_rsa \
| | --os ubuntu \
| | --time-limit 300 \
| | --concurrency 30 \
| | --recovery-time 25 \
| | --test-count 1 \
| | -n 10.128.0.11 -n 10.128.15.193 -n 10.128.0.189 -n 10.128.0.184 -n 10.128.0.9 \
| | --test g2 --nemesis split \
| | > invoke.log 2>&1 \
| | "
| | ```
| Wraps: (3) exit status 255
| Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (4) exit status 10
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/jepsen/g2/split](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=artifacts#/jepsen/g2/split)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fg2%2Fsplit.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest jepsen split failed on home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go runtime goexit usr local go src runtime asm s wraps output in run bash wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test nemesis split invoke log returned stderr error ssh problem exit status ssh problem wraps node command with error bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test nemesis split invoke log wraps exit status error types errors ssh hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack errutil withprefix main withcommanddetails exec exiterror more artifacts powered by
| 1
|
224,585
| 17,193,916,682
|
IssuesEvent
|
2021-07-16 14:42:04
|
pyrocms/pyrocms
|
https://api.github.com/repos/pyrocms/pyrocms
|
closed
|
[videos] New video tutorials!
|
Type: Documentation
|
To start off, I had considerable difficulties creating a custom ajax contact form with captcha (yes, i guess this is beginners :) ).
Please add your suggestions for new videos here!
|
1.0
|
[videos] New video tutorials! - To start off, I had considerable difficulties creating a custom ajax contact form with captcha (yes, i guess this is beginners :) ).
Please add your suggestions for new videos here!
|
non_test
|
new video tutorials to start off i had considerable difficulties creating a custom ajax contact form with captcha yes i guess this is beginners please add your suggestions for new videos here
| 0
|
121,497
| 10,170,396,029
|
IssuesEvent
|
2019-08-08 05:07:59
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed test: TestMergeQueue
|
C-test-failure O-robot
|
The following tests appear to have failed on master (testrace): TestMergeQueue/sanity, TestMergeQueue/combined-threshold, TestMergeQueue, TestMergeQueue/sticky-bit, TestMergeQueue/lhs-undersize, TestMergeQueue/both-empty, TestMergeQueue/non-collocated
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestMergeQueue).
[#1428523](https://teamcity.cockroachdb.com/viewLog.html?buildId=1428523):
```
TestMergeQueue/non-collocated
...orage.go:829 [s2,r10/?:{b-c}] applied PREEMPTIVE snapshot in 320ms [clear=0ms batch=229ms entries=0ms commit=90ms]
I190807 19:00:07.904950 157073 storage/replica_command.go:1188 [s1,r10/1:{b-c}] change replicas (ADD_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n1,s1):1, next=2, gen=15]
I190807 19:00:07.915726 157073 storage/replica_raft.go:290 [s1,r10/1:{b-c},txn=93480a64] proposing ADD_REPLICA((n2,s2):2): updated=(n1,s1):1,(n2,s2):2 next=3
I190807 19:00:08.332247 157073 storage/replica_command.go:1188 [s2,r10/2:{b-c}] change replicas (REMOVE_REPLICA (n1,s1):1): existing descriptor r10:{b-c} [(n1,s1):1, (n2,s2):2, next=3, gen=16]
I190807 19:00:08.360906 157073 storage/replica_raft.go:290 [s2,r10/2:{b-c},txn=f9bde25a] proposing REMOVE_REPLICA((n1,s1):1): updated=(n2,s2):2 next=3
I190807 19:00:08.414632 170862 storage/store.go:2530 [replicaGC,s1,r10/1:{b-c}] removing replica r10/1
I190807 19:00:08.416044 170862 storage/replica_destroy.go:146 [replicaGC,s1,r10/1:{b-c}] removed 6 (1+5) keys in 1ms [clear=0ms commit=0ms]
W190807 19:00:08.416810 170888 storage/replica_command.go:1725 [merge,s2,r10/2:{b-c}] while adding target n1,s1: snapshot failed: (n1,s1):?: remote couldn't accept PREEMPTIVE snapshot a0c5420f at applied index 21 with error: [n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})
E190807 19:00:08.425903 170713 storage/queue.go:1027 [replicate,s2,r10/2:{b-c}] 1 matching stores are currently throttled: [[n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})]
I190807 19:00:08.426036 170713 storage/queue.go:1127 [replicate] purgatory is now empty
I190807 19:00:08.426725 170888 storage/store_snapshot.go:775 [merge,s2,r10/2:{b-c}] sending PREEMPTIVE snapshot 5a6de673 at applied index 21
I190807 19:00:08.427752 170888 storage/store_snapshot.go:818 [merge,s2,r10/2:{b-c}] streamed snapshot to (n1,s1):?: kv pairs: 21, log entries: 0, rate-limit: 8.0 MiB/sec, 0.00s
I190807 19:00:08.430891 170882 storage/replica_raftstorage.go:823 [s1,r10/?:{-}] applying PREEMPTIVE snapshot at index 21 (id=5a6de673, encoded size=1102, 1 rocksdb batches, 0 log entries)
I190807 19:00:08.432871 170882 storage/replica_raftstorage.go:829 [s1,r10/?:{b-c}] applied PREEMPTIVE snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I190807 19:00:08.442618 170888 storage/replica_command.go:1188 [merge,s2,r10/2:{b-c}] change replicas (ADD_REPLICA (n1,s1):3): existing descriptor r10:{b-c} [(n2,s2):2, next=3, gen=17]
I190807 19:00:08.460010 170888 storage/replica_raft.go:290 [merge,s2,r10/2:{b-c},txn=4cb784e3] proposing ADD_REPLICA((n1,s1):3): updated=(n2,s2):2,(n1,s1):3 next=4
I190807 19:00:08.495173 170888 storage/replica_command.go:1188 [merge,s1,r10/3:{b-c}] change replicas (REMOVE_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n2,s2):2, (n1,s1):3, next=4, gen=18]
I190807 19:00:08.510995 170888 storage/replica_raft.go:290 [merge,s1,r10/3:{b-c},txn=647f6935] proposing REMOVE_REPLICA((n2,s2):2): updated=(n1,s1):3 next=4
I190807 19:00:08.526619 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.538611 171003 storage/store.go:2530 [replicaGC,s2,r10/2:{b-c}] removing replica r10/2
I190807 19:00:08.540067 171003 storage/replica_destroy.go:146 [replicaGC,s2,r10/2:{b-c}] removed 7 (0+7) keys in 1ms [clear=0ms commit=0ms]
I190807 19:00:08.689040 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.877603 85667 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=bac592ef] removing replica r10/3
TestMergeQueue
--- FAIL: testrace/TestMergeQueue (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:57:01.616659 85583 gossip/gossip.go:394 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39043" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
W190807 18:57:01.710337 85583 gossip/gossip.go:1498 [n2] no incoming or outgoing connections
I190807 18:57:01.711139 85583 gossip/gossip.go:394 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:33469" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
I190807 18:57:01.730521 85876 gossip/client.go:124 [n2] started gossip client to 127.0.0.1:39043
I190807 18:57:01.795336 85583 storage/client_test.go:495 gossip network initialized
I190807 18:57:01.803889 85583 storage/replica_command.go:283 [s1,r1/1:/M{in-ax}] initiating a split of this range at key "a" [r2] (manual)
I190807 18:57:01.850586 85583 storage/replica_command.go:283 [s1,r2/1:{a-/Max}] initiating a split of this range at key "b" [r3] (manual)
I190807 18:57:01.899827 85583 storage/replica_command.go:283 [s1,r3/1:{b-/Max}] initiating a split of this range at key "c" [r4] (manual)
I190807 18:57:03.850154 85892 gossip/gossip.go:1512 [n2] node has connected to cluster via gossip
TestMergeQueue/lhs-undersize
--- FAIL: testrace/TestMergeQueue/lhs-undersize (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:08.782829 85658 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:59:08.807560 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.808047 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:59:08.843036 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.847447 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:59:09.705064 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:11.884279 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:11.886797 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:59:11.887846 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:59:23.526013 85698 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.0s [applied=1, batches=1, state_assertions=0]
I190807 18:59:23.529053 134859 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r8] (manual)
I190807 18:59:25.341106 134859 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r8:{b-c} [(n1,s1):1, next=2, gen=11] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=4.62+0.00 --> 4.62qps) below threshold (size=32 MiB, qps=4.62))
I190807 18:59:25.537883 85653 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=f8f4916e] removing replica r8/1
TestMergeQueue/combined-threshold
--- FAIL: testrace/TestMergeQueue/combined-threshold (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:36.625896 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 9.2s [applied=1, batches=1, state_assertions=0]
I190807 18:59:40.539715 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:40.540850 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 32 MiB (reasons: size=false used=true avail=false)
I190807 18:59:40.541589 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 32 MiB in 0.0s
W190807 18:59:45.926124 85710 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 7.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:45.932898 147186 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r9] (manual)
I190807 18:59:46.913641 156969 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r9:{b-c} [(n1,s1):1, next=2, gen=13] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=0.00+0.00 --> 0.00qps) below threshold (size=32 MiB, qps=0.00))
I190807 18:59:47.128755 85651 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=3ac231da] removing replica r9/1
TestMergeQueue/sticky-bit
...heduler).Start.func2(0x5778e20, 0xc001f28660)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc000345e10, 0xc000dc55e0, 0xc000345e00)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 85697 [sync.Cond.Wait]:
runtime.goparkunlock(...)
/usr/local/go/src/runtime/proc.go:307
sync.runtime_notifyListWait(0xc0023d8290, 0xc000000a9c)
/usr/local/go/src/runtime/sema.go:510 +0xf9
sync.(*Cond).Wait(0xc0023d8280)
/usr/local/go/src/sync/cond.go:56 +0x8e
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc001140ab0, 0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:192 +0x9c
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0025a63a0, 0xc000dc55e0, 0xc0025a6390)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 171028 [select]:
github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher.(*RequestBatcher).Send(0xc0003acfc0, 0x5778da0, 0xc0000de010, 0xa, 0x57d9da0, 0xc000795200, 0x0, 0x1, 0x0, 0xc00038bcc0)
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher/batcher.go:240 +0x251
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).gcTxnRecord(0xc0003acd20, 0x5778da0, 0xc0000de010, 0xa, 0xc0016c7560, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:772 +0x571
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).cleanupFinishedTxnIntents.func2(0x5778da0, 0xc0000de010)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:812 +0x91
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1(0xc0011b6aa0, 0x5778da0, 0xc0000de010, 0xc000496880, 0x33, 0x0, 0x0, 0xc002cd4f60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:321 +0xf4
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:316 +0x14b
goroutine 85736 [select]:
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1(0x5778e20, 0xc001f29c80)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:748 +0x1ef
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0002faf00, 0xc000dc55e0, 0xc000416f40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
****************************************************************************
This node experienced a fatal error (printed above), and as a result the
process is terminating.
Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
problem in CockroachDB. With your help, the support team at Cockroach Labs
will try to determine the root cause, recommend next steps, and we can
improve CockroachDB based on your report.
Please submit a crash report by following the instructions here:
https://github.com/cockroachdb/cockroach/issues/new/choose
If you would rather not post publicly, please contact us directly at:
support@cockroachlabs.com
The Cockroach Labs team appreciates your feedback.
TestMergeQueue/sanity
...269] removing replica r5/1
I190807 18:57:56.238789 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:56.239373 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:58.884367 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:58.885544 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:59.419582 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:59.420066 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.237228 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.237968 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.416758 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.417527 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.061599 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.062161 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.064319 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.064742 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.540987 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1-2/2 ("a"-"c") for 32 MiB (reasons: size=false used=true avail=false)
W190807 18:58:11.103858 85661 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.5s [applied=1, batches=1, state_assertions=0]
W190807 18:58:11.560338 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 13.6s [applied=1, batches=1, state_assertions=0]
I190807 18:58:11.562378 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.562787 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.584071 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.584547 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.894549 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1-2/2 ("a"-"c") for 32 MiB in 2.4s
W190807 18:58:25.326862 85705 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.8s [applied=1, batches=1, state_assertions=0]
I190807 18:58:25.331543 96953 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r6] (manual)
TestMergeQueue/both-empty
--- FAIL: testrace/TestMergeQueue/both-empty (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:58:27.185018 122024 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r6:{b-c} [(n1,s1):1, next=2, gen=7] into this range (lhs+rhs has (size=0 B+0 B qps=4.82+0.00 --> 4.82qps) below threshold (size=0 B, qps=4.82))
I190807 18:58:27.790959 85705 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=7ac14de8] removing replica r6/1
W190807 18:58:38.204739 85668 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:38.213803 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.214397 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:38.266936 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.269692 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:39.805151 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:39.805623 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:58:39.807038 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:41.885436 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:58:41.888047 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:58:41.890478 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:58:54.874866 85670 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:54.877563 122008 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r7] (manual)
I190807 18:58:56.925560 134529 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r7:{b-c} [(n1,s1):1, next=2, gen=9] into this range (lhs+rhs has (size=0 B+0 B qps=3.95+0.00 --> 3.95qps) below threshold (size=0 B, qps=3.95))
I190807 18:58:57.181493 85682 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=42f7db6a] removing replica r7/1
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed test: TestMergeQueue - The following tests appear to have failed on master (testrace): TestMergeQueue/sanity, TestMergeQueue/combined-threshold, TestMergeQueue, TestMergeQueue/sticky-bit, TestMergeQueue/lhs-undersize, TestMergeQueue/both-empty, TestMergeQueue/non-collocated
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestMergeQueue).
[#1428523](https://teamcity.cockroachdb.com/viewLog.html?buildId=1428523):
```
TestMergeQueue/non-collocated
...orage.go:829 [s2,r10/?:{b-c}] applied PREEMPTIVE snapshot in 320ms [clear=0ms batch=229ms entries=0ms commit=90ms]
I190807 19:00:07.904950 157073 storage/replica_command.go:1188 [s1,r10/1:{b-c}] change replicas (ADD_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n1,s1):1, next=2, gen=15]
I190807 19:00:07.915726 157073 storage/replica_raft.go:290 [s1,r10/1:{b-c},txn=93480a64] proposing ADD_REPLICA((n2,s2):2): updated=(n1,s1):1,(n2,s2):2 next=3
I190807 19:00:08.332247 157073 storage/replica_command.go:1188 [s2,r10/2:{b-c}] change replicas (REMOVE_REPLICA (n1,s1):1): existing descriptor r10:{b-c} [(n1,s1):1, (n2,s2):2, next=3, gen=16]
I190807 19:00:08.360906 157073 storage/replica_raft.go:290 [s2,r10/2:{b-c},txn=f9bde25a] proposing REMOVE_REPLICA((n1,s1):1): updated=(n2,s2):2 next=3
I190807 19:00:08.414632 170862 storage/store.go:2530 [replicaGC,s1,r10/1:{b-c}] removing replica r10/1
I190807 19:00:08.416044 170862 storage/replica_destroy.go:146 [replicaGC,s1,r10/1:{b-c}] removed 6 (1+5) keys in 1ms [clear=0ms commit=0ms]
W190807 19:00:08.416810 170888 storage/replica_command.go:1725 [merge,s2,r10/2:{b-c}] while adding target n1,s1: snapshot failed: (n1,s1):?: remote couldn't accept PREEMPTIVE snapshot a0c5420f at applied index 21 with error: [n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})
E190807 19:00:08.425903 170713 storage/queue.go:1027 [replicate,s2,r10/2:{b-c}] 1 matching stores are currently throttled: [[n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})]
I190807 19:00:08.426036 170713 storage/queue.go:1127 [replicate] purgatory is now empty
I190807 19:00:08.426725 170888 storage/store_snapshot.go:775 [merge,s2,r10/2:{b-c}] sending PREEMPTIVE snapshot 5a6de673 at applied index 21
I190807 19:00:08.427752 170888 storage/store_snapshot.go:818 [merge,s2,r10/2:{b-c}] streamed snapshot to (n1,s1):?: kv pairs: 21, log entries: 0, rate-limit: 8.0 MiB/sec, 0.00s
I190807 19:00:08.430891 170882 storage/replica_raftstorage.go:823 [s1,r10/?:{-}] applying PREEMPTIVE snapshot at index 21 (id=5a6de673, encoded size=1102, 1 rocksdb batches, 0 log entries)
I190807 19:00:08.432871 170882 storage/replica_raftstorage.go:829 [s1,r10/?:{b-c}] applied PREEMPTIVE snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I190807 19:00:08.442618 170888 storage/replica_command.go:1188 [merge,s2,r10/2:{b-c}] change replicas (ADD_REPLICA (n1,s1):3): existing descriptor r10:{b-c} [(n2,s2):2, next=3, gen=17]
I190807 19:00:08.460010 170888 storage/replica_raft.go:290 [merge,s2,r10/2:{b-c},txn=4cb784e3] proposing ADD_REPLICA((n1,s1):3): updated=(n2,s2):2,(n1,s1):3 next=4
I190807 19:00:08.495173 170888 storage/replica_command.go:1188 [merge,s1,r10/3:{b-c}] change replicas (REMOVE_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n2,s2):2, (n1,s1):3, next=4, gen=18]
I190807 19:00:08.510995 170888 storage/replica_raft.go:290 [merge,s1,r10/3:{b-c},txn=647f6935] proposing REMOVE_REPLICA((n2,s2):2): updated=(n1,s1):3 next=4
I190807 19:00:08.526619 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.538611 171003 storage/store.go:2530 [replicaGC,s2,r10/2:{b-c}] removing replica r10/2
I190807 19:00:08.540067 171003 storage/replica_destroy.go:146 [replicaGC,s2,r10/2:{b-c}] removed 7 (0+7) keys in 1ms [clear=0ms commit=0ms]
I190807 19:00:08.689040 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.877603 85667 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=bac592ef] removing replica r10/3
TestMergeQueue
--- FAIL: testrace/TestMergeQueue (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:57:01.616659 85583 gossip/gossip.go:394 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39043" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
W190807 18:57:01.710337 85583 gossip/gossip.go:1498 [n2] no incoming or outgoing connections
I190807 18:57:01.711139 85583 gossip/gossip.go:394 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:33469" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
I190807 18:57:01.730521 85876 gossip/client.go:124 [n2] started gossip client to 127.0.0.1:39043
I190807 18:57:01.795336 85583 storage/client_test.go:495 gossip network initialized
I190807 18:57:01.803889 85583 storage/replica_command.go:283 [s1,r1/1:/M{in-ax}] initiating a split of this range at key "a" [r2] (manual)
I190807 18:57:01.850586 85583 storage/replica_command.go:283 [s1,r2/1:{a-/Max}] initiating a split of this range at key "b" [r3] (manual)
I190807 18:57:01.899827 85583 storage/replica_command.go:283 [s1,r3/1:{b-/Max}] initiating a split of this range at key "c" [r4] (manual)
I190807 18:57:03.850154 85892 gossip/gossip.go:1512 [n2] node has connected to cluster via gossip
TestMergeQueue/lhs-undersize
--- FAIL: testrace/TestMergeQueue/lhs-undersize (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:08.782829 85658 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:59:08.807560 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.808047 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:59:08.843036 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.847447 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:59:09.705064 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:11.884279 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:11.886797 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:59:11.887846 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:59:23.526013 85698 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.0s [applied=1, batches=1, state_assertions=0]
I190807 18:59:23.529053 134859 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r8] (manual)
I190807 18:59:25.341106 134859 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r8:{b-c} [(n1,s1):1, next=2, gen=11] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=4.62+0.00 --> 4.62qps) below threshold (size=32 MiB, qps=4.62))
I190807 18:59:25.537883 85653 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=f8f4916e] removing replica r8/1
TestMergeQueue/combined-threshold
--- FAIL: testrace/TestMergeQueue/combined-threshold (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:36.625896 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 9.2s [applied=1, batches=1, state_assertions=0]
I190807 18:59:40.539715 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:40.540850 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 32 MiB (reasons: size=false used=true avail=false)
I190807 18:59:40.541589 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 32 MiB in 0.0s
W190807 18:59:45.926124 85710 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 7.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:45.932898 147186 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r9] (manual)
I190807 18:59:46.913641 156969 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r9:{b-c} [(n1,s1):1, next=2, gen=13] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=0.00+0.00 --> 0.00qps) below threshold (size=32 MiB, qps=0.00))
I190807 18:59:47.128755 85651 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=3ac231da] removing replica r9/1
TestMergeQueue/sticky-bit
...heduler).Start.func2(0x5778e20, 0xc001f28660)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc000345e10, 0xc000dc55e0, 0xc000345e00)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 85697 [sync.Cond.Wait]:
runtime.goparkunlock(...)
/usr/local/go/src/runtime/proc.go:307
sync.runtime_notifyListWait(0xc0023d8290, 0xc000000a9c)
/usr/local/go/src/runtime/sema.go:510 +0xf9
sync.(*Cond).Wait(0xc0023d8280)
/usr/local/go/src/sync/cond.go:56 +0x8e
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc001140ab0, 0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:192 +0x9c
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0025a63a0, 0xc000dc55e0, 0xc0025a6390)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 171028 [select]:
github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher.(*RequestBatcher).Send(0xc0003acfc0, 0x5778da0, 0xc0000de010, 0xa, 0x57d9da0, 0xc000795200, 0x0, 0x1, 0x0, 0xc00038bcc0)
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher/batcher.go:240 +0x251
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).gcTxnRecord(0xc0003acd20, 0x5778da0, 0xc0000de010, 0xa, 0xc0016c7560, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:772 +0x571
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).cleanupFinishedTxnIntents.func2(0x5778da0, 0xc0000de010)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:812 +0x91
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1(0xc0011b6aa0, 0x5778da0, 0xc0000de010, 0xc000496880, 0x33, 0x0, 0x0, 0xc002cd4f60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:321 +0xf4
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:316 +0x14b
goroutine 85736 [select]:
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1(0x5778e20, 0xc001f29c80)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:748 +0x1ef
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0002faf00, 0xc000dc55e0, 0xc000416f40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
****************************************************************************
This node experienced a fatal error (printed above), and as a result the
process is terminating.
Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
problem in CockroachDB. With your help, the support team at Cockroach Labs
will try to determine the root cause, recommend next steps, and we can
improve CockroachDB based on your report.
Please submit a crash report by following the instructions here:
https://github.com/cockroachdb/cockroach/issues/new/choose
If you would rather not post publicly, please contact us directly at:
support@cockroachlabs.com
The Cockroach Labs team appreciates your feedback.
TestMergeQueue/sanity
...269] removing replica r5/1
I190807 18:57:56.238789 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:56.239373 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:58.884367 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:58.885544 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:59.419582 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:59.420066 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.237228 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.237968 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.416758 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.417527 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.061599 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.062161 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.064319 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.064742 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.540987 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1-2/2 ("a"-"c") for 32 MiB (reasons: size=false used=true avail=false)
W190807 18:58:11.103858 85661 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.5s [applied=1, batches=1, state_assertions=0]
W190807 18:58:11.560338 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 13.6s [applied=1, batches=1, state_assertions=0]
I190807 18:58:11.562378 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.562787 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.584071 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.584547 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.894549 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1-2/2 ("a"-"c") for 32 MiB in 2.4s
W190807 18:58:25.326862 85705 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.8s [applied=1, batches=1, state_assertions=0]
I190807 18:58:25.331543 96953 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r6] (manual)
TestMergeQueue/both-empty
--- FAIL: testrace/TestMergeQueue/both-empty (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:58:27.185018 122024 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r6:{b-c} [(n1,s1):1, next=2, gen=7] into this range (lhs+rhs has (size=0 B+0 B qps=4.82+0.00 --> 4.82qps) below threshold (size=0 B, qps=4.82))
I190807 18:58:27.790959 85705 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=7ac14de8] removing replica r6/1
W190807 18:58:38.204739 85668 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:38.213803 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.214397 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:38.266936 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.269692 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:39.805151 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:39.805623 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:58:39.807038 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:41.885436 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:58:41.888047 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:58:41.890478 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:58:54.874866 85670 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:54.877563 122008 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r7] (manual)
I190807 18:58:56.925560 134529 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r7:{b-c} [(n1,s1):1, next=2, gen=9] into this range (lhs+rhs has (size=0 B+0 B qps=3.95+0.00 --> 3.95qps) below threshold (size=0 B, qps=3.95))
I190807 18:58:57.181493 85682 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=42f7db6a] removing replica r7/1
```
Please assign, take a look and update the issue accordingly.
|
test
|
teamcity failed test testmergequeue the following tests appear to have failed on master testrace testmergequeue sanity testmergequeue combined threshold testmergequeue testmergequeue sticky bit testmergequeue lhs undersize testmergequeue both empty testmergequeue non collocated you may want to check testmergequeue non collocated orage go applied preemptive snapshot in storage replica command go change replicas add replica existing descriptor b c storage replica raft go proposing add replica updated next storage replica command go change replicas remove replica existing descriptor b c storage replica raft go proposing remove replica updated next storage store go removing replica storage replica destroy go removed keys in storage replica command go while adding target snapshot failed remote couldn t accept preemptive snapshot at applied index with error cannot apply snapshot snapshot intersects existing range initiated gc incoming b c storage queue go matching stores are currently throttled cannot apply snapshot snapshot intersects existing range initiated gc incoming b c storage queue go purgatory is now empty storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica existing descriptor b c storage replica raft go proposing add replica updated next storage replica command go change replicas remove replica existing descriptor b c storage replica raft go proposing remove replica updated next storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica storage replica destroy go removed keys in storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica testmergequeue fail testrace testmergequeue test ended in panic stdout gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at cluster name gossip gossip go no incoming or outgoing connections gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at cluster name gossip client go started gossip client to storage client test go gossip network initialized storage replica command go initiating a split of this range at key a manual storage replica command go initiating a split of this range at key b manual storage replica command go initiating a split of this range at key c manual gossip gossip go node has connected to cluster via gossip testmergequeue lhs undersize fail testrace testmergequeue lhs undersize test ended in panic stdout storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size mib mib qps below threshold size mib qps storage store go removing replica testmergequeue combined threshold fail testrace testmergequeue combined threshold test ended in panic stdout storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size mib mib qps below threshold size mib qps storage store go removing replica testmergequeue sticky bit heduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine runtime goparkunlock usr local go src runtime proc go sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg internal client requestbatcher requestbatcher send go src github com cockroachdb cockroach pkg internal client requestbatcher batcher go github com cockroachdb cockroach pkg storage intentresolver intentresolver gctxnrecord go src github com cockroachdb cockroach pkg storage intentresolver intent resolver go github com cockroachdb cockroach pkg storage intentresolver intentresolver cleanupfinishedtxnintents go src github com cockroachdb cockroach pkg storage intentresolver intent resolver go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go this node experienced a fatal error printed above and as a result the process is terminating fatal errors can occur due to faulty hardware disks memory clocks or a problem in cockroachdb with your help the support team at cockroach labs will try to determine the root cause recommend next steps and we can improve cockroachdb based on your report please submit a crash report by following the instructions here if you would rather not post publicly please contact us directly at support cockroachlabs com the cockroach labs team appreciates your feedback testmergequeue sanity removing replica storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage compactor compactor go processing compaction a c for mib reasons size false used true avail false storage store go handle raft ready storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage compactor compactor go processed compaction a c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual testmergequeue both empty fail testrace testmergequeue both empty test ended in panic stdout storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica please assign take a look and update the issue accordingly
| 1
|
129,685
| 10,582,508,353
|
IssuesEvent
|
2019-10-08 11:43:22
|
apache/incubator-shardingsphere
|
https://api.github.com/repos/apache/incubator-shardingsphere
|
closed
|
Supplement test cases for sharding-integration-test module
|
test
|
There are several test cases which are not finished in sharding-integration-test module, need to supplement them in the featue.
|
1.0
|
Supplement test cases for sharding-integration-test module - There are several test cases which are not finished in sharding-integration-test module, need to supplement them in the featue.
|
test
|
supplement test cases for sharding integration test module there are several test cases which are not finished in sharding integration test module need to supplement them in the featue
| 1
|
738,556
| 25,566,392,613
|
IssuesEvent
|
2022-11-30 14:35:37
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
zoom.us - "Virtual Background" option is missing from the background settings
|
status-needsinfo browser-firefox priority-normal severity-critical status-needsinfo-denschub os-linux os-win engine-gecko
|
<!-- @browser: Firefox 106.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:106.0) Gecko/20100101 Firefox/106.0 -->
<!-- @reported_with: unknown -->
**URL**: https://zoom.us/
**Browser / Version**: Firefox 106.0
**Operating System**: Linux
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Firefox does not offer Virtual Backgrounds (blur) in Zoom meetings
**Steps to Reproduce**:
Virtual backgrounds (in meeting) are not available for Firefox.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
zoom.us - "Virtual Background" option is missing from the background settings - <!-- @browser: Firefox 106.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:106.0) Gecko/20100101 Firefox/106.0 -->
<!-- @reported_with: unknown -->
**URL**: https://zoom.us/
**Browser / Version**: Firefox 106.0
**Operating System**: Linux
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Firefox does not offer Virtual Backgrounds (blur) in Zoom meetings
**Steps to Reproduce**:
Virtual backgrounds (in meeting) are not available for Firefox.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
zoom us virtual background option is missing from the background settings url browser version firefox operating system linux tested another browser yes chrome problem type something else description firefox does not offer virtual backgrounds blur in zoom meetings steps to reproduce virtual backgrounds in meeting are not available for firefox browser configuration none from with ❤️
| 0
|
315,528
| 9,621,640,589
|
IssuesEvent
|
2019-05-14 11:09:35
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.bestbuy.com - see bug description
|
browser-firefox-mobile engine-gecko priority-important
|
<!-- @browser: Firefox Mobile 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:65.0) Gecko/65.0 Firefox/65.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001
**Browser / Version**: Firefox Mobile 65.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: If this is a ghost site how is my location dusplayed
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/5/f3fe9ee7-cb8b-4cc9-a7d8-ec90b84ca11e.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506082951</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: default</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[console.info(EventManager: Debug mode enabled via window.debugMode) https://assets.bbystatic.com/header-footer/store-locator.7bbb7e7eb239c54ecd82.0.js:1:159515]', u'[JavaScript Warning: "The resource at https://bestbuy.demdex.net/dest5.html?d_nsid=0#https%3A%2F%2Fwww.bestbuy.com%2Fsite%2Fdishwashers%2Fbuilt-in-dishwashers%2Fabcat0905001.c%3Fid%3Dabcat0905001 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "The resource at https://nexus.ensighten.com/bestbuy/tagframe_dev/Bootstrap.js was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://nexus.ensighten.com/bestbuy/tagframe_dev/Bootstrap.js." {file: "https://img.bbystatic.com/BestBuy_US/js/tracking/ens-index.html?gpt=true&asn=true&prt=none&smv=true&eng=true&tfs=true" line: 3}]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=MC&d_rtbd=json&d_ver=2&d_verify=1&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&ts=1557778404669 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "The resource at https://nexus.ensighten.com/bestbuy/prod/Bootstrap.js was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=MC&d_rtbd=json&d_ver=2&d_verify=1&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&ts=1557778404669. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Loading failed for the <script> with source https://nexus.ensighten.com/bestbuy/prod/Bootstrap.js." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 1}]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&ts=1557778404894 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&ts=1557778404894. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&d_cid_ic=visitortoken%01844020f2-75bb-11e9-9055-0a03487d4b28&ts=1557778404973 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&d_cid_ic=visitortoken%01844020f2-75bb-11e9-9055-0a03487d4b28&ts=1557778404973. (Reason: CORS request did not succeed)."]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.bestbuy.com - see bug description - <!-- @browser: Firefox Mobile 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:65.0) Gecko/65.0 Firefox/65.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001
**Browser / Version**: Firefox Mobile 65.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: If this is a ghost site how is my location dusplayed
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/5/f3fe9ee7-cb8b-4cc9-a7d8-ec90b84ca11e.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506082951</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: default</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://logx.optimizely.com/log/decision. (Reason: CORS request did not succeed)."]', u'[console.info(EventManager: Debug mode enabled via window.debugMode) https://assets.bbystatic.com/header-footer/store-locator.7bbb7e7eb239c54ecd82.0.js:1:159515]', u'[JavaScript Warning: "The resource at https://bestbuy.demdex.net/dest5.html?d_nsid=0#https%3A%2F%2Fwww.bestbuy.com%2Fsite%2Fdishwashers%2Fbuilt-in-dishwashers%2Fabcat0905001.c%3Fid%3Dabcat0905001 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "The resource at https://nexus.ensighten.com/bestbuy/tagframe_dev/Bootstrap.js was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://nexus.ensighten.com/bestbuy/tagframe_dev/Bootstrap.js." {file: "https://img.bbystatic.com/BestBuy_US/js/tracking/ens-index.html?gpt=true&asn=true&prt=none&smv=true&eng=true&tfs=true" line: 3}]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=MC&d_rtbd=json&d_ver=2&d_verify=1&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&ts=1557778404669 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "The resource at https://nexus.ensighten.com/bestbuy/prod/Bootstrap.js was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=MC&d_rtbd=json&d_ver=2&d_verify=1&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&ts=1557778404669. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Loading failed for the <script> with source https://nexus.ensighten.com/bestbuy/prod/Bootstrap.js." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 1}]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&ts=1557778404894 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&ts=1557778404894. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "The resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&d_cid_ic=visitortoken%01844020f2-75bb-11e9-9055-0a03487d4b28&ts=1557778404973 was blocked because content blocking is enabled." {file: "https://www.bestbuy.com/site/dishwashers/built-in-dishwashers/abcat0905001.c?id=abcat0905001" line: 0}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://dpm.demdex.net/id?d_visid_ver=3.1.2&d_fieldgroup=AAM&d_rtbd=json&d_ver=2&d_orgid=F6301253512D2BDB0A490D45%40AdobeOrg&d_nsid=0&d_mid=77636043971760990884317743444566127323&d_cid_ic=visitortoken%01844020f2-75bb-11e9-9055-0a03487d4b28&ts=1557778404973. (Reason: CORS request did not succeed)."]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description if this is a ghost site how is my location dusplayed steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked true basic gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel default console messages u u u u u u u u u u u u u u u u u u from with ❤️
| 0
|
232,652
| 18,894,757,734
|
IssuesEvent
|
2021-11-15 16:39:48
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
acceptance: TestComposeGSSPython failed
|
C-test-failure O-robot branch-master
|
acceptance.TestComposeGSSPython [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3734013&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3734013&tab=artifacts#/) on master @ [eeec6ff9f0053cfc105fb735b19fe0608683c996](https://github.com/cockroachdb/cockroach/commits/eeec6ff9f0053cfc105fb735b19fe0608683c996):
```
Digest: sha256:635f0aa53d99017b38d1a0aa5b2082f7812b03e3cdb299103fe77b5c8a07f1d2
Status: Downloaded newer image for alpine:3.14
---> 0a97eee8041e
Step 2/5 : RUN apk add --no-cache krb5-server && rm -rf /var/cache/apk/*
---> Running in e110aac3ba6a
fetch https://dl-cdn.alpinelinux.org/alpine/v3.14/main/x86_64/APKINDEX.tar.gz
fetch https://dl-cdn.alpinelinux.org/alpine/v3.14/community/x86_64/APKINDEX.tar.gz
(1/13) Installing libev (4.33-r0)
(2/13) Installing libverto (0.3.2-r0)
(3/13) Installing libverto-libev (0.3.2-r0)
(4/13) Installing libcom_err (1.46.2-r0)
(5/13) Installing krb5-conf (1.0-r2)
(6/13) Installing keyutils-libs (1.6.3-r0)
(7/13) Installing krb5-libs (1.18.4-r0)
(8/13) Installing gdbm (1.19-r0)
(9/13) Installing libsasl (2.1.27-r12)
(10/13) Installing libldap (2.4.58-r0)
(11/13) Installing krb5-server-ldap (1.18.4-r0)
(12/13) Installing e2fsprogs-libs (1.46.2-r0)
(13/13) Installing krb5-server (1.18.4-r0)
Executing busybox-1.33.1-r6.trigger
OK: 10 MiB in 27 packages
Removing intermediate container e110aac3ba6a
---> 0fd734218073
Step 3/5 : COPY krb5.conf /etc/krb5.conf
---> 3c66ac4b3d15
Step 4/5 : RUN kdb5_util create -s -P kpass && kadmin.local -q "addprinc -pw psql tester@MY.EX" && kadmin.local -q "addprinc -randkey postgres/gss_cockroach_1.gss_default@MY.EX"
---> Running in a5f4d8316a37
Loading random data
Initializing database '/var/lib/krb5kdc/principal' for realm 'MY.EX',
master key name 'K/M@MY.EX'
Authenticating as principal root/admin@MY.EX with password.
[91mNo policy specified for tester@MY.EX; defaulting to no policy
[0mPrincipal "tester@MY.EX" created.
Authenticating as principal root/admin@MY.EX with password.
[91mNo policy specified for postgres/gss_cockroach_1.gss_default@MY.EX; defaulting to no policy
[0mPrincipal "postgres/gss_cockroach_1.gss_default@MY.EX" created.
Removing intermediate container a5f4d8316a37
---> bca075c621e0
Step 5/5 : CMD ["/start.sh"]
---> Running in 93efb98a79b0
Removing intermediate container 93efb98a79b0
---> 822c14daec55
Successfully built 822c14daec55
Successfully tagged gss_kdc:latest
Pulling cockroach (ubuntu:xenial-20170214)...
Head https://registry-1.docker.io/v2/library/ubuntu/manifests/xenial-20170214: unknown: unable to complete public repo access query from database: context deadline exceeded
compose_test.go:54: exit status 1
--- FAIL: TestComposeGSSPython (4.71s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestComposeGSSPython.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
acceptance: TestComposeGSSPython failed - acceptance.TestComposeGSSPython [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3734013&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3734013&tab=artifacts#/) on master @ [eeec6ff9f0053cfc105fb735b19fe0608683c996](https://github.com/cockroachdb/cockroach/commits/eeec6ff9f0053cfc105fb735b19fe0608683c996):
```
Digest: sha256:635f0aa53d99017b38d1a0aa5b2082f7812b03e3cdb299103fe77b5c8a07f1d2
Status: Downloaded newer image for alpine:3.14
---> 0a97eee8041e
Step 2/5 : RUN apk add --no-cache krb5-server && rm -rf /var/cache/apk/*
---> Running in e110aac3ba6a
fetch https://dl-cdn.alpinelinux.org/alpine/v3.14/main/x86_64/APKINDEX.tar.gz
fetch https://dl-cdn.alpinelinux.org/alpine/v3.14/community/x86_64/APKINDEX.tar.gz
(1/13) Installing libev (4.33-r0)
(2/13) Installing libverto (0.3.2-r0)
(3/13) Installing libverto-libev (0.3.2-r0)
(4/13) Installing libcom_err (1.46.2-r0)
(5/13) Installing krb5-conf (1.0-r2)
(6/13) Installing keyutils-libs (1.6.3-r0)
(7/13) Installing krb5-libs (1.18.4-r0)
(8/13) Installing gdbm (1.19-r0)
(9/13) Installing libsasl (2.1.27-r12)
(10/13) Installing libldap (2.4.58-r0)
(11/13) Installing krb5-server-ldap (1.18.4-r0)
(12/13) Installing e2fsprogs-libs (1.46.2-r0)
(13/13) Installing krb5-server (1.18.4-r0)
Executing busybox-1.33.1-r6.trigger
OK: 10 MiB in 27 packages
Removing intermediate container e110aac3ba6a
---> 0fd734218073
Step 3/5 : COPY krb5.conf /etc/krb5.conf
---> 3c66ac4b3d15
Step 4/5 : RUN kdb5_util create -s -P kpass && kadmin.local -q "addprinc -pw psql tester@MY.EX" && kadmin.local -q "addprinc -randkey postgres/gss_cockroach_1.gss_default@MY.EX"
---> Running in a5f4d8316a37
Loading random data
Initializing database '/var/lib/krb5kdc/principal' for realm 'MY.EX',
master key name 'K/M@MY.EX'
Authenticating as principal root/admin@MY.EX with password.
[91mNo policy specified for tester@MY.EX; defaulting to no policy
[0mPrincipal "tester@MY.EX" created.
Authenticating as principal root/admin@MY.EX with password.
[91mNo policy specified for postgres/gss_cockroach_1.gss_default@MY.EX; defaulting to no policy
[0mPrincipal "postgres/gss_cockroach_1.gss_default@MY.EX" created.
Removing intermediate container a5f4d8316a37
---> bca075c621e0
Step 5/5 : CMD ["/start.sh"]
---> Running in 93efb98a79b0
Removing intermediate container 93efb98a79b0
---> 822c14daec55
Successfully built 822c14daec55
Successfully tagged gss_kdc:latest
Pulling cockroach (ubuntu:xenial-20170214)...
Head https://registry-1.docker.io/v2/library/ubuntu/manifests/xenial-20170214: unknown: unable to complete public repo access query from database: context deadline exceeded
compose_test.go:54: exit status 1
--- FAIL: TestComposeGSSPython (4.71s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestComposeGSSPython.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
acceptance testcomposegsspython failed acceptance testcomposegsspython with on master digest status downloaded newer image for alpine step run apk add no cache server rm rf var cache apk running in fetch fetch installing libev installing libverto installing libverto libev installing libcom err installing conf installing keyutils libs installing libs installing gdbm installing libsasl installing libldap installing server ldap installing libs installing server executing busybox trigger ok mib in packages removing intermediate container step copy conf etc conf step run util create s p kpass kadmin local q addprinc pw psql tester my ex kadmin local q addprinc randkey postgres gss cockroach gss default my ex running in loading random data initializing database var lib principal for realm my ex master key name k m my ex authenticating as principal root admin my ex with password policy specified for tester my ex defaulting to no policy tester my ex created authenticating as principal root admin my ex with password policy specified for postgres gss cockroach gss default my ex defaulting to no policy postgres gss cockroach gss default my ex created removing intermediate container step cmd running in removing intermediate container successfully built successfully tagged gss kdc latest pulling cockroach ubuntu xenial head unknown unable to complete public repo access query from database context deadline exceeded compose test go exit status fail testcomposegsspython help see also in this failure goflags json cc cockroachdb sql experience
| 1
|
215,671
| 16,686,725,873
|
IssuesEvent
|
2021-06-08 08:48:08
|
ita-social-projects/EventsExpress
|
https://api.github.com/repos/ita-social-projects/EventsExpress
|
opened
|
Verify that admin has a possibility to delete UnitOfMeasuring and Unit removal is synchronized with the event card
|
test
|
**Date Time**
06-08-2021 12:00
**Priority**
High
**Description**
Every admin has a possibility to delete UnitOfMeasuring. When the unit is deleted, it will be fully deleted from all the event cards.
Pre-requisites
1. Go to https://eventsexpress-test.azurewebsites.net
2. Authorize as the admin (e.g. User - admin@gmail.com, Password - 1qaz1qaz)
3. Move to the “Units of measuring” page
4. Select already existing 'Unit name'
5. Authorize as simply user (e.g. User - eventsexpress.testuser2@gmail.com, Password - Yaremchuk2001)
6. Move to the “Add event” page
7. As an authorized user create any event
**Test Cases**
| S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated |
|:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:|
| 1 | After creating the event move to 'Home' page | <br> | The currently event was created | <br> | <br> | <br> |
| 2 | Click on 'View' button in this event and choose the 'Unit name' from dropdown list below 'Count' label | 'kilometers' | <br> | <br> | <br> | <br> |
| 3 | Click on 'Save' button | <br> | The 'Short Name' of this unit is shown ('km') | <br> | <br> | <br> |
| 4 | Authorize as the admin | User - admin@gmail.com, Password - 1qaz1qaz | <br> | <br> | <br> | <br> |
| 5 | Click on 'Delete' button near the unit '#kolimeters' | <br> | An pop-up 'Do you really want to delete a unit of measuring' with a 'Unit name' and 'Short name' below and buttons 'Yes' and 'No' in it appear | <br> | <br> | <br> |
| 6 | Click on 'No' button | <br> | The 'Unit name' isn't deleted | <br> | <br> | <br> |
| 7 | Click on 'Yes' button | <br> | The 'Unit name' isn't deleted | <br> | <br> | <br> |
| 8 | Authorize as simply user , move to ' Home page' and click on 'View' button in this event | <br> | The unit is fully deleted from all the event cards | <br> | <br> | <br> |
**Screenshots**
If applicable, add screenshots to help explain your test.
**Environment:**
OS: [Windows 10]
Browser [Google Chrome]
Version [ 20210607.1]
**Additional context**
Link to story #347
|
1.0
|
Verify that admin has a possibility to delete UnitOfMeasuring and Unit removal is synchronized with the event card - **Date Time**
06-08-2021 12:00
**Priority**
High
**Description**
Every admin has a possibility to delete UnitOfMeasuring. When the unit is deleted, it will be fully deleted from all the event cards.
Pre-requisites
1. Go to https://eventsexpress-test.azurewebsites.net
2. Authorize as the admin (e.g. User - admin@gmail.com, Password - 1qaz1qaz)
3. Move to the “Units of measuring” page
4. Select already existing 'Unit name'
5. Authorize as simply user (e.g. User - eventsexpress.testuser2@gmail.com, Password - Yaremchuk2001)
6. Move to the “Add event” page
7. As an authorized user create any event
**Test Cases**
| S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated |
|:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:|
| 1 | After creating the event move to 'Home' page | <br> | The currently event was created | <br> | <br> | <br> |
| 2 | Click on 'View' button in this event and choose the 'Unit name' from dropdown list below 'Count' label | 'kilometers' | <br> | <br> | <br> | <br> |
| 3 | Click on 'Save' button | <br> | The 'Short Name' of this unit is shown ('km') | <br> | <br> | <br> |
| 4 | Authorize as the admin | User - admin@gmail.com, Password - 1qaz1qaz | <br> | <br> | <br> | <br> |
| 5 | Click on 'Delete' button near the unit '#kolimeters' | <br> | An pop-up 'Do you really want to delete a unit of measuring' with a 'Unit name' and 'Short name' below and buttons 'Yes' and 'No' in it appear | <br> | <br> | <br> |
| 6 | Click on 'No' button | <br> | The 'Unit name' isn't deleted | <br> | <br> | <br> |
| 7 | Click on 'Yes' button | <br> | The 'Unit name' isn't deleted | <br> | <br> | <br> |
| 8 | Authorize as simply user , move to ' Home page' and click on 'View' button in this event | <br> | The unit is fully deleted from all the event cards | <br> | <br> | <br> |
**Screenshots**
If applicable, add screenshots to help explain your test.
**Environment:**
OS: [Windows 10]
Browser [Google Chrome]
Version [ 20210607.1]
**Additional context**
Link to story #347
|
test
|
verify that admin has a possibility to delete unitofmeasuring and unit removal is synchronized with the event card date time priority high description every admin has a possibility to delete unitofmeasuring when the unit is deleted it will be fully deleted from all the event cards pre requisites go to authorize as the admin e g user admin gmail com password move to the “units of measuring” page select already existing unit name authorize as simply user e g user eventsexpress gmail com password move to the “add event” page as an authorized user create any event test cases s action test data expected result actual result p f automated after creating the event move to home page the currently event was created click on view button in this event and choose the unit name from dropdown list below count label kilometers click on save button the short name of this unit is shown km authorize as the admin user admin gmail com password click on delete button near the unit kolimeters an pop up do you really want to delete a unit of measuring with a unit name and short name below and buttons yes and no in it appear click on no button the unit name isn t deleted click on yes button the unit name isn t deleted authorize as simply user move to home page and click on view button in this event the unit is fully deleted from all the event cards screenshots if applicable add screenshots to help explain your test environment os browser version additional context link to story
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.