Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
9,553
12,514,584,126
IssuesEvent
2020-06-03 05:42:49
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
opened
Improve processing framework for mesh layers
Feature Request Mesh Processing
related to https://github.com/lutraconsulting/qgis-crayfish-plugin/issues/448 see https://github.com/qgis/QGIS/blob/3adbed503d5c1621d4dc8e4e217ce37835eceaae/src/core/processing/qgsprocessing.h#L44 We have to change TypeMesh to multiple enums like `TypeMeshEdges, TypeMeshVectices, TypeMeshFaces, TypeMeshAnyElementType` so in the processing algorithms we can select which type of mesh data we want to handle
1.0
Improve processing framework for mesh layers - related to https://github.com/lutraconsulting/qgis-crayfish-plugin/issues/448 see https://github.com/qgis/QGIS/blob/3adbed503d5c1621d4dc8e4e217ce37835eceaae/src/core/processing/qgsprocessing.h#L44 We have to change TypeMesh to multiple enums like `TypeMeshEdges, TypeMeshVectices, TypeMeshFaces, TypeMeshAnyElementType` so in the processing algorithms we can select which type of mesh data we want to handle
process
improve processing framework for mesh layers related to see we have to change typemesh to multiple enums like typemeshedges typemeshvectices typemeshfaces typemeshanyelementtype so in the processing algorithms we can select which type of mesh data we want to handle
1
17,208
22,793,151,761
IssuesEvent
2022-07-10 10:05:58
giacomorebecchi/food-pricing
https://api.github.com/repos/giacomorebecchi/food-pricing
opened
Empty description text is translated as "nan"
preprocessing
We should either add a custom word (e.g. "EMPTY_DESCRIPTION"), or leave it empty.
1.0
Empty description text is translated as "nan" - We should either add a custom word (e.g. "EMPTY_DESCRIPTION"), or leave it empty.
process
empty description text is translated as nan we should either add a custom word e g empty description or leave it empty
1
508,684
14,704,586,614
IssuesEvent
2021-01-04 16:43:16
GemsTracker/gemstracker-library
https://api.github.com/repos/GemsTracker/gemstracker-library
closed
Extend length if gla_remote_ip
bug impact major priority urgent
Extend the length of gla_remote_ip to allow storage of more complex $request->getClientIp() results, e.g. when working from behind a firewall.
1.0
Extend length if gla_remote_ip - Extend the length of gla_remote_ip to allow storage of more complex $request->getClientIp() results, e.g. when working from behind a firewall.
non_process
extend length if gla remote ip extend the length of gla remote ip to allow storage of more complex request getclientip results e g when working from behind a firewall
0
9,077
12,148,979,467
IssuesEvent
2020-04-24 15:24:58
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Alternative to downloading locally
Pri2 cxp doc-enhancement machine-learning/svc team-data-science-process/subsvc triaged
It would be helpful to know if/how one can avoid downloading locally for serverless apps. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2f45a6b5-0fea-7fbb-5d4d-37e0e2583fd7 * Version Independent ID: 7be6f792-09f8-c22f-86b6-d0f690e9b3a4 * Content: [Explore data in Azure blob storage with pandas - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/explore-data-blob#feedback) * Content Source: [articles/machine-learning/team-data-science-process/explore-data-blob.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/explore-data-blob.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
1.0
Alternative to downloading locally - It would be helpful to know if/how one can avoid downloading locally for serverless apps. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2f45a6b5-0fea-7fbb-5d4d-37e0e2583fd7 * Version Independent ID: 7be6f792-09f8-c22f-86b6-d0f690e9b3a4 * Content: [Explore data in Azure blob storage with pandas - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/explore-data-blob#feedback) * Content Source: [articles/machine-learning/team-data-science-process/explore-data-blob.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/explore-data-blob.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
process
alternative to downloading locally it would be helpful to know if how one can avoid downloading locally for serverless apps document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service machine learning sub service team data science process github login marktab microsoft alias tdsp
1
21,282
28,455,615,831
IssuesEvent
2023-04-17 06:48:33
sap-tutorials/sap-build-process-automation
https://api.github.com/repos/sap-tutorials/sap-build-process-automation
closed
Run the Business Process
tutorial:spa-run-process
Tutorials: https://developers.sap.com/tutorials/spa-run-process.html -------------------------- Hi, during the deployment of the order process. I'm getting asked to enter run variables and to enter a trigger. Both options are not mentioned on the tutorial page. I guess it makes sense to add at least a hint to these panels. Moreover, instead of Undeploy a Redeploy button is shown. BR, Ingo
1.0
Run the Business Process - Tutorials: https://developers.sap.com/tutorials/spa-run-process.html -------------------------- Hi, during the deployment of the order process. I'm getting asked to enter run variables and to enter a trigger. Both options are not mentioned on the tutorial page. I guess it makes sense to add at least a hint to these panels. Moreover, instead of Undeploy a Redeploy button is shown. BR, Ingo
process
run the business process tutorials hi during the deployment of the order process i m getting asked to enter run variables and to enter a trigger both options are not mentioned on the tutorial page i guess it makes sense to add at least a hint to these panels moreover instead of undeploy a redeploy button is shown br ingo
1
42,190
6,969,592,257
IssuesEvent
2017-12-11 06:30:25
benwiley4000/react-responsive-audio-player
https://api.github.com/repos/benwiley4000/react-responsive-audio-player
opened
Remove propType annotations from AudioPlayer.js
documentation
Maybe we'll bring this back if we do some jsdoc type thing in the future, but for now it's too much to maintain multiple descriptions of the proptypes accepted by `AudioPlayer`.
1.0
Remove propType annotations from AudioPlayer.js - Maybe we'll bring this back if we do some jsdoc type thing in the future, but for now it's too much to maintain multiple descriptions of the proptypes accepted by `AudioPlayer`.
non_process
remove proptype annotations from audioplayer js maybe we ll bring this back if we do some jsdoc type thing in the future but for now it s too much to maintain multiple descriptions of the proptypes accepted by audioplayer
0
59,140
8,338,297,858
IssuesEvent
2018-09-28 13:56:34
coenvalk/conductAR
https://api.github.com/repos/coenvalk/conductAR
closed
Usage Documentation for Music Playback
documentation
Create documents to explain usage for music playback scripts.
1.0
Usage Documentation for Music Playback - Create documents to explain usage for music playback scripts.
non_process
usage documentation for music playback create documents to explain usage for music playback scripts
0
17,130
22,649,071,951
IssuesEvent
2022-07-01 11:41:56
PyCQA/pylint
https://api.github.com/repos/PyCQA/pylint
closed
Different results with -j4 and -j8
Bug :beetle: topic-multiprocessing
I'm not sure if this is a duplicate of #2573 or #374 because I don't know if those are particular to the specific checks they mention. ### Steps to reproduce 1. `git clone https://github.com/nedbat/coveragepy; cd coveragepy` 2. `git checkout 50cc68846f` 3. `pip install tox` 4. `tox -e lint --notest` 5. `.tox/lint/bin/python -m pylint --notes= -j 4 coverage tests doc ci igor.py setup.py __main__.py` 6. No violations are reported 7. `.tox/lint/bin/python -m pylint --notes= -j 8 coverage tests doc ci igor.py setup.py __main__.py` ``` ************* Module coverage.pytracer coverage/pytracer.py:140 E: self.should_start_context is not callable (not-callable) coverage/pytracer.py:165 E: self.should_trace is not callable (not-callable) coverage/pytracer.py:166 E: 'self.should_trace_cache' does not support item assignment (unsupported-assignment-operation) coverage/pytracer.py:171 E: Value 'self.data' doesn't support membership test (unsupported-membership-test) coverage/pytracer.py:172 E: 'self.data' does not support item assignment (unsupported-assignment-operation) coverage/pytracer.py:173 E: Value 'self.data' is unsubscriptable (unsubscriptable-object) coverage/pytracer.py:259 E: self.warn is not callable (not-callable) ``` A different manifestation of what might be the same underlying cause: in my CI, running -j4, I get occasional failures with these same violations reported. ### Expected behavior The number of processes shouldn't change the results, only the performance. ### pylint --version output Result of `pylint --version` output: ``` $ .tox/lint/bin/python -m pylint --version pylint 2.7.4 astroid 2.5.3 Python 3.8.8 (default, Mar 21 2021, 13:50:20) [Clang 12.0.0 (clang-1200.0.32.29)] ``` <!-- Additional dependencies: ``` pandas==0.23.2 marshmallow==3.10.0 ... ``` -->
1.0
Different results with -j4 and -j8 - I'm not sure if this is a duplicate of #2573 or #374 because I don't know if those are particular to the specific checks they mention. ### Steps to reproduce 1. `git clone https://github.com/nedbat/coveragepy; cd coveragepy` 2. `git checkout 50cc68846f` 3. `pip install tox` 4. `tox -e lint --notest` 5. `.tox/lint/bin/python -m pylint --notes= -j 4 coverage tests doc ci igor.py setup.py __main__.py` 6. No violations are reported 7. `.tox/lint/bin/python -m pylint --notes= -j 8 coverage tests doc ci igor.py setup.py __main__.py` ``` ************* Module coverage.pytracer coverage/pytracer.py:140 E: self.should_start_context is not callable (not-callable) coverage/pytracer.py:165 E: self.should_trace is not callable (not-callable) coverage/pytracer.py:166 E: 'self.should_trace_cache' does not support item assignment (unsupported-assignment-operation) coverage/pytracer.py:171 E: Value 'self.data' doesn't support membership test (unsupported-membership-test) coverage/pytracer.py:172 E: 'self.data' does not support item assignment (unsupported-assignment-operation) coverage/pytracer.py:173 E: Value 'self.data' is unsubscriptable (unsubscriptable-object) coverage/pytracer.py:259 E: self.warn is not callable (not-callable) ``` A different manifestation of what might be the same underlying cause: in my CI, running -j4, I get occasional failures with these same violations reported. ### Expected behavior The number of processes shouldn't change the results, only the performance. ### pylint --version output Result of `pylint --version` output: ``` $ .tox/lint/bin/python -m pylint --version pylint 2.7.4 astroid 2.5.3 Python 3.8.8 (default, Mar 21 2021, 13:50:20) [Clang 12.0.0 (clang-1200.0.32.29)] ``` <!-- Additional dependencies: ``` pandas==0.23.2 marshmallow==3.10.0 ... ``` -->
process
different results with and i m not sure if this is a duplicate of or because i don t know if those are particular to the specific checks they mention steps to reproduce git clone cd coveragepy git checkout pip install tox tox e lint notest tox lint bin python m pylint notes j coverage tests doc ci igor py setup py main py no violations are reported tox lint bin python m pylint notes j coverage tests doc ci igor py setup py main py module coverage pytracer coverage pytracer py e self should start context is not callable not callable coverage pytracer py e self should trace is not callable not callable coverage pytracer py e self should trace cache does not support item assignment unsupported assignment operation coverage pytracer py e value self data doesn t support membership test unsupported membership test coverage pytracer py e self data does not support item assignment unsupported assignment operation coverage pytracer py e value self data is unsubscriptable unsubscriptable object coverage pytracer py e self warn is not callable not callable a different manifestation of what might be the same underlying cause in my ci running i get occasional failures with these same violations reported expected behavior the number of processes shouldn t change the results only the performance pylint version output result of pylint version output tox lint bin python m pylint version pylint astroid python default mar additional dependencies pandas marshmallow
1
17,894
23,868,683,822
IssuesEvent
2022-09-07 13:15:18
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[BUG] SQL Connector key fields with avro format returns null schema
compute/data-processing type/bug
![image.png](https://images.zenhubusercontent.com/61564d60ef4304714f6337a4/552b8d7b-7b7c-455a-b8b4-8912512eac20) When running the PulsarTableITCase with the key and partial value, the tests failed because for the keySerailization the returned schema was null. I loooked into the code it should not return null. It might be a bug in the flink-avro format and we need to find out why if we want to use the format for key + value.
1.0
[BUG] SQL Connector key fields with avro format returns null schema - ![image.png](https://images.zenhubusercontent.com/61564d60ef4304714f6337a4/552b8d7b-7b7c-455a-b8b4-8912512eac20) When running the PulsarTableITCase with the key and partial value, the tests failed because for the keySerailization the returned schema was null. I loooked into the code it should not return null. It might be a bug in the flink-avro format and we need to find out why if we want to use the format for key + value.
process
sql connector key fields with avro format returns null schema when running the pulsartableitcase with the key and partial value the tests failed because for the keyserailization the returned schema was null i loooked into the code it should not return null it might be a bug in the flink avro format and we need to find out why if we want to use the format for key value
1
215,813
24,196,526,426
IssuesEvent
2022-09-24 01:12:41
haeli05/source
https://api.github.com/repos/haeli05/source
opened
CVE-2020-36604 (Medium) detected in hoek-8.2.4.tgz
security vulnerability
## CVE-2020-36604 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-8.2.4.tgz</b></p></summary> <p>General purpose node utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/@hapi/hoek/-/hoek-8.2.4.tgz">https://registry.npmjs.org/@hapi/hoek/-/hoek-8.2.4.tgz</a></p> <p>Path to dependency file: /FrontEnd/package.json</p> <p>Path to vulnerable library: /FrontEnd/node_modules/@hapi/hoek/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.1.1.tgz (Root Library) - workbox-webpack-plugin-4.3.1.tgz - workbox-build-4.3.1.tgz - joi-15.1.1.tgz - :x: **hoek-8.2.4.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> hoek before 8.5.1 and 9.x before 9.0.3 allows prototype poisoning in the clone function. <p>Publish Date: 2022-09-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36604>CVE-2020-36604</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-23</p> <p>Fix Resolution (@hapi/hoek): 8.5.1</p> <p>Direct dependency fix Resolution (react-scripts): 3.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36604 (Medium) detected in hoek-8.2.4.tgz - ## CVE-2020-36604 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-8.2.4.tgz</b></p></summary> <p>General purpose node utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/@hapi/hoek/-/hoek-8.2.4.tgz">https://registry.npmjs.org/@hapi/hoek/-/hoek-8.2.4.tgz</a></p> <p>Path to dependency file: /FrontEnd/package.json</p> <p>Path to vulnerable library: /FrontEnd/node_modules/@hapi/hoek/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.1.1.tgz (Root Library) - workbox-webpack-plugin-4.3.1.tgz - workbox-build-4.3.1.tgz - joi-15.1.1.tgz - :x: **hoek-8.2.4.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> hoek before 8.5.1 and 9.x before 9.0.3 allows prototype poisoning in the clone function. <p>Publish Date: 2022-09-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36604>CVE-2020-36604</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-23</p> <p>Fix Resolution (@hapi/hoek): 8.5.1</p> <p>Direct dependency fix Resolution (react-scripts): 3.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in hoek tgz cve medium severity vulnerability vulnerable library hoek tgz general purpose node utilities library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules hapi hoek package json dependency hierarchy react scripts tgz root library workbox webpack plugin tgz workbox build tgz joi tgz x hoek tgz vulnerable library vulnerability details hoek before and x before allows prototype poisoning in the clone function publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution hapi hoek direct dependency fix resolution react scripts step up your open source security game with mend
0
410,553
11,993,566,094
IssuesEvent
2020-04-08 12:15:20
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
[4.1 beta 1]: Can't create manual order
bug component: order priority: high
**Describe the bug** Can't seem to create manual order while testing 4.1 beta 1. Tried on 2 Pressable test sites. **To Reproduce** Steps to reproduce the behavior: 1. Create any product on the site; 2. Try creating manual order; 3. See error: ![Screen Shot 2020-04-07 at 7 45 32 PM](https://user-images.githubusercontent.com/19143190/78729746-d9dc1580-7908-11ea-8941-4266a8f6911c.png) Error: ``` Uncaught Error: Call to a member function format() on null in /srv/htdocs/wp-content/plugins/woocommerce/includes/tracks/events/class-wc-order-tracking.php on line 30 track_order_viewed() wp-includes/class-wp-hook.php:287 apply_filters() wp-includes/class-wp-hook.php:311 do_action() wp-includes/plugin.php:478 do_action() wp-content/plugins/woocommerce/includes/admin/meta-boxes/class-wc-meta-box-order-data.php:310 output() wp-admin/includes/template.php:1345 do_meta_boxes() wp-admin/edit-form-advanced.php:677 require() wp-admin/post-new.php:75 ``` **Screenshots** See above. **Expected behavior** I expected to be able to create manual order without any issues. **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [x] I can reproduce this bug consistently using the steps above. **WordPress Environment** <details> ``` ` ### WordPress Environment ### WordPress address (URL): https://jamosova2.mystagingwebsite.com Site address (URL): https://jamosova2.mystagingwebsite.com WC Version: 4.1.0 REST API Version: ✔ 1.0.7 WC Blocks Version: ✔ 2.5.16 Action Scheduler Version: ✔ 3.1.4 WC Admin Version: ✔ 1.1.0 Log Directory Writable: ✔ WP Version: 5.4 WP Multisite: – WP Memory Limit: 256 MB WP Debug Mode: – WP Cron: ✔ Language: en_US External object cache: ✔ ### Server Environment ### Server Info: nginx PHP Version: 7.4.4 PHP Post Max Size: 2 GB PHP Time Limit: 1200 PHP Max Input Vars: 6144 cURL Version: 7.69.1 OpenSSL/1.1.0l SUHOSIN Installed: – MySQL Version: 5.5.5-10.3.21-MariaDB-log Max Upload Size: 2 GB Default Timezone is UTC: ✔ fsockopen/cURL: ✔ SoapClient: ✔ DOMDocument: ✔ GZip: ✔ Multibyte String: ✔ Remote Post: ✔ Remote Get: ✔ ### Database ### WC Database Version: 4.1.0 WC Database Prefix: wp_ Total Database Size: 11.61MB Database Data Size: 6.47MB Database Index Size: 5.14MB wp_woocommerce_sessions: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_order_items: Data: 0.05MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_order_itemmeta: Data: 0.25MB + Index: 0.25MB + Engine InnoDB wp_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_actions: Data: 0.31MB + Index: 0.36MB + Engine InnoDB wp_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_logs: Data: 0.31MB + Index: 0.25MB + Engine InnoDB wp_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_comments: Data: 0.06MB + Index: 0.09MB + Engine InnoDB wp_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_ms_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_options: Data: 1.17MB + Index: 0.09MB + Engine InnoDB wp_postmeta: Data: 1.52MB + Index: 1.88MB + Engine InnoDB wp_posts: Data: 1.30MB + Index: 0.22MB + Engine InnoDB wp_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_termmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_terms: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_term_relationships: Data: 0.08MB + Index: 0.05MB + Engine InnoDB wp_term_taxonomy: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_usermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_deposits_payment_plans: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_deposits_payment_plans_schedule: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_product_meta_lookup: Data: 0.06MB + Index: 0.09MB + Engine InnoDB wp_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_gzd_dhl_labelmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_dhl_labels: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipmentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipments: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipment_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipment_items: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_gzd_shipping_provider: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_gzd_shipping_providermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wsal_metadata: Data: 0.39MB + Index: 0.61MB + Engine InnoDB wp_wsal_occurrences: Data: 0.05MB + Index: 0.03MB + Engine InnoDB wp_wsal_options: Data: 0.02MB + Index: 0.02MB + Engine InnoDB ### Post Type Counts ### attachment: 204 jetpack_migration: 2 jp_img_sitemap: 4 jp_sitemap: 4 jp_sitemap_master: 4 nav_menu_item: 11 page: 9 post: 23 product: 259 product_variation: 188 revision: 45 shop_coupon: 6 shop_order: 150 shop_order_refund: 18 shop_subscription: 2 ### Security ### Secure connection (HTTPS): ✔ Hide errors from visitors: ✔ ### Active Plugins (5) ### Query Monitor: by John Blackbourn – 3.5.2 Akismet Anti-Spam: by Automattic – 4.1.4 Jetpack by WordPress.com: by Automattic – 8.4.1 WooCommerce Beta Tester: by WooCommerce – 2.0.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce: by Automattic – 4.1.0-beta.1 ### Inactive Plugins (30) ### Advanced Custom Fields: by Elliot Condon – 5.8.7 Email Templates: by Damian Logghe – 1.3.1.2 Facebook for WooCommerce: by Facebook – 1.10.0 – Installed version not tested with active version of WooCommerce 4.1.0 Germanized for WooCommerce: by Vendidero – 3.1.4 – Installed version not tested with active version of WooCommerce 4.1.0 Gutenberg: by Gutenberg Team – 7.1.0 Loco Translate: by Tim Whitlock – 2.3.1 Payment Gateway Based Fees and Discounts for WooCommerce: by Tyche Softwares – 2.6 – Installed version not tested with active version of WooCommerce 4.1.0 Storefront Powerpack: by WooCommerce – 1.5.0 SVG Support: by Benbodhi – 2.3.15 WooCommerce Admin: by WooCommerce – 0.24.0 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Blocks: by Automattic – 2.5.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Customer/Order CSV Export: by SkyVerge – 4.3.5 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Deposits: by Automattic – 1.3.3 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Dynamic Pricing: by Lucas Stark – 3.1.10 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce PayPal Checkout Gateway: by WooCommerce – 1.6.18 (update to version 1.6.20 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce PayPal Powered by Braintree Gateway: by WooCommerce – 2.0.3 (update to version 2.3.8 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Print Invoices/Packing Lists: by SkyVerge – 3.6.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Product Add-ons: by WooCommerce – 2.9.7 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Sequential Order Numbers Pro: by SkyVerge – 1.11.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Services: by Automattic – 1.22.2 (update to version 1.22.5 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stamps.com API integration: by WooCommerce – 1.3.15 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stamps.com Export Suite: by SkyVerge – 2.10.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Store Catalog PDF Download: by WooCommerce – 1.0.8 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stripe Gateway: by WooCommerce – 4.3.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Subscriptions: by WooCommerce – 3.0.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce UPS Shipping: by WooCommerce – 3.2.19 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Wholesale Prices: by Rymera Web Co – 1.11.1 – Installed version not tested with active version of WooCommerce 4.1.0 WP Crontrol: by John Blackbourn & crontributors – 1.7.1 WP Migrate DB: by Delicious Brains – 0.9.2 WP Security Audit Log: by WP White Security – 4.0.1 ### Dropin Plugins (3) ### advanced-cache.php: advanced-cache.php db.php: Query Monitor Database Class object-cache.php: Memcached ### Settings ### API Enabled: ✔ Force SSL: ✔ Currency: USD ($) Currency Position: left Thousand Separator: , Decimal Separator: . Number of Decimals: 2 Taxonomies: Product Types: external (external) grouped (grouped) simple (simple) subscription (subscription) variable (variable) variable subscription (variable-subscription) Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog) exclude-from-search (exclude-from-search) featured (featured) outofstock (outofstock) rated-1 (rated-1) rated-2 (rated-2) rated-3 (rated-3) rated-4 (rated-4) rated-5 (rated-5) Connected to WooCommerce.com: – ### WC Pages ### Shop base: #4 - /shop/ Cart: #5 - /cart/ Checkout: #6 - /checkout/ My account: #7 - /my-account/ Terms and conditions: ❌ Page not set ### Theme ### Name: Storefront Version: 2.5.5 Author URL: https://woocommerce.com/ Child Theme: ❌ – If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme WooCommerce Support: ✔ ### Templates ### Overrides: – ### Action Scheduler ### Complete: 1,047 Oldest: 2020-03-09 06:51:33 -0400 Newest: 2020-04-07 19:52:08 -0400 Pending: 4 Oldest: 2020-04-07 19:52:14 -0400 Newest: 2020-05-06 09:06:20 -0400 ` ``` </details>
1.0
[4.1 beta 1]: Can't create manual order - **Describe the bug** Can't seem to create manual order while testing 4.1 beta 1. Tried on 2 Pressable test sites. **To Reproduce** Steps to reproduce the behavior: 1. Create any product on the site; 2. Try creating manual order; 3. See error: ![Screen Shot 2020-04-07 at 7 45 32 PM](https://user-images.githubusercontent.com/19143190/78729746-d9dc1580-7908-11ea-8941-4266a8f6911c.png) Error: ``` Uncaught Error: Call to a member function format() on null in /srv/htdocs/wp-content/plugins/woocommerce/includes/tracks/events/class-wc-order-tracking.php on line 30 track_order_viewed() wp-includes/class-wp-hook.php:287 apply_filters() wp-includes/class-wp-hook.php:311 do_action() wp-includes/plugin.php:478 do_action() wp-content/plugins/woocommerce/includes/admin/meta-boxes/class-wc-meta-box-order-data.php:310 output() wp-admin/includes/template.php:1345 do_meta_boxes() wp-admin/edit-form-advanced.php:677 require() wp-admin/post-new.php:75 ``` **Screenshots** See above. **Expected behavior** I expected to be able to create manual order without any issues. **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [x] I can reproduce this bug consistently using the steps above. **WordPress Environment** <details> ``` ` ### WordPress Environment ### WordPress address (URL): https://jamosova2.mystagingwebsite.com Site address (URL): https://jamosova2.mystagingwebsite.com WC Version: 4.1.0 REST API Version: ✔ 1.0.7 WC Blocks Version: ✔ 2.5.16 Action Scheduler Version: ✔ 3.1.4 WC Admin Version: ✔ 1.1.0 Log Directory Writable: ✔ WP Version: 5.4 WP Multisite: – WP Memory Limit: 256 MB WP Debug Mode: – WP Cron: ✔ Language: en_US External object cache: ✔ ### Server Environment ### Server Info: nginx PHP Version: 7.4.4 PHP Post Max Size: 2 GB PHP Time Limit: 1200 PHP Max Input Vars: 6144 cURL Version: 7.69.1 OpenSSL/1.1.0l SUHOSIN Installed: – MySQL Version: 5.5.5-10.3.21-MariaDB-log Max Upload Size: 2 GB Default Timezone is UTC: ✔ fsockopen/cURL: ✔ SoapClient: ✔ DOMDocument: ✔ GZip: ✔ Multibyte String: ✔ Remote Post: ✔ Remote Get: ✔ ### Database ### WC Database Version: 4.1.0 WC Database Prefix: wp_ Total Database Size: 11.61MB Database Data Size: 6.47MB Database Index Size: 5.14MB wp_woocommerce_sessions: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_order_items: Data: 0.05MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_order_itemmeta: Data: 0.25MB + Index: 0.25MB + Engine InnoDB wp_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_actions: Data: 0.31MB + Index: 0.36MB + Engine InnoDB wp_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_actionscheduler_logs: Data: 0.31MB + Index: 0.25MB + Engine InnoDB wp_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_comments: Data: 0.06MB + Index: 0.09MB + Engine InnoDB wp_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_ms_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_options: Data: 1.17MB + Index: 0.09MB + Engine InnoDB wp_postmeta: Data: 1.52MB + Index: 1.88MB + Engine InnoDB wp_posts: Data: 1.30MB + Index: 0.22MB + Engine InnoDB wp_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_termmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_terms: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_term_relationships: Data: 0.08MB + Index: 0.05MB + Engine InnoDB wp_term_taxonomy: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_usermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_deposits_payment_plans: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_deposits_payment_plans_schedule: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB wp_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wc_product_meta_lookup: Data: 0.06MB + Index: 0.09MB + Engine InnoDB wp_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB wp_woocommerce_gzd_dhl_labelmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_dhl_labels: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipmentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipments: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipment_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_woocommerce_gzd_shipment_items: Data: 0.02MB + Index: 0.06MB + Engine InnoDB wp_woocommerce_gzd_shipping_provider: Data: 0.02MB + Index: 0.00MB + Engine InnoDB wp_woocommerce_gzd_shipping_providermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB wp_wsal_metadata: Data: 0.39MB + Index: 0.61MB + Engine InnoDB wp_wsal_occurrences: Data: 0.05MB + Index: 0.03MB + Engine InnoDB wp_wsal_options: Data: 0.02MB + Index: 0.02MB + Engine InnoDB ### Post Type Counts ### attachment: 204 jetpack_migration: 2 jp_img_sitemap: 4 jp_sitemap: 4 jp_sitemap_master: 4 nav_menu_item: 11 page: 9 post: 23 product: 259 product_variation: 188 revision: 45 shop_coupon: 6 shop_order: 150 shop_order_refund: 18 shop_subscription: 2 ### Security ### Secure connection (HTTPS): ✔ Hide errors from visitors: ✔ ### Active Plugins (5) ### Query Monitor: by John Blackbourn – 3.5.2 Akismet Anti-Spam: by Automattic – 4.1.4 Jetpack by WordPress.com: by Automattic – 8.4.1 WooCommerce Beta Tester: by WooCommerce – 2.0.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce: by Automattic – 4.1.0-beta.1 ### Inactive Plugins (30) ### Advanced Custom Fields: by Elliot Condon – 5.8.7 Email Templates: by Damian Logghe – 1.3.1.2 Facebook for WooCommerce: by Facebook – 1.10.0 – Installed version not tested with active version of WooCommerce 4.1.0 Germanized for WooCommerce: by Vendidero – 3.1.4 – Installed version not tested with active version of WooCommerce 4.1.0 Gutenberg: by Gutenberg Team – 7.1.0 Loco Translate: by Tim Whitlock – 2.3.1 Payment Gateway Based Fees and Discounts for WooCommerce: by Tyche Softwares – 2.6 – Installed version not tested with active version of WooCommerce 4.1.0 Storefront Powerpack: by WooCommerce – 1.5.0 SVG Support: by Benbodhi – 2.3.15 WooCommerce Admin: by WooCommerce – 0.24.0 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Blocks: by Automattic – 2.5.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Customer/Order CSV Export: by SkyVerge – 4.3.5 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Deposits: by Automattic – 1.3.3 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Dynamic Pricing: by Lucas Stark – 3.1.10 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce PayPal Checkout Gateway: by WooCommerce – 1.6.18 (update to version 1.6.20 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce PayPal Powered by Braintree Gateway: by WooCommerce – 2.0.3 (update to version 2.3.8 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Print Invoices/Packing Lists: by SkyVerge – 3.6.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Product Add-ons: by WooCommerce – 2.9.7 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Sequential Order Numbers Pro: by SkyVerge – 1.11.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Services: by Automattic – 1.22.2 (update to version 1.22.5 is available) – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stamps.com API integration: by WooCommerce – 1.3.15 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stamps.com Export Suite: by SkyVerge – 2.10.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Store Catalog PDF Download: by WooCommerce – 1.0.8 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Stripe Gateway: by WooCommerce – 4.3.2 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Subscriptions: by WooCommerce – 3.0.1 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce UPS Shipping: by WooCommerce – 3.2.19 – Installed version not tested with active version of WooCommerce 4.1.0 WooCommerce Wholesale Prices: by Rymera Web Co – 1.11.1 – Installed version not tested with active version of WooCommerce 4.1.0 WP Crontrol: by John Blackbourn & crontributors – 1.7.1 WP Migrate DB: by Delicious Brains – 0.9.2 WP Security Audit Log: by WP White Security – 4.0.1 ### Dropin Plugins (3) ### advanced-cache.php: advanced-cache.php db.php: Query Monitor Database Class object-cache.php: Memcached ### Settings ### API Enabled: ✔ Force SSL: ✔ Currency: USD ($) Currency Position: left Thousand Separator: , Decimal Separator: . Number of Decimals: 2 Taxonomies: Product Types: external (external) grouped (grouped) simple (simple) subscription (subscription) variable (variable) variable subscription (variable-subscription) Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog) exclude-from-search (exclude-from-search) featured (featured) outofstock (outofstock) rated-1 (rated-1) rated-2 (rated-2) rated-3 (rated-3) rated-4 (rated-4) rated-5 (rated-5) Connected to WooCommerce.com: – ### WC Pages ### Shop base: #4 - /shop/ Cart: #5 - /cart/ Checkout: #6 - /checkout/ My account: #7 - /my-account/ Terms and conditions: ❌ Page not set ### Theme ### Name: Storefront Version: 2.5.5 Author URL: https://woocommerce.com/ Child Theme: ❌ – If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme WooCommerce Support: ✔ ### Templates ### Overrides: – ### Action Scheduler ### Complete: 1,047 Oldest: 2020-03-09 06:51:33 -0400 Newest: 2020-04-07 19:52:08 -0400 Pending: 4 Oldest: 2020-04-07 19:52:14 -0400 Newest: 2020-05-06 09:06:20 -0400 ` ``` </details>
non_process
can t create manual order describe the bug can t seem to create manual order while testing beta tried on pressable test sites to reproduce steps to reproduce the behavior create any product on the site try creating manual order see error error uncaught error call to a member function format on null in srv htdocs wp content plugins woocommerce includes tracks events class wc order tracking php on line track order viewed wp includes class wp hook php apply filters wp includes class wp hook php do action wp includes plugin php do action wp content plugins woocommerce includes admin meta boxes class wc meta box order data php output wp admin includes template php do meta boxes wp admin edit form advanced php require wp admin post new php screenshots see above expected behavior i expected to be able to create manual order without any issues isolating the problem mark completed items with an i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above wordpress environment wordpress environment wordpress address url site address url wc version rest api version ✔ wc blocks version ✔ action scheduler version ✔ wc admin version ✔ log directory writable ✔ wp version wp multisite – wp memory limit mb wp debug mode – wp cron ✔ language en us external object cache ✔ server environment server info nginx php version php post max size gb php time limit php max input vars curl version openssl suhosin installed – mysql version mariadb log max upload size gb default timezone is utc ✔ fsockopen curl ✔ soapclient ✔ domdocument ✔ gzip ✔ multibyte string ✔ remote post ✔ remote get ✔ database wc database version wc database prefix wp total database size database data size database index size wp woocommerce sessions data index engine innodb wp woocommerce api keys data index engine innodb wp woocommerce attribute taxonomies data index engine innodb wp woocommerce downloadable product permissions data index engine innodb wp woocommerce order items data index engine innodb wp woocommerce order itemmeta data index engine innodb wp woocommerce tax rates data index engine innodb wp woocommerce tax rate locations data index engine innodb wp woocommerce shipping zones data index engine innodb wp woocommerce shipping zone locations data index engine innodb wp woocommerce shipping zone methods data index engine innodb wp woocommerce payment tokens data index engine innodb wp woocommerce payment tokenmeta data index engine innodb wp woocommerce log data index engine innodb wp actionscheduler actions data index engine innodb wp actionscheduler claims data index engine innodb wp actionscheduler groups data index engine innodb wp actionscheduler logs data index engine innodb wp commentmeta data index engine innodb wp comments data index engine innodb wp links data index engine innodb wp ms snippets data index engine innodb wp options data index engine innodb wp postmeta data index engine innodb wp posts data index engine innodb wp snippets data index engine innodb wp termmeta data index engine innodb wp terms data index engine innodb wp term relationships data index engine innodb wp term taxonomy data index engine innodb wp usermeta data index engine innodb wp users data index engine innodb wp wc admin notes data index engine innodb wp wc admin note actions data index engine innodb wp wc category lookup data index engine innodb wp wc customer lookup data index engine innodb wp wc deposits payment plans data index engine innodb wp wc deposits payment plans schedule data index engine innodb wp wc download log data index engine innodb wp wc order coupon lookup data index engine innodb wp wc order product lookup data index engine innodb wp wc order stats data index engine innodb wp wc order tax lookup data index engine innodb wp wc product meta lookup data index engine innodb wp wc reserved stock data index engine innodb wp wc tax rate classes data index engine innodb wp wc webhooks data index engine innodb wp woocommerce gzd dhl labelmeta data index engine innodb wp woocommerce gzd dhl labels data index engine innodb wp woocommerce gzd shipmentmeta data index engine innodb wp woocommerce gzd shipments data index engine innodb wp woocommerce gzd shipment itemmeta data index engine innodb wp woocommerce gzd shipment items data index engine innodb wp woocommerce gzd shipping provider data index engine innodb wp woocommerce gzd shipping providermeta data index engine innodb wp wsal metadata data index engine innodb wp wsal occurrences data index engine innodb wp wsal options data index engine innodb post type counts attachment jetpack migration jp img sitemap jp sitemap jp sitemap master nav menu item page post product product variation revision shop coupon shop order shop order refund shop subscription security secure connection https ✔ hide errors from visitors ✔ active plugins query monitor by john blackbourn – akismet anti spam by automattic – jetpack by wordpress com by automattic – woocommerce beta tester by woocommerce – – installed version not tested with active version of woocommerce woocommerce by automattic – beta inactive plugins advanced custom fields by elliot condon – email templates by damian logghe – facebook for woocommerce by facebook – – installed version not tested with active version of woocommerce germanized for woocommerce by vendidero – – installed version not tested with active version of woocommerce gutenberg by gutenberg team – loco translate by tim whitlock – payment gateway based fees and discounts for woocommerce by tyche softwares – – installed version not tested with active version of woocommerce storefront powerpack by woocommerce – svg support by benbodhi – woocommerce admin by woocommerce – – installed version not tested with active version of woocommerce woocommerce blocks by automattic – – installed version not tested with active version of woocommerce woocommerce customer order csv export by skyverge – – installed version not tested with active version of woocommerce woocommerce deposits by automattic – – installed version not tested with active version of woocommerce woocommerce dynamic pricing by lucas stark – – installed version not tested with active version of woocommerce woocommerce paypal checkout gateway by woocommerce – update to version is available – installed version not tested with active version of woocommerce woocommerce paypal powered by braintree gateway by woocommerce – update to version is available – installed version not tested with active version of woocommerce woocommerce print invoices packing lists by skyverge – – installed version not tested with active version of woocommerce woocommerce product add ons by woocommerce – – installed version not tested with active version of woocommerce woocommerce sequential order numbers pro by skyverge – – installed version not tested with active version of woocommerce woocommerce services by automattic – update to version is available – installed version not tested with active version of woocommerce woocommerce stamps com api integration by woocommerce – – installed version not tested with active version of woocommerce woocommerce stamps com export suite by skyverge – – installed version not tested with active version of woocommerce woocommerce store catalog pdf download by woocommerce – – installed version not tested with active version of woocommerce woocommerce stripe gateway by woocommerce – – installed version not tested with active version of woocommerce woocommerce subscriptions by woocommerce – – installed version not tested with active version of woocommerce woocommerce ups shipping by woocommerce – – installed version not tested with active version of woocommerce woocommerce wholesale prices by rymera web co – – installed version not tested with active version of woocommerce wp crontrol by john blackbourn crontributors – wp migrate db by delicious brains – wp security audit log by wp white security – dropin plugins advanced cache php advanced cache php db php query monitor database class object cache php memcached settings api enabled ✔ force ssl ✔ currency usd currency position left thousand separator decimal separator number of decimals taxonomies product types external external grouped grouped simple simple subscription subscription variable variable variable subscription variable subscription taxonomies product visibility exclude from catalog exclude from catalog exclude from search exclude from search featured featured outofstock outofstock rated rated rated rated rated rated rated rated rated rated connected to woocommerce com – wc pages shop base shop cart cart checkout checkout my account my account terms and conditions ❌ page not set theme name storefront version author url child theme ❌ – if you are modifying woocommerce on a parent theme that you did not build personally we recommend using a child theme see how to create a child theme woocommerce support ✔ templates overrides – action scheduler complete oldest newest pending oldest newest
0
64,105
6,892,926,140
IssuesEvent
2017-11-22 23:38:00
null511/PiServerLite
https://api.github.com/repos/null511/PiServerLite
closed
HTTPS Support
enhancement Testing
Need an option to enable strict use of HTTPS. - When enabled, automatically redirect **all** incoming _HTTP_ requests to _HTTPS_. - Support non-standard port settings. ie, **!**_443_
1.0
HTTPS Support - Need an option to enable strict use of HTTPS. - When enabled, automatically redirect **all** incoming _HTTP_ requests to _HTTPS_. - Support non-standard port settings. ie, **!**_443_
non_process
https support need an option to enable strict use of https when enabled automatically redirect all incoming http requests to https support non standard port settings ie
0
10,291
13,145,268,188
IssuesEvent
2020-08-08 02:30:44
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Add_locale processor not refreshing timezone info.
:Processors Stalled enhancement libbeat needs_team
- Version: 6.3.1 - Operating System: Win 10 x64 (maybe all?) - Discuss Forum URL: [https://discuss.elastic.co/t/add-locale-processor-not-refreshing-timezone-info/141397](https://discuss.elastic.co/t/add-locale-processor-not-refreshing-timezone-info/141397) - Steps to Reproduce: - Add locale processor - Start a filebeat - Change system timezone - Events will be emitted with original timezone until beat restart From Discuss Forum: > Recently added the add_locale processor to a Filebeat instance and its adding the current timezone offset to logs correctly. However, if I change the timezone without restarting Filebeat, the timezone offset reported is the old one. This appears to happen with Winlogbeat as well. For timezone changes to take effect in beats, does a beat need to be restarted or is something bugged in the way the timezone is determined? > I could see this being a problem with long running beats when a timezone change to standard time to daylight savings.
1.0
Add_locale processor not refreshing timezone info. - - Version: 6.3.1 - Operating System: Win 10 x64 (maybe all?) - Discuss Forum URL: [https://discuss.elastic.co/t/add-locale-processor-not-refreshing-timezone-info/141397](https://discuss.elastic.co/t/add-locale-processor-not-refreshing-timezone-info/141397) - Steps to Reproduce: - Add locale processor - Start a filebeat - Change system timezone - Events will be emitted with original timezone until beat restart From Discuss Forum: > Recently added the add_locale processor to a Filebeat instance and its adding the current timezone offset to logs correctly. However, if I change the timezone without restarting Filebeat, the timezone offset reported is the old one. This appears to happen with Winlogbeat as well. For timezone changes to take effect in beats, does a beat need to be restarted or is something bugged in the way the timezone is determined? > I could see this being a problem with long running beats when a timezone change to standard time to daylight savings.
process
add locale processor not refreshing timezone info version operating system win maybe all discuss forum url steps to reproduce add locale processor start a filebeat change system timezone events will be emitted with original timezone until beat restart from discuss forum recently added the add locale processor to a filebeat instance and its adding the current timezone offset to logs correctly however if i change the timezone without restarting filebeat the timezone offset reported is the old one this appears to happen with winlogbeat as well for timezone changes to take effect in beats does a beat need to be restarted or is something bugged in the way the timezone is determined i could see this being a problem with long running beats when a timezone change to standard time to daylight savings
1
19,191
25,318,760,776
IssuesEvent
2022-11-18 00:44:49
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
ANALISTA DE SISTEMAS SR na [MAXIFROTA]
SALVADOR BACK-END DESENVOLVIMENTO DE SOFTWARE GESTÃO DE PROJETOS JAVA SENIOR SQL MOBILE REQUISITOS GRAILS GROOVY PROCESSOS INOVAÇÃO Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## ANALISTA DE SISTEMAS SR - Realizar levantamento de requisitos e especificações para testes e melhorias dos sistemas. - Desenvolver sistemas e relatórios. - Aplicar conhecimentos tecnicos, utilizando novas ferramentas e tecnologias para melhorias nos processos da area. - Realizar consultas no banco de dados e disponibilizar para os usuarios. Verificar o correto funcionamento das rotinas dos sistemas e aplicacoes da empresa reportando as areas responsaveis. - Elaborar e realizar levantamentos sobre informacoes e dados, para estudo e implantacao de sistemas. - Avaliar e identificar necessidades de soluções tecnologicas, elaborando planejamento de projetos e entendimento das necessidades do negOcio e dos clientes. - Dar suporte a usuarios de sistemas. - Dar apoio a supervisao e fazer levantamento de indicadores da area. ## Local Salvador - Bahia ## Benefícios - Assistência médica - Assistência odontologica - Vale-alimentacao - Vale-refeicao - Vale-transporte ## Requisitos **Obrigatórios:** - Experiencia minima de 03 anos com suporte a usuario e desenvolvimento de sistemas em ambiente WEB, com conhecimento em ferramentas mobile - Escolaridade Ensino Superior em Tecnologia da Informacao, lnformatica, Ciencias da Computacao ou areas afins. - Java - Grails - Groovy - SQL **Desejáveis:** - Desejável Pós Graduacao na area de atuacao - Experiencia em banco de dados - Lideranca técnica de times de desenvolvimento ## MAXIFROTA Através de produtos que aliam meios de pagamentos e sistemas de gestão, oferecemos uma solução integrada e inteligente que atua diretamente sobre os custos com o abastecimento e a manutenção de veículos, principais despesas e desafios para gerenciamento de frotas. ## Gente MaxiFrota Caso tenha interesse em fazer parte da nossa equipe, preencha o formulário abaixo. Seu currículo será encaminhado para o setor responsável e, caso existam vagas com o seu perfil, entraremos em contato https://www.maxifrota.com.br/trabalhe-conosco
1.0
ANALISTA DE SISTEMAS SR na [MAXIFROTA] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## ANALISTA DE SISTEMAS SR - Realizar levantamento de requisitos e especificações para testes e melhorias dos sistemas. - Desenvolver sistemas e relatórios. - Aplicar conhecimentos tecnicos, utilizando novas ferramentas e tecnologias para melhorias nos processos da area. - Realizar consultas no banco de dados e disponibilizar para os usuarios. Verificar o correto funcionamento das rotinas dos sistemas e aplicacoes da empresa reportando as areas responsaveis. - Elaborar e realizar levantamentos sobre informacoes e dados, para estudo e implantacao de sistemas. - Avaliar e identificar necessidades de soluções tecnologicas, elaborando planejamento de projetos e entendimento das necessidades do negOcio e dos clientes. - Dar suporte a usuarios de sistemas. - Dar apoio a supervisao e fazer levantamento de indicadores da area. ## Local Salvador - Bahia ## Benefícios - Assistência médica - Assistência odontologica - Vale-alimentacao - Vale-refeicao - Vale-transporte ## Requisitos **Obrigatórios:** - Experiencia minima de 03 anos com suporte a usuario e desenvolvimento de sistemas em ambiente WEB, com conhecimento em ferramentas mobile - Escolaridade Ensino Superior em Tecnologia da Informacao, lnformatica, Ciencias da Computacao ou areas afins. - Java - Grails - Groovy - SQL **Desejáveis:** - Desejável Pós Graduacao na area de atuacao - Experiencia em banco de dados - Lideranca técnica de times de desenvolvimento ## MAXIFROTA Através de produtos que aliam meios de pagamentos e sistemas de gestão, oferecemos uma solução integrada e inteligente que atua diretamente sobre os custos com o abastecimento e a manutenção de veículos, principais despesas e desafios para gerenciamento de frotas. ## Gente MaxiFrota Caso tenha interesse em fazer parte da nossa equipe, preencha o formulário abaixo. Seu currículo será encaminhado para o setor responsável e, caso existam vagas com o seu perfil, entraremos em contato https://www.maxifrota.com.br/trabalhe-conosco
process
analista de sistemas sr na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na analista de sistemas sr realizar levantamento de requisitos e especificações para testes e melhorias dos sistemas desenvolver sistemas e relatórios aplicar conhecimentos tecnicos utilizando novas ferramentas e tecnologias para melhorias nos processos da area realizar consultas no banco de dados e disponibilizar para os usuarios verificar o correto funcionamento das rotinas dos sistemas e aplicacoes da empresa reportando as areas responsaveis elaborar e realizar levantamentos sobre informacoes e dados para estudo e implantacao de sistemas avaliar e identificar necessidades de soluções tecnologicas elaborando planejamento de projetos e entendimento das necessidades do negocio e dos clientes dar suporte a usuarios de sistemas dar apoio a supervisao e fazer levantamento de indicadores da area local salvador bahia benefícios assistência médica assistência odontologica vale alimentacao vale refeicao vale transporte requisitos obrigatórios experiencia minima de anos com suporte a usuario e desenvolvimento de sistemas em ambiente web com conhecimento em ferramentas mobile escolaridade ensino superior em tecnologia da informacao lnformatica ciencias da computacao ou areas afins java grails groovy sql desejáveis desejável pós graduacao na area de atuacao experiencia em banco de dados lideranca técnica de times de desenvolvimento maxifrota através de produtos que aliam meios de pagamentos e sistemas de gestão oferecemos uma solução integrada e inteligente que atua diretamente sobre os custos com o abastecimento e a manutenção de veículos principais despesas e desafios para gerenciamento de frotas gente maxifrota caso tenha interesse em fazer parte da nossa equipe preencha o formulário abaixo seu currículo será encaminhado para o setor responsável e caso existam vagas com o seu perfil entraremos em contato
1
2,998
5,988,167,687
IssuesEvent
2017-06-02 03:11:41
P0cL4bs/WiFi-Pumpkin
https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin
closed
Error bs4 python module
in process priority solved
Hello, this is error: ~# wifi-pumpkin Traceback (most recent call last): File "wifi-pumpkin.py", line 39, in <module> from core.loaders.checker.networkmanager import CLI_NetworkManager,UI_NetworkManager File "/usr/share/WiFi-Pumpkin/core/loaders/checker/networkmanager.py", line 4, in <module> from core.main import Initialize File "/usr/share/WiFi-Pumpkin/core/main.py", line 32, in <module> from core.widgets.tabmodels import ( File "/usr/share/WiFi-Pumpkin/core/widgets/tabmodels.py", line 7, in <module> from core.utility.threads import ThreadPopen File "/usr/share/WiFi-Pumpkin/core/utility/threads.py", line 19, in <module> from core.servers.proxy.http.controller.handler import MasterHandler File "/usr/share/WiFi-Pumpkin/core/servers/proxy/http/controller/handler.py", line 1, in <module> from plugins.extension import * File "/usr/share/WiFi-Pumpkin/plugins/extension/html_inject.py", line 2, in <module> from bs4 import BeautifulSoup File "build/bdist.linux-x86_64/egg/bs4/__init__.py", line 30, in <module> File "build/bdist.linux-x86_64/egg/bs4/builder/__init__.py", line 314, in <module> File "build/bdist.linux-x86_64/egg/bs4/builder/_html5lib.py", line 70, in <module> AttributeError: 'module' object has no attribute '_base' I have update html5lib.py but the error is the same. Please help me Thanks Network adapter TpLink wn722n no vmachine Kali linux rolling wifi pumpkin latest version
1.0
Error bs4 python module - Hello, this is error: ~# wifi-pumpkin Traceback (most recent call last): File "wifi-pumpkin.py", line 39, in <module> from core.loaders.checker.networkmanager import CLI_NetworkManager,UI_NetworkManager File "/usr/share/WiFi-Pumpkin/core/loaders/checker/networkmanager.py", line 4, in <module> from core.main import Initialize File "/usr/share/WiFi-Pumpkin/core/main.py", line 32, in <module> from core.widgets.tabmodels import ( File "/usr/share/WiFi-Pumpkin/core/widgets/tabmodels.py", line 7, in <module> from core.utility.threads import ThreadPopen File "/usr/share/WiFi-Pumpkin/core/utility/threads.py", line 19, in <module> from core.servers.proxy.http.controller.handler import MasterHandler File "/usr/share/WiFi-Pumpkin/core/servers/proxy/http/controller/handler.py", line 1, in <module> from plugins.extension import * File "/usr/share/WiFi-Pumpkin/plugins/extension/html_inject.py", line 2, in <module> from bs4 import BeautifulSoup File "build/bdist.linux-x86_64/egg/bs4/__init__.py", line 30, in <module> File "build/bdist.linux-x86_64/egg/bs4/builder/__init__.py", line 314, in <module> File "build/bdist.linux-x86_64/egg/bs4/builder/_html5lib.py", line 70, in <module> AttributeError: 'module' object has no attribute '_base' I have update html5lib.py but the error is the same. Please help me Thanks Network adapter TpLink wn722n no vmachine Kali linux rolling wifi pumpkin latest version
process
error python module hello this is error wifi pumpkin traceback most recent call last file wifi pumpkin py line in from core loaders checker networkmanager import cli networkmanager ui networkmanager file usr share wifi pumpkin core loaders checker networkmanager py line in from core main import initialize file usr share wifi pumpkin core main py line in from core widgets tabmodels import file usr share wifi pumpkin core widgets tabmodels py line in from core utility threads import threadpopen file usr share wifi pumpkin core utility threads py line in from core servers proxy http controller handler import masterhandler file usr share wifi pumpkin core servers proxy http controller handler py line in from plugins extension import file usr share wifi pumpkin plugins extension html inject py line in from import beautifulsoup file build bdist linux egg init py line in file build bdist linux egg builder init py line in file build bdist linux egg builder py line in attributeerror module object has no attribute base i have update py but the error is the same please help me thanks network adapter tplink no vmachine kali linux rolling wifi pumpkin latest version
1
106,686
4,282,380,749
IssuesEvent
2016-07-15 08:55:02
opencv/opencv
https://api.github.com/repos/opencv/opencv
reopened
BFMatcher SigSegv (OpenCL implementation)
affected: master bug category: t-api priority: normal
My application very rarely segfaults around BFMatching.knnMatch (OpenCL via transparent API). After investigation I created sample code that also segfaults: ```cpp #include <opencv2/features2d.hpp> #include <opencv2/imgcodecs.hpp> #include <opencv2/opencv.hpp> #include <opencv2/core/ocl.hpp> #include <vector> #include <iostream> #include <stdio.h> #include <tuple> using namespace std; using namespace cv; void print_ocl_device_name() { vector<ocl::PlatformInfo> platforms; ocl::getPlatfomsInfo(platforms); for (size_t i = 0; i < platforms.size(); i++) { const cv::ocl::PlatformInfo* platform = &platforms[i]; cout << "Platform Name: " << platform->name().c_str() << "\n"; for (int j = 0; j < platform->deviceNumber(); j++) { cv::ocl::Device current_device; platform->getDevice(current_device, j); int deviceType = current_device.type(); cout << "Device " << j << ": " << current_device.name() << "\n"; } } ocl::Device default_device = ocl::Device::getDefault(); cout << "Used device: " << default_device.name() << "\n"; } int main(void) { print_ocl_device_name(); RNG r(239); Mat desc1(5100, 64, CV_32F); Mat desc2(5046, 64, CV_32F); r.fill(desc1, RNG::UNIFORM, Scalar::all(0.0), Scalar::all(1.0)); r.fill(desc2, RNG::UNIFORM, Scalar::all(0.0), Scalar::all(1.0)); for (int i = 0; i < 1000 * 1000; i++) { BFMatcher matcher(NORM_L2); UMat udesc1, udesc2; desc1.copyTo(udesc1); desc2.copyTo(udesc2); vector< vector<DMatch> > nn_matches; matcher.knnMatch(udesc1, udesc2, nn_matches, 2); printf("%d\n", i); } return 0; } ``` **Note**: this happens very rare. I see it after tens of thousands of iterations (tens of minutes on fast GPU). ## Environment OpenCV version: **3.0.0** I have segfaults on Nvidia GTX 680 (**OpenCL 1.1**): ``` Device name: GeForce GTX 680 Nvidia driver verison: 331.113 Platform Version: OpenCL 1.1 CUDA 6.0.1 ``` The same behaivour on Nvidia GTX 580 with 352.30 driver (**OpenCL 1.1**). On CPU device (not CPU implementation, here I mean CPU as OpenCL device) I was trying - but for now there is no segfaults. But I think this is only because of speed difference between CPU and GPU. (On GPU it takes tens of minutes or even more - CPU slower up to to one or two orders). Or these can be explained driver implementation details, or the reason, that CPU RAM is the same that the Host RAM. **GDB backtraces**: ``` Thread 7 (Thread 0x7f64ad44a700 (LWP 2212)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64ad44a700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 6 (Thread 0x7f64acc49700 (LWP 2213)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64acc49700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 5 (Thread 0x7f64adc4b700 (LWP 2211)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64adc4b700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 4 (Thread 0x7f64b6209780 (LWP 2208)): #0 0x00007f64b48cc774 in __GI___libc_malloc (bytes=32) at malloc.c:2903 #1 0x00007f64b4e84688 in operator new(unsigned long) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #2 0x00007f64b5d8b42e in allocate (this=0x26997e0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/ext/new_allocator.h:104 #3 allocate (__a=..., __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/ext/alloc_traits.h:182 #4 _M_allocate (this=0x26997e0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/bits/stl_vector.h:170 #5 _M_allocate_and_copy<cv::DMatch*> (this=0x26997e0, __last=0x0, __first=0x0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/bits/stl_vector.h:1224 #6 std::vector<cv::DMatch, std::allocator<cv::DMatch> >::reserve (this=this@entry=0x26997e0, __n=__n@entry=2) at /usr/include/c++/4.9/bits/vector.tcc:75 #7 0x00007f64b5d913b0 in ocl_knnMatchConvert (compactResult=<optimized out>, matches=std::vector of length 4825, capacity 5100 = {...}, distance=..., trainIdx=...) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:240 #8 ocl_knnMatchDownload (compactResult=<optimized out>, matches=std::vector of length 4825, capacity 5100 = {...}, distance=..., trainIdx=...) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:270 #9 cv::ocl_knnMatch (query=..., _train=..., matches=std::vector of length 4825, capacity 5100 = {...}, k=k@entry=2, dstType=<optimized out>, compactResult=compactResult@entry=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:713 #10 0x00007f64b5d938ac in cv::BFMatcher::knnMatchImpl (this=0x262b720, _queryDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=2, _masks=..., compactResult=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:786 #11 0x00007f64b5d8bade in cv::DescriptorMatcher::knnMatch (this=0x262b720, queryDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=knn@entry=2, masks=..., compactResult=compactResult@entry=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:630 #12 0x00007f64b5d8c68c in cv::DescriptorMatcher::knnMatch (this=<optimized out>, queryDescriptors=..., trainDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=2, mask=..., compactResult=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:577 #13 0x0000000000401c13 in main () at bfmatcher_segfault.cpp:50 Thread 3 (Thread 0x7f64af9ee700 (LWP 2209)): #0 0x00007f64b493984d in poll () at ../sysdeps/unix/syscall-template.S:81 #1 0x00007f64b1785373 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b1135f15 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64af9ee700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 2 (Thread 0x7f64ae44c700 (LWP 2210)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64ae44c700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 1 (Thread 0x7f64a7fff700 (LWP 2214)): #0 0x00007f64b0ffdae2 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #1 0x00007f64b0ffe514 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b2dad0a5 in start_thread (arg=0x7f64a7fff700) at pthread_create.c:309 #4 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 ``` **Valgrind** can't help because of too big slow down. **AddressSanitizer** just fails on startup with SigSegv, this seems to be bad combination of ASan and OpenCL. ## Workaround It seems, that explicit sync needed after calculating and before result downloading. If I change `return k.run(2, globalSize, localSize, false);` to `return k.run(2, globalSize, localSize, true);`, i.e. pass sync=true - I stops encounter SigSegv. This fix must be applied in: - ocl_matchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L114 - ocl_knnMatchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L214 - ocl_radiusMatchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L327 ## Results clEnqueueMapBuffer fetches data in concurrent with unfinished kernel, and this is **undefined behaviour**. May be `sync=false` is just randomly passed param? While line at https://github.com/Itseez/opencv/blob/3.0.0/modules/core/src/ocl.cpp#L3382 should change `sync` to `true`? But this is not so - I checked, this `if` branch does not executes in my scenario. The reason is in https://github.com/Itseez/opencv/blob/3.0.0/modules/core/src/ocl.cpp#L3119 - `m.u->tempUMat()` fails that if-branch. Because UMatData.flags=3 for all Kernel.addUMat(UMat) calls, while TEMP_UMAT=8. What is TEMP_UMAT? Are those umats should be tempUMat? And even if so - all three Kernel.run in matchers.cpp should be called with sync=true, because there are no async behaviour. It seems that such bug with OpenCL methods result fetching can be in a lot of places, can you check them please? It is quite dangerous bug. (I checked LKOptFlow - there Kernel.run called with sync=true) Also, there are no bugfixes bug-porting to 3.0.0 version, because there is no 3.0.0 stable branch. Why so? 3.0.0 branch like 2.4 branch seems to be useful.
1.0
BFMatcher SigSegv (OpenCL implementation) - My application very rarely segfaults around BFMatching.knnMatch (OpenCL via transparent API). After investigation I created sample code that also segfaults: ```cpp #include <opencv2/features2d.hpp> #include <opencv2/imgcodecs.hpp> #include <opencv2/opencv.hpp> #include <opencv2/core/ocl.hpp> #include <vector> #include <iostream> #include <stdio.h> #include <tuple> using namespace std; using namespace cv; void print_ocl_device_name() { vector<ocl::PlatformInfo> platforms; ocl::getPlatfomsInfo(platforms); for (size_t i = 0; i < platforms.size(); i++) { const cv::ocl::PlatformInfo* platform = &platforms[i]; cout << "Platform Name: " << platform->name().c_str() << "\n"; for (int j = 0; j < platform->deviceNumber(); j++) { cv::ocl::Device current_device; platform->getDevice(current_device, j); int deviceType = current_device.type(); cout << "Device " << j << ": " << current_device.name() << "\n"; } } ocl::Device default_device = ocl::Device::getDefault(); cout << "Used device: " << default_device.name() << "\n"; } int main(void) { print_ocl_device_name(); RNG r(239); Mat desc1(5100, 64, CV_32F); Mat desc2(5046, 64, CV_32F); r.fill(desc1, RNG::UNIFORM, Scalar::all(0.0), Scalar::all(1.0)); r.fill(desc2, RNG::UNIFORM, Scalar::all(0.0), Scalar::all(1.0)); for (int i = 0; i < 1000 * 1000; i++) { BFMatcher matcher(NORM_L2); UMat udesc1, udesc2; desc1.copyTo(udesc1); desc2.copyTo(udesc2); vector< vector<DMatch> > nn_matches; matcher.knnMatch(udesc1, udesc2, nn_matches, 2); printf("%d\n", i); } return 0; } ``` **Note**: this happens very rare. I see it after tens of thousands of iterations (tens of minutes on fast GPU). ## Environment OpenCV version: **3.0.0** I have segfaults on Nvidia GTX 680 (**OpenCL 1.1**): ``` Device name: GeForce GTX 680 Nvidia driver verison: 331.113 Platform Version: OpenCL 1.1 CUDA 6.0.1 ``` The same behaivour on Nvidia GTX 580 with 352.30 driver (**OpenCL 1.1**). On CPU device (not CPU implementation, here I mean CPU as OpenCL device) I was trying - but for now there is no segfaults. But I think this is only because of speed difference between CPU and GPU. (On GPU it takes tens of minutes or even more - CPU slower up to to one or two orders). Or these can be explained driver implementation details, or the reason, that CPU RAM is the same that the Host RAM. **GDB backtraces**: ``` Thread 7 (Thread 0x7f64ad44a700 (LWP 2212)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64ad44a700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 6 (Thread 0x7f64acc49700 (LWP 2213)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64acc49700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 5 (Thread 0x7f64adc4b700 (LWP 2211)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64adc4b700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 4 (Thread 0x7f64b6209780 (LWP 2208)): #0 0x00007f64b48cc774 in __GI___libc_malloc (bytes=32) at malloc.c:2903 #1 0x00007f64b4e84688 in operator new(unsigned long) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #2 0x00007f64b5d8b42e in allocate (this=0x26997e0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/ext/new_allocator.h:104 #3 allocate (__a=..., __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/ext/alloc_traits.h:182 #4 _M_allocate (this=0x26997e0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/bits/stl_vector.h:170 #5 _M_allocate_and_copy<cv::DMatch*> (this=0x26997e0, __last=0x0, __first=0x0, __n=<error reading variable: Cannot access memory at address 0xfffffffffffffcc8>) at /usr/include/c++/4.9/bits/stl_vector.h:1224 #6 std::vector<cv::DMatch, std::allocator<cv::DMatch> >::reserve (this=this@entry=0x26997e0, __n=__n@entry=2) at /usr/include/c++/4.9/bits/vector.tcc:75 #7 0x00007f64b5d913b0 in ocl_knnMatchConvert (compactResult=<optimized out>, matches=std::vector of length 4825, capacity 5100 = {...}, distance=..., trainIdx=...) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:240 #8 ocl_knnMatchDownload (compactResult=<optimized out>, matches=std::vector of length 4825, capacity 5100 = {...}, distance=..., trainIdx=...) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:270 #9 cv::ocl_knnMatch (query=..., _train=..., matches=std::vector of length 4825, capacity 5100 = {...}, k=k@entry=2, dstType=<optimized out>, compactResult=compactResult@entry=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:713 #10 0x00007f64b5d938ac in cv::BFMatcher::knnMatchImpl (this=0x262b720, _queryDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=2, _masks=..., compactResult=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:786 #11 0x00007f64b5d8bade in cv::DescriptorMatcher::knnMatch (this=0x262b720, queryDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=knn@entry=2, masks=..., compactResult=compactResult@entry=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:630 #12 0x00007f64b5d8c68c in cv::DescriptorMatcher::knnMatch (this=<optimized out>, queryDescriptors=..., trainDescriptors=..., matches=std::vector of length 4825, capacity 5100 = {...}, knn=2, mask=..., compactResult=false) at /home/polarnick/libraries/opencv/opencv_3.0.0/modules/features2d/src/matchers.cpp:577 #13 0x0000000000401c13 in main () at bfmatcher_segfault.cpp:50 Thread 3 (Thread 0x7f64af9ee700 (LWP 2209)): #0 0x00007f64b493984d in poll () at ../sysdeps/unix/syscall-template.S:81 #1 0x00007f64b1785373 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b1135f15 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64af9ee700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 2 (Thread 0x7f64ae44c700 (LWP 2210)): #0 sem_wait () at ../nptl/sysdeps/unix/sysv/linux/x86_64/sem_wait.S:85 #1 0x00007f64b1784a7e in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b0ffe548 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #4 0x00007f64b2dad0a5 in start_thread (arg=0x7f64ae44c700) at pthread_create.c:309 #5 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 Thread 1 (Thread 0x7f64a7fff700 (LWP 2214)): #0 0x00007f64b0ffdae2 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #1 0x00007f64b0ffe514 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #2 0x00007f64b17872d9 in ?? () from /usr/lib/x86_64-linux-gnu/libnvidia-opencl.so.1 #3 0x00007f64b2dad0a5 in start_thread (arg=0x7f64a7fff700) at pthread_create.c:309 #4 0x00007f64b4943cfd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111 ``` **Valgrind** can't help because of too big slow down. **AddressSanitizer** just fails on startup with SigSegv, this seems to be bad combination of ASan and OpenCL. ## Workaround It seems, that explicit sync needed after calculating and before result downloading. If I change `return k.run(2, globalSize, localSize, false);` to `return k.run(2, globalSize, localSize, true);`, i.e. pass sync=true - I stops encounter SigSegv. This fix must be applied in: - ocl_matchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L114 - ocl_knnMatchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L214 - ocl_radiusMatchSingle: https://github.com/Itseez/opencv/blob/3.0.0/modules/features2d/src/matchers.cpp#L327 ## Results clEnqueueMapBuffer fetches data in concurrent with unfinished kernel, and this is **undefined behaviour**. May be `sync=false` is just randomly passed param? While line at https://github.com/Itseez/opencv/blob/3.0.0/modules/core/src/ocl.cpp#L3382 should change `sync` to `true`? But this is not so - I checked, this `if` branch does not executes in my scenario. The reason is in https://github.com/Itseez/opencv/blob/3.0.0/modules/core/src/ocl.cpp#L3119 - `m.u->tempUMat()` fails that if-branch. Because UMatData.flags=3 for all Kernel.addUMat(UMat) calls, while TEMP_UMAT=8. What is TEMP_UMAT? Are those umats should be tempUMat? And even if so - all three Kernel.run in matchers.cpp should be called with sync=true, because there are no async behaviour. It seems that such bug with OpenCL methods result fetching can be in a lot of places, can you check them please? It is quite dangerous bug. (I checked LKOptFlow - there Kernel.run called with sync=true) Also, there are no bugfixes bug-porting to 3.0.0 version, because there is no 3.0.0 stable branch. Why so? 3.0.0 branch like 2.4 branch seems to be useful.
non_process
bfmatcher sigsegv opencl implementation my application very rarely segfaults around bfmatching knnmatch opencl via transparent api after investigation i created sample code that also segfaults cpp include include include include include include include include using namespace std using namespace cv void print ocl device name vector platforms ocl getplatfomsinfo platforms for size t i i platforms size i const cv ocl platforminfo platform platforms cout name c str n for int j j devicenumber j cv ocl device current device platform getdevice current device j int devicetype current device type cout device j current device name n ocl device default device ocl device getdefault cout used device default device name n int main void print ocl device name rng r mat cv mat cv r fill rng uniform scalar all scalar all r fill rng uniform scalar all scalar all for int i i i bfmatcher matcher norm umat copyto copyto vector nn matches matcher knnmatch nn matches printf d n i return note this happens very rare i see it after tens of thousands of iterations tens of minutes on fast gpu environment opencv version i have segfaults on nvidia gtx opencl device name geforce gtx nvidia driver verison platform version opencl cuda the same behaivour on nvidia gtx with driver opencl on cpu device not cpu implementation here i mean cpu as opencl device i was trying but for now there is no segfaults but i think this is only because of speed difference between cpu and gpu on gpu it takes tens of minutes or even more cpu slower up to to one or two orders or these can be explained driver implementation details or the reason that cpu ram is the same that the host ram gdb backtraces thread thread lwp sem wait at nptl sysdeps unix sysv linux sem wait s in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s thread thread lwp sem wait at nptl sysdeps unix sysv linux sem wait s in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s thread thread lwp sem wait at nptl sysdeps unix sysv linux sem wait s in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s thread thread lwp in gi libc malloc bytes at malloc c in operator new unsigned long from usr lib linux gnu libstdc so in allocate this n at usr include c ext new allocator h allocate a n at usr include c ext alloc traits h m allocate this n at usr include c bits stl vector h m allocate and copy this last first n at usr include c bits stl vector h std vector reserve this this entry n n entry at usr include c bits vector tcc in ocl knnmatchconvert compactresult matches std vector of length capacity distance trainidx at home polarnick libraries opencv opencv modules src matchers cpp ocl knnmatchdownload compactresult matches std vector of length capacity distance trainidx at home polarnick libraries opencv opencv modules src matchers cpp cv ocl knnmatch query train matches std vector of length capacity k k entry dsttype compactresult compactresult entry false at home polarnick libraries opencv opencv modules src matchers cpp in cv bfmatcher knnmatchimpl this querydescriptors matches std vector of length capacity knn masks compactresult false at home polarnick libraries opencv opencv modules src matchers cpp in cv descriptormatcher knnmatch this querydescriptors matches std vector of length capacity knn knn entry masks compactresult compactresult entry false at home polarnick libraries opencv opencv modules src matchers cpp in cv descriptormatcher knnmatch this querydescriptors traindescriptors matches std vector of length capacity knn mask compactresult false at home polarnick libraries opencv opencv modules src matchers cpp in main at bfmatcher segfault cpp thread thread lwp in poll at sysdeps unix syscall template s in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s thread thread lwp sem wait at nptl sysdeps unix sysv linux sem wait s in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s thread thread lwp in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in from usr lib linux gnu libnvidia opencl so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s valgrind can t help because of too big slow down addresssanitizer just fails on startup with sigsegv this seems to be bad combination of asan and opencl workaround it seems that explicit sync needed after calculating and before result downloading if i change return k run globalsize localsize false to return k run globalsize localsize true i e pass sync true i stops encounter sigsegv this fix must be applied in ocl matchsingle ocl knnmatchsingle ocl radiusmatchsingle results clenqueuemapbuffer fetches data in concurrent with unfinished kernel and this is undefined behaviour may be sync false is just randomly passed param while line at should change sync to true but this is not so i checked this if branch does not executes in my scenario the reason is in m u tempumat fails that if branch because umatdata flags for all kernel addumat umat calls while temp umat what is temp umat are those umats should be tempumat and even if so all three kernel run in matchers cpp should be called with sync true because there are no async behaviour it seems that such bug with opencl methods result fetching can be in a lot of places can you check them please it is quite dangerous bug i checked lkoptflow there kernel run called with sync true also there are no bugfixes bug porting to version because there is no stable branch why so branch like branch seems to be useful
0
129,552
10,577,922,524
IssuesEvent
2019-10-07 21:13:43
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jepsen/monotonic/strobe-skews failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/8d3bdd75d0080cdb089b65d55f4e5fcf4a3ea905 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen/monotonic/strobe-skews PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1524120&tab=artifacts#/jepsen/monotonic/strobe-skews ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191007-1524120/jepsen/monotonic/strobe-skews/run_1 jepsen.go:262,jepsen.go:320,test_runner.go:689: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1570424814-47-n6cpu4:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.128.0.36 -n 10.128.0.35 -n 10.128.0.34 -n 10.128.0.32 -n 10.128.0.30 \ --test monotonic --nemesis strobe-skews \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 1 : exit status 1 ```
2.0
roachtest: jepsen/monotonic/strobe-skews failed - SHA: https://github.com/cockroachdb/cockroach/commits/8d3bdd75d0080cdb089b65d55f4e5fcf4a3ea905 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen/monotonic/strobe-skews PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1524120&tab=artifacts#/jepsen/monotonic/strobe-skews ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191007-1524120/jepsen/monotonic/strobe-skews/run_1 jepsen.go:262,jepsen.go:320,test_runner.go:689: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1570424814-47-n6cpu4:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.128.0.36 -n 10.128.0.35 -n 10.128.0.34 -n 10.128.0.32 -n 10.128.0.30 \ --test monotonic --nemesis strobe-skews \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 1 : exit status 1 ```
non_process
roachtest jepsen monotonic strobe skews failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen monotonic strobe skews pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts jepsen monotonic strobe skews run jepsen go jepsen go test runner go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test monotonic nemesis strobe skews invoke log returned stderr stdout error exit status exit status
0
49,632
6,035,589,600
IssuesEvent
2017-06-09 14:14:53
IronLanguages/ironpython2
https://api.github.com/repos/IronLanguages/ironpython2
opened
Re-enable test_expandvars_nonascii in test_ntpath and test_genericpath
mono Test Failure
_From @slide on April 9, 2017 4:46_ Tests are failing on Mono for different encoding output. _Copied from original issue: IronLanguages/main#1621_
1.0
Re-enable test_expandvars_nonascii in test_ntpath and test_genericpath - _From @slide on April 9, 2017 4:46_ Tests are failing on Mono for different encoding output. _Copied from original issue: IronLanguages/main#1621_
non_process
re enable test expandvars nonascii in test ntpath and test genericpath from slide on april tests are failing on mono for different encoding output copied from original issue ironlanguages main
0
6,563
9,648,888,266
IssuesEvent
2019-05-17 17:32:02
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Bug: GPA doesn't come over from USAJOBS to Open Opps on Education row
Apply Process Bug State Dept.
Environment: UAT Steps to Reproduce: 1) entered information into USAJOBS UAT profile including an education row with a GPA 2) Apply to an internship in Open Opportunities - Education row comes in but the GPA is blank <img width="1440" alt="Screen Shot 2019-05-02 at 1 32 42 PM" src="https://user-images.githubusercontent.com/18709918/57094757-4acd8300-6cdf-11e9-82db-31ef1fcb6c16.png"> <img width="1440" alt="Screen Shot 2019-05-02 at 1 31 17 PM" src="https://user-images.githubusercontent.com/18709918/57094763-4dc87380-6cdf-11e9-8da6-1907eebba580.png">
1.0
Bug: GPA doesn't come over from USAJOBS to Open Opps on Education row - Environment: UAT Steps to Reproduce: 1) entered information into USAJOBS UAT profile including an education row with a GPA 2) Apply to an internship in Open Opportunities - Education row comes in but the GPA is blank <img width="1440" alt="Screen Shot 2019-05-02 at 1 32 42 PM" src="https://user-images.githubusercontent.com/18709918/57094757-4acd8300-6cdf-11e9-82db-31ef1fcb6c16.png"> <img width="1440" alt="Screen Shot 2019-05-02 at 1 31 17 PM" src="https://user-images.githubusercontent.com/18709918/57094763-4dc87380-6cdf-11e9-8da6-1907eebba580.png">
process
bug gpa doesn t come over from usajobs to open opps on education row environment uat steps to reproduce entered information into usajobs uat profile including an education row with a gpa apply to an internship in open opportunities education row comes in but the gpa is blank img width alt screen shot at pm src img width alt screen shot at pm src
1
307,758
26,560,845,848
IssuesEvent
2023-01-20 15:44:39
numaproj/numaflow
https://api.github.com/repos/numaproj/numaflow
opened
Add Source Data Transformer E2E test
enhancement testing area/source
# Summary What change needs making? # Use Cases When would you use this? --- <!-- Issue Author: Don't delete this message to encourage other users to support your issue! --> **Message from the maintainers**: If you wish to see this enhancement implemented please add a 👍 reaction to this issue! We often sort issues this way to know what to prioritize.
1.0
Add Source Data Transformer E2E test - # Summary What change needs making? # Use Cases When would you use this? --- <!-- Issue Author: Don't delete this message to encourage other users to support your issue! --> **Message from the maintainers**: If you wish to see this enhancement implemented please add a 👍 reaction to this issue! We often sort issues this way to know what to prioritize.
non_process
add source data transformer test summary what change needs making use cases when would you use this message from the maintainers if you wish to see this enhancement implemented please add a 👍 reaction to this issue we often sort issues this way to know what to prioritize
0
43,706
5,683,701,111
IssuesEvent
2017-04-13 13:24:19
fabric8io/fabric8-ux
https://api.github.com/repos/fabric8io/fabric8-ux
closed
VISUALS (SPECS): IA
visual design
Verification Requirements: Refer to previous visual story: [VISUALS: IA](https://github.com/fabric8io/fabric8-ux/issues/93), then: - Add "Settings" to the top level nav - Figure out reasonable layout - Figure out User icon, Area icon, etc - Add specifications to previous visuals work - Post new visual specs to InVision, review with UX team members - Link to InVision prototypes from this story - Tag story as "Ready for Development"
1.0
VISUALS (SPECS): IA - Verification Requirements: Refer to previous visual story: [VISUALS: IA](https://github.com/fabric8io/fabric8-ux/issues/93), then: - Add "Settings" to the top level nav - Figure out reasonable layout - Figure out User icon, Area icon, etc - Add specifications to previous visuals work - Post new visual specs to InVision, review with UX team members - Link to InVision prototypes from this story - Tag story as "Ready for Development"
non_process
visuals specs ia verification requirements refer to previous visual story then add settings to the top level nav figure out reasonable layout figure out user icon area icon etc add specifications to previous visuals work post new visual specs to invision review with ux team members link to invision prototypes from this story tag story as ready for development
0
8,379
11,540,598,420
IssuesEvent
2020-02-18 00:43:51
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Two way application data between USAJOBS and Open Opps
Apply Process State Dept.
Applicants are not understanding that once they click apply, any updates on USAJOBS will no longer be pulled over. Michelle has suggested two way updates. Other options: Refresh button modal on apply
1.0
Two way application data between USAJOBS and Open Opps - Applicants are not understanding that once they click apply, any updates on USAJOBS will no longer be pulled over. Michelle has suggested two way updates. Other options: Refresh button modal on apply
process
two way application data between usajobs and open opps applicants are not understanding that once they click apply any updates on usajobs will no longer be pulled over michelle has suggested two way updates other options refresh button modal on apply
1
240,492
7,802,377,653
IssuesEvent
2018-06-10 11:52:59
gama-platform/gama
https://api.github.com/repos/gama-platform/gama
closed
Dynamic parameter view interface
> Enhancement Concerns GAML Concerns Interface Priority Low
Clearly not a top priority enhancement I was just thinking that it could be nice (actually maybe it's already possible) to have a dynamic parameter view interface (show or not the parameter) One simple example. Imagine you have a parameter that enable to compute an interaction graph between agent and then you have a parameter that enable to change the distance parameter to compute the graph. Typically when ComputeGraph is false you don't want to have the distance parameters visible because it will have no effect if you change it. Now when I write ``` bool drawInteraction <- false parameter: "Draw Interaction:" category: "Visualization"; int distance parameter: 'distance ' category: "Visualization" min: 1 max:200 <- 100; ``` The user can play with the distance slider and it will have no effect on the model because draw interaction is set to false. However I don't know what could be a simple syntax in GAML
1.0
Dynamic parameter view interface - Clearly not a top priority enhancement I was just thinking that it could be nice (actually maybe it's already possible) to have a dynamic parameter view interface (show or not the parameter) One simple example. Imagine you have a parameter that enable to compute an interaction graph between agent and then you have a parameter that enable to change the distance parameter to compute the graph. Typically when ComputeGraph is false you don't want to have the distance parameters visible because it will have no effect if you change it. Now when I write ``` bool drawInteraction <- false parameter: "Draw Interaction:" category: "Visualization"; int distance parameter: 'distance ' category: "Visualization" min: 1 max:200 <- 100; ``` The user can play with the distance slider and it will have no effect on the model because draw interaction is set to false. However I don't know what could be a simple syntax in GAML
non_process
dynamic parameter view interface clearly not a top priority enhancement i was just thinking that it could be nice actually maybe it s already possible to have a dynamic parameter view interface show or not the parameter one simple example imagine you have a parameter that enable to compute an interaction graph between agent and then you have a parameter that enable to change the distance parameter to compute the graph typically when computegraph is false you don t want to have the distance parameters visible because it will have no effect if you change it now when i write bool drawinteraction false parameter draw interaction category visualization int distance parameter distance category visualization min max the user can play with the distance slider and it will have no effect on the model because draw interaction is set to false however i don t know what could be a simple syntax in gaml
0
9,670
12,676,214,306
IssuesEvent
2020-06-19 04:24:15
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Detailed report for async_wrap in `process.memoryUsage()`
async_wrap feature request process stalled
I'm curious if we could report total memory used by different kinds of AsyncWrap instances. This should provide some useful insights when fighting memory leaks. cc @trevnorris @nodejs/diagnostics ?
1.0
Detailed report for async_wrap in `process.memoryUsage()` - I'm curious if we could report total memory used by different kinds of AsyncWrap instances. This should provide some useful insights when fighting memory leaks. cc @trevnorris @nodejs/diagnostics ?
process
detailed report for async wrap in process memoryusage i m curious if we could report total memory used by different kinds of asyncwrap instances this should provide some useful insights when fighting memory leaks cc trevnorris nodejs diagnostics
1
26,013
6,734,345,128
IssuesEvent
2017-10-18 17:43:16
Tat5ato/Phantasmic-Mind
https://api.github.com/repos/Tat5ato/Phantasmic-Mind
closed
Code for Item Consumption
Code Secondary
The items that had been thought of that the player is going to use are buffs and debuffs. The two affects not only the player, but also the fear gauge-the alcohol and anti-depressants affects it. As such, it has to call back to the health and fear gauge when used (or some other methods, I'm not too knowledgeable about coding so I don't know how to properly do it).
1.0
Code for Item Consumption - The items that had been thought of that the player is going to use are buffs and debuffs. The two affects not only the player, but also the fear gauge-the alcohol and anti-depressants affects it. As such, it has to call back to the health and fear gauge when used (or some other methods, I'm not too knowledgeable about coding so I don't know how to properly do it).
non_process
code for item consumption the items that had been thought of that the player is going to use are buffs and debuffs the two affects not only the player but also the fear gauge the alcohol and anti depressants affects it as such it has to call back to the health and fear gauge when used or some other methods i m not too knowledgeable about coding so i don t know how to properly do it
0
64,914
6,926,512,661
IssuesEvent
2017-11-30 19:26:03
phetsims/faradays-law
https://api.github.com/repos/phetsims/faradays-law
closed
Artifacts present in Firefox
status:ready-to-test type:question
Test device: MacBook Pro Operating System: 10.9.5 Browser: Firefox 32.0.3 Problem description: Artifacts present in Firefox Steps to reproduce: -Open Sim (iFrame applicable too) -Drag the magnet slowly -Artifacts will be present Severity: Screenshots: ![screen shot 2014-10-07 at 12 13 46 pm](https://cloud.githubusercontent.com/assets/7717368/4547561/2e2584dc-4e4e-11e4-9fdb-a782d1cde7fa.png) Troubleshooting information (do not edit): Name: Faraday's Law URL: http://www.colorado.edu/physics/phet/dev/html/faradays-law/1.0.0-rc.1/faradays-law_en.html Version: 1.0.0-rc.1 Features missing: touch User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.9; rv:32.0) Gecko/20100101 Firefox/32.0 Language: en-US Window: 1280x603 Pixel Ratio: 1/1 WebGL: WebGL 1.0 GLSL: WebGL GLSL ES 1.0 Vendor: Mozilla (Mozilla) Vertex: attribs: 16 varying: 16 uniform: 1024 Texture: size: 4096 imageUnits: 16 (vertex: 16, combined: 16) Max viewport: 8192x8192 OES_texture_float: true
1.0
Artifacts present in Firefox - Test device: MacBook Pro Operating System: 10.9.5 Browser: Firefox 32.0.3 Problem description: Artifacts present in Firefox Steps to reproduce: -Open Sim (iFrame applicable too) -Drag the magnet slowly -Artifacts will be present Severity: Screenshots: ![screen shot 2014-10-07 at 12 13 46 pm](https://cloud.githubusercontent.com/assets/7717368/4547561/2e2584dc-4e4e-11e4-9fdb-a782d1cde7fa.png) Troubleshooting information (do not edit): Name: Faraday's Law URL: http://www.colorado.edu/physics/phet/dev/html/faradays-law/1.0.0-rc.1/faradays-law_en.html Version: 1.0.0-rc.1 Features missing: touch User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.9; rv:32.0) Gecko/20100101 Firefox/32.0 Language: en-US Window: 1280x603 Pixel Ratio: 1/1 WebGL: WebGL 1.0 GLSL: WebGL GLSL ES 1.0 Vendor: Mozilla (Mozilla) Vertex: attribs: 16 varying: 16 uniform: 1024 Texture: size: 4096 imageUnits: 16 (vertex: 16, combined: 16) Max viewport: 8192x8192 OES_texture_float: true
non_process
artifacts present in firefox test device macbook pro operating system browser firefox problem description artifacts present in firefox steps to reproduce open sim iframe applicable too drag the magnet slowly artifacts will be present severity screenshots troubleshooting information do not edit name faraday s law url version rc features missing touch user agent mozilla macintosh intel mac os x rv gecko firefox language en us window pixel ratio webgl webgl glsl webgl glsl es vendor mozilla mozilla vertex attribs varying uniform texture size imageunits vertex combined max viewport oes texture float true
0
14,832
18,169,577,414
IssuesEvent
2021-09-27 18:17:17
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Remove 60 second timeout from the Druid driver
Type:Bug Priority:P3 Querying/Processor Database/Druid
**Describe the bug** This is a continuation of #12003, but specifically for the Druid driver. **IMPORTANT** If you are seeing a timeout, then 99.99% of the time it will have absolutely nothing to do with this issue. You are looking for #12423 - please upvote by clicking :+1: on the first post, do not comment. https://github.com/metabase/metabase/blob/8d3ae7b0ea6675d605a539c950ad26def79e344a/modules/drivers/druid/src/metabase/driver/druid/query_processor.clj#L139 **Expected behavior** None of the other drivers have a timeout, it's a legacy artifact from the olden days. Should either be removed or changed to an environment variable, so it can be modified. ---- Do **not** upvote this issue unless you know that you are explicitly hitting the timeout in the Druid driver, and then provide a comment with "Diagnostic Info" and the Logs from Admin > Troubleshooting.
1.0
Remove 60 second timeout from the Druid driver - **Describe the bug** This is a continuation of #12003, but specifically for the Druid driver. **IMPORTANT** If you are seeing a timeout, then 99.99% of the time it will have absolutely nothing to do with this issue. You are looking for #12423 - please upvote by clicking :+1: on the first post, do not comment. https://github.com/metabase/metabase/blob/8d3ae7b0ea6675d605a539c950ad26def79e344a/modules/drivers/druid/src/metabase/driver/druid/query_processor.clj#L139 **Expected behavior** None of the other drivers have a timeout, it's a legacy artifact from the olden days. Should either be removed or changed to an environment variable, so it can be modified. ---- Do **not** upvote this issue unless you know that you are explicitly hitting the timeout in the Druid driver, and then provide a comment with "Diagnostic Info" and the Logs from Admin > Troubleshooting.
process
remove second timeout from the druid driver describe the bug this is a continuation of but specifically for the druid driver important if you are seeing a timeout then of the time it will have absolutely nothing to do with this issue you are looking for please upvote by clicking on the first post do not comment expected behavior none of the other drivers have a timeout it s a legacy artifact from the olden days should either be removed or changed to an environment variable so it can be modified do not upvote this issue unless you know that you are explicitly hitting the timeout in the druid driver and then provide a comment with diagnostic info and the logs from admin troubleshooting
1
755
2,622,257,016
IssuesEvent
2015-03-04 00:57:05
golang/go
https://api.github.com/repos/golang/go
closed
build: too many open files in buildlet
builder
Just observed this new failure when starting a Windows buildlet: ``` builder: windows-amd64-gce rev: 0cdf1d2bc265b6ed8e1b9834ec9f6c59dcecdcfd vm name: buildlet-windows-amd64-gce-0cdf1d2b-rnbb9a0b started: 2015-03-03 00:49:42.356721167 +0000 UTC started: 2015-03-03 00:50:52.996985212 +0000 UTC success: false Events: 2015-03-03T00:49:43Z instance_create_requested +23.2s 2015-03-03T00:50:06Z instance_created +0.1s 2015-03-03T00:50:06Z waiting_for_buildlet +39.0s 2015-03-03T00:50:45Z buildlet_up +0.0s 2015-03-03T00:50:45Z start_write_version_tar +0.0s 2015-03-03T00:50:45Z start_fetch_gerrit_tgz +0.1s 2015-03-03T00:50:45Z start_write_go_tar +0.0s 2015-03-03T00:50:45Z start_write_go14_tar +7.2s 2015-03-03T00:50:52Z end_write_go_tar Build log: Error: Post http://10.240.20.29/writetgz?dir=go1.4: dial tcp 10.240.20.29:80: too many open files ``` Does the buildlet have an fd leak? /cc @adg
1.0
build: too many open files in buildlet - Just observed this new failure when starting a Windows buildlet: ``` builder: windows-amd64-gce rev: 0cdf1d2bc265b6ed8e1b9834ec9f6c59dcecdcfd vm name: buildlet-windows-amd64-gce-0cdf1d2b-rnbb9a0b started: 2015-03-03 00:49:42.356721167 +0000 UTC started: 2015-03-03 00:50:52.996985212 +0000 UTC success: false Events: 2015-03-03T00:49:43Z instance_create_requested +23.2s 2015-03-03T00:50:06Z instance_created +0.1s 2015-03-03T00:50:06Z waiting_for_buildlet +39.0s 2015-03-03T00:50:45Z buildlet_up +0.0s 2015-03-03T00:50:45Z start_write_version_tar +0.0s 2015-03-03T00:50:45Z start_fetch_gerrit_tgz +0.1s 2015-03-03T00:50:45Z start_write_go_tar +0.0s 2015-03-03T00:50:45Z start_write_go14_tar +7.2s 2015-03-03T00:50:52Z end_write_go_tar Build log: Error: Post http://10.240.20.29/writetgz?dir=go1.4: dial tcp 10.240.20.29:80: too many open files ``` Does the buildlet have an fd leak? /cc @adg
non_process
build too many open files in buildlet just observed this new failure when starting a windows buildlet builder windows gce rev vm name buildlet windows gce started utc started utc success false events instance create requested instance created waiting for buildlet buildlet up start write version tar start fetch gerrit tgz start write go tar start write tar end write go tar build log error post dial tcp too many open files does the buildlet have an fd leak cc adg
0
4,716
17,347,747,511
IssuesEvent
2021-07-29 03:05:55
JacobLinCool/BA
https://api.github.com/repos/JacobLinCool/BA
closed
Automation (2021/29/7 10:55:26 AM)
automation
**Updated.** (2021/29/7 11:00:20 AM) ## 登入: 完成 ``` [2021/29/7 10:55:28 AM] 開始執行帳號登入程序 [2021/29/7 10:55:36 AM] 正在檢測登入狀態 [2021/29/7 10:55:39 AM] 登入狀態: 未登入 [2021/29/7 10:55:42 AM] 嘗試登入中 [2021/29/7 10:55:52 AM] 已嘗試登入,重新檢測登入狀態 [2021/29/7 10:55:52 AM] 正在檢測登入狀態 [2021/29/7 10:55:55 AM] 登入狀態: 已登入 [2021/29/7 10:55:55 AM] 帳號登入程序已完成 ``` ## 簽到: 完成 ``` [2021/29/7 10:55:55 AM] 開始執行自動簽到程序 [2021/29/7 10:55:55 AM] 正在檢測簽到狀態 [2021/29/7 10:55:59 AM] 簽到狀態: 已簽到 [2021/29/7 10:56:00 AM] 自動簽到程序已完成 [2021/29/7 10:56:00 AM] 開始執行自動觀看雙倍簽到獎勵廣告程序 [2021/29/7 10:56:00 AM] 正在檢測雙倍簽到獎勵狀態 [2021/29/7 10:56:05 AM] 雙倍簽到獎勵狀態: 已獲得雙倍簽到獎勵 [2021/29/7 10:56:05 AM] 自動觀看雙倍簽到獎勵廣告程序已完成 ``` ## 答題: 完成 ``` [2021/29/7 10:56:06 AM] 開始執行動畫瘋自動答題程序 [2021/29/7 10:56:06 AM] 正在檢測答題狀態 [2021/29/7 10:56:12 AM] 今日已經答過題目了 [2021/29/7 10:56:13 AM] 動畫瘋自動答題程序已完成 ``` ## 抽獎: 執行中 ``` [2021/29/7 10:56:13 AM] 開始執行福利社自動抽抽樂程序 [2021/29/7 10:56:13 AM] 正在尋找抽抽樂 [2021/29/7 10:56:17 AM] 找到 8 個抽抽樂 [2021/29/7 10:56:17 AM] 1: 一個打四個!綠聯 充電器 GaN快充版 3C1A [2021/29/7 10:56:17 AM] 2: 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! [2021/29/7 10:56:17 AM] 3: 又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎 [2021/29/7 10:56:17 AM] 4: 你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈 [2021/29/7 10:56:17 AM] 5: 熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂! [2021/29/7 10:56:17 AM] 6: XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐! [2021/29/7 10:56:17 AM] 7: GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版! [2021/29/7 10:56:17 AM] 8: EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:56:17 AM] 正在嘗試執行第 1 個抽抽樂: 一個打四個!綠聯 充電器 GaN快充版 3C1A [2021/29/7 10:56:21 AM] 第 1 個抽抽樂(一個打四個!綠聯 充電器 GaN快充版 3C1A)的廣告免費次數已用完 [2021/29/7 10:56:21 AM] 正在嘗試執行第 2 個抽抽樂: 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! [2021/29/7 10:56:24 AM] 第 2 個抽抽樂(《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂!)的廣告免費次數已用完 [2021/29/7 10:56:24 AM] 正在嘗試執行第 3 個抽抽樂: 又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎 [2021/29/7 10:56:27 AM] 第 3 個抽抽樂(又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎)的廣告免費次數已用完 [2021/29/7 10:56:27 AM] 正在嘗試執行第 4 個抽抽樂: 你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈 [2021/29/7 10:56:29 AM] 第 4 個抽抽樂(你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈)的廣告免費次數已用完 [2021/29/7 10:56:29 AM] 正在嘗試執行第 5 個抽抽樂: 熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂! [2021/29/7 10:56:32 AM] 第 5 個抽抽樂(熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂!)的廣告免費次數已用完 [2021/29/7 10:56:32 AM] 正在嘗試執行第 6 個抽抽樂: XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐! [2021/29/7 10:56:35 AM] 第 6 個抽抽樂(XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐!)的廣告免費次數已用完 [2021/29/7 10:56:35 AM] 正在嘗試執行第 7 個抽抽樂: GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版! [2021/29/7 10:56:37 AM] 第 7 個抽抽樂(GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版!)的廣告免費次數已用完 [2021/29/7 10:56:37 AM] 正在嘗試執行第 8 個抽抽樂: EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:56:40 AM] 正在執行第 1 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:56:46 AM] 正在觀看廣告 [2021/29/7 10:57:30 AM] 未進入結算頁面,重試中 [2021/29/7 10:57:32 AM] 正在執行第 2 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:57:38 AM] 正在觀看廣告 [2021/29/7 10:58:22 AM] 正在確認結算頁面 [2021/29/7 10:58:26 AM] 已完成一次抽抽樂:EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:58:29 AM] 正在執行第 3 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:58:35 AM] 正在觀看廣告 [2021/29/7 10:59:19 AM] 未進入結算頁面,重試中 [2021/29/7 10:59:21 AM] 正在執行第 4 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:59:27 AM] 正在觀看廣告 [2021/29/7 11:00:11 AM] 未進入結算頁面,重試中 [2021/29/7 11:00:13 AM] 正在執行第 5 次抽獎,可能需要多達 1 分鐘 [2021/29/7 11:00:20 AM] 正在觀看廣告 ```
1.0
Automation (2021/29/7 10:55:26 AM) - **Updated.** (2021/29/7 11:00:20 AM) ## 登入: 完成 ``` [2021/29/7 10:55:28 AM] 開始執行帳號登入程序 [2021/29/7 10:55:36 AM] 正在檢測登入狀態 [2021/29/7 10:55:39 AM] 登入狀態: 未登入 [2021/29/7 10:55:42 AM] 嘗試登入中 [2021/29/7 10:55:52 AM] 已嘗試登入,重新檢測登入狀態 [2021/29/7 10:55:52 AM] 正在檢測登入狀態 [2021/29/7 10:55:55 AM] 登入狀態: 已登入 [2021/29/7 10:55:55 AM] 帳號登入程序已完成 ``` ## 簽到: 完成 ``` [2021/29/7 10:55:55 AM] 開始執行自動簽到程序 [2021/29/7 10:55:55 AM] 正在檢測簽到狀態 [2021/29/7 10:55:59 AM] 簽到狀態: 已簽到 [2021/29/7 10:56:00 AM] 自動簽到程序已完成 [2021/29/7 10:56:00 AM] 開始執行自動觀看雙倍簽到獎勵廣告程序 [2021/29/7 10:56:00 AM] 正在檢測雙倍簽到獎勵狀態 [2021/29/7 10:56:05 AM] 雙倍簽到獎勵狀態: 已獲得雙倍簽到獎勵 [2021/29/7 10:56:05 AM] 自動觀看雙倍簽到獎勵廣告程序已完成 ``` ## 答題: 完成 ``` [2021/29/7 10:56:06 AM] 開始執行動畫瘋自動答題程序 [2021/29/7 10:56:06 AM] 正在檢測答題狀態 [2021/29/7 10:56:12 AM] 今日已經答過題目了 [2021/29/7 10:56:13 AM] 動畫瘋自動答題程序已完成 ``` ## 抽獎: 執行中 ``` [2021/29/7 10:56:13 AM] 開始執行福利社自動抽抽樂程序 [2021/29/7 10:56:13 AM] 正在尋找抽抽樂 [2021/29/7 10:56:17 AM] 找到 8 個抽抽樂 [2021/29/7 10:56:17 AM] 1: 一個打四個!綠聯 充電器 GaN快充版 3C1A [2021/29/7 10:56:17 AM] 2: 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! [2021/29/7 10:56:17 AM] 3: 又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎 [2021/29/7 10:56:17 AM] 4: 你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈 [2021/29/7 10:56:17 AM] 5: 熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂! [2021/29/7 10:56:17 AM] 6: XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐! [2021/29/7 10:56:17 AM] 7: GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版! [2021/29/7 10:56:17 AM] 8: EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:56:17 AM] 正在嘗試執行第 1 個抽抽樂: 一個打四個!綠聯 充電器 GaN快充版 3C1A [2021/29/7 10:56:21 AM] 第 1 個抽抽樂(一個打四個!綠聯 充電器 GaN快充版 3C1A)的廣告免費次數已用完 [2021/29/7 10:56:21 AM] 正在嘗試執行第 2 個抽抽樂: 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! [2021/29/7 10:56:24 AM] 第 2 個抽抽樂(《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂!)的廣告免費次數已用完 [2021/29/7 10:56:24 AM] 正在嘗試執行第 3 個抽抽樂: 又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎 [2021/29/7 10:56:27 AM] 第 3 個抽抽樂(又到了白色鍵盤的季節!irocks K71M RGB 機械鍵盤抽獎)的廣告免費次數已用完 [2021/29/7 10:56:27 AM] 正在嘗試執行第 4 個抽抽樂: 你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈 [2021/29/7 10:56:29 AM] 第 4 個抽抽樂(你的明眸護眼法寶 - Awesome LED觸控式可調雙光源螢幕掛燈)的廣告免費次數已用完 [2021/29/7 10:56:29 AM] 正在嘗試執行第 5 個抽抽樂: 熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂! [2021/29/7 10:56:32 AM] 第 5 個抽抽樂(熱銷百萬!2021最新款 「TruEgos Super 2NC雙工抗噪真無線」-限時抽抽樂!)的廣告免費次數已用完 [2021/29/7 10:56:32 AM] 正在嘗試執行第 6 個抽抽樂: XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐! [2021/29/7 10:56:35 AM] 第 6 個抽抽樂(XPG競爆你的電競生活,好禮大方送,附送 MANA 電競口香糖 8/11 上市前搶先嚐!)的廣告免費次數已用完 [2021/29/7 10:56:35 AM] 正在嘗試執行第 7 個抽抽樂: GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版! [2021/29/7 10:56:37 AM] 第 7 個抽抽樂(GoKids玩樂小子|深入絕地:暗黑世界傳說-跨越16年的經典RPG遊戲,史詩繁中再版!)的廣告免費次數已用完 [2021/29/7 10:56:37 AM] 正在嘗試執行第 8 個抽抽樂: EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:56:40 AM] 正在執行第 1 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:56:46 AM] 正在觀看廣告 [2021/29/7 10:57:30 AM] 未進入結算頁面,重試中 [2021/29/7 10:57:32 AM] 正在執行第 2 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:57:38 AM] 正在觀看廣告 [2021/29/7 10:58:22 AM] 正在確認結算頁面 [2021/29/7 10:58:26 AM] 已完成一次抽抽樂:EPOS |Sennheiser 最強電競耳機─王者回歸,GSP 602抽起來 [2021/29/7 10:58:29 AM] 正在執行第 3 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:58:35 AM] 正在觀看廣告 [2021/29/7 10:59:19 AM] 未進入結算頁面,重試中 [2021/29/7 10:59:21 AM] 正在執行第 4 次抽獎,可能需要多達 1 分鐘 [2021/29/7 10:59:27 AM] 正在觀看廣告 [2021/29/7 11:00:11 AM] 未進入結算頁面,重試中 [2021/29/7 11:00:13 AM] 正在執行第 5 次抽獎,可能需要多達 1 分鐘 [2021/29/7 11:00:20 AM] 正在觀看廣告 ```
non_process
automation am updated am 登入 完成 開始執行帳號登入程序 正在檢測登入狀態 登入狀態 未登入 嘗試登入中 已嘗試登入,重新檢測登入狀態 正在檢測登入狀態 登入狀態 已登入 帳號登入程序已完成 簽到 完成 開始執行自動簽到程序 正在檢測簽到狀態 簽到狀態 已簽到 自動簽到程序已完成 開始執行自動觀看雙倍簽到獎勵廣告程序 正在檢測雙倍簽到獎勵狀態 雙倍簽到獎勵狀態 已獲得雙倍簽到獎勵 自動觀看雙倍簽到獎勵廣告程序已完成 答題 完成 開始執行動畫瘋自動答題程序 正在檢測答題狀態 今日已經答過題目了 動畫瘋自動答題程序已完成 抽獎 執行中 開始執行福利社自動抽抽樂程序 正在尋找抽抽樂 找到 個抽抽樂 一個打四個!綠聯 充電器 gan快充版 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! 又到了白色鍵盤的季節!irocks rgb 機械鍵盤抽獎 你的明眸護眼法寶 awesome led觸控式可調雙光源螢幕掛燈 熱銷百萬! 「truegos super 」 限時抽抽樂! xpg競爆你的電競生活,好禮大方送,附送 mana 電競口香糖 上市前搶先嚐! gokids玩樂小子|深入絕地:暗黑世界傳說 ,史詩繁中再版! epos |sennheiser 最強電競耳機─王者回歸,gsp 正在嘗試執行第 個抽抽樂: 一個打四個!綠聯 充電器 gan快充版 第 個抽抽樂(一個打四個!綠聯 充電器 gan快充版 )的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: 《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂! 第 個抽抽樂(《信星科技》一指雙用,五指操控,飛智黃蜂搖桿抽抽樂!)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: 又到了白色鍵盤的季節!irocks rgb 機械鍵盤抽獎 第 個抽抽樂(又到了白色鍵盤的季節!irocks rgb 機械鍵盤抽獎)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: 你的明眸護眼法寶 awesome led觸控式可調雙光源螢幕掛燈 第 個抽抽樂(你的明眸護眼法寶 awesome led觸控式可調雙光源螢幕掛燈)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: 熱銷百萬! 「truegos super 」 限時抽抽樂! 第 個抽抽樂(熱銷百萬! 「truegos super 」 限時抽抽樂!)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: xpg競爆你的電競生活,好禮大方送,附送 mana 電競口香糖 上市前搶先嚐! 第 個抽抽樂(xpg競爆你的電競生活,好禮大方送,附送 mana 電競口香糖 上市前搶先嚐!)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: gokids玩樂小子|深入絕地:暗黑世界傳說 ,史詩繁中再版! 第 個抽抽樂(gokids玩樂小子|深入絕地:暗黑世界傳說 ,史詩繁中再版!)的廣告免費次數已用完 正在嘗試執行第 個抽抽樂: epos |sennheiser 最強電競耳機─王者回歸,gsp 正在執行第 次抽獎,可能需要多達 分鐘 正在觀看廣告 未進入結算頁面,重試中 正在執行第 次抽獎,可能需要多達 分鐘 正在觀看廣告 正在確認結算頁面 已完成一次抽抽樂:epos |sennheiser 最強電競耳機─王者回歸,gsp 正在執行第 次抽獎,可能需要多達 分鐘 正在觀看廣告 未進入結算頁面,重試中 正在執行第 次抽獎,可能需要多達 分鐘 正在觀看廣告 未進入結算頁面,重試中 正在執行第 次抽獎,可能需要多達 分鐘 正在觀看廣告
0
7,505
6,924,352,193
IssuesEvent
2017-11-30 12:26:12
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
Allow to overwrite security config
Feature Security SecurityBundle
| Q | A | ---------------- | ----- | Bug report? | no | Feature request? | yes | BC Break report? | no | RFC? | no | Symfony version | 4.1 Recall for https://github.com/symfony/symfony/issues/22308 :) ```yaml # security.yaml firewalls: # .... api: pattern: ^/api stateless: true anonymous: true # ... ``` ```php # bundle extension public function prepend(ContainerBuilder $container) { $security = $container->getExtensionConfig('security'); $security[0]['firewalls']['api']['anonymous'] = false; $security[0]['firewalls']['api']['fos_oauth'] = true; $container->loadFromExtension('security', $security[0]); } ``` Result ``` Configuration path "security.access_control" cannot be overwritten. You have to define all options for this path, and any of its sub-paths in one configuration section. ``` I understand `loadFromExtension` adds a new config branch (`$configs[] = $config`). But im trying to fake 1 config branch, so i guess i need reflection to clear current branch first. Can we create API for that? Or what about allowing overwrites on existing elements? Thx!
True
Allow to overwrite security config - | Q | A | ---------------- | ----- | Bug report? | no | Feature request? | yes | BC Break report? | no | RFC? | no | Symfony version | 4.1 Recall for https://github.com/symfony/symfony/issues/22308 :) ```yaml # security.yaml firewalls: # .... api: pattern: ^/api stateless: true anonymous: true # ... ``` ```php # bundle extension public function prepend(ContainerBuilder $container) { $security = $container->getExtensionConfig('security'); $security[0]['firewalls']['api']['anonymous'] = false; $security[0]['firewalls']['api']['fos_oauth'] = true; $container->loadFromExtension('security', $security[0]); } ``` Result ``` Configuration path "security.access_control" cannot be overwritten. You have to define all options for this path, and any of its sub-paths in one configuration section. ``` I understand `loadFromExtension` adds a new config branch (`$configs[] = $config`). But im trying to fake 1 config branch, so i guess i need reflection to clear current branch first. Can we create API for that? Or what about allowing overwrites on existing elements? Thx!
non_process
allow to overwrite security config q a bug report no feature request yes bc break report no rfc no symfony version recall for yaml security yaml firewalls api pattern api stateless true anonymous true php bundle extension public function prepend containerbuilder container security container getextensionconfig security security false security true container loadfromextension security security result configuration path security access control cannot be overwritten you have to define all options for this path and any of its sub paths in one configuration section i understand loadfromextension adds a new config branch configs config but im trying to fake config branch so i guess i need reflection to clear current branch first can we create api for that or what about allowing overwrites on existing elements thx
0
15,550
19,703,502,581
IssuesEvent
2022-01-12 19:07:56
googleapis/java-os-config
https://api.github.com/repos/googleapis/java-os-config
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'os-config' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'os-config' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname os config invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
137,456
18,752,717,193
IssuesEvent
2021-11-05 05:53:40
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2018-7273 (Medium) detected in linux-stablev4.17.12
security vulnerability
## CVE-2018-7273 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stablev4.17.12</b></p></summary> <p> <p>Linux kernel stable tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git>https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/floppy.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/floppy.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 4.15.4, the floppy driver reveals the addresses of kernel functions and global variables using printk calls within the function show_floppy in drivers/block/floppy.c. An attacker can read this information from dmesg and use the addresses to find the locations of kernel code and data and bypass kernel security protections such as KASLR. <p>Publish Date: 2018-02-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7273>CVE-2018-7273</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-7273 (Medium) detected in linux-stablev4.17.12 - ## CVE-2018-7273 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stablev4.17.12</b></p></summary> <p> <p>Linux kernel stable tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git>https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/floppy.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/floppy.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 4.15.4, the floppy driver reveals the addresses of kernel functions and global variables using printk calls within the function show_floppy in drivers/block/floppy.c. An attacker can read this information from dmesg and use the addresses to find the locations of kernel code and data and bypass kernel security protections such as KASLR. <p>Publish Date: 2018-02-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7273>CVE-2018-7273</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux cve medium severity vulnerability vulnerable library linux linux kernel stable tree library home page a href found in head commit a href found in base branch master vulnerable source files drivers block floppy c drivers block floppy c vulnerability details in the linux kernel through the floppy driver reveals the addresses of kernel functions and global variables using printk calls within the function show floppy in drivers block floppy c an attacker can read this information from dmesg and use the addresses to find the locations of kernel code and data and bypass kernel security protections such as kaslr publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href step up your open source security game with whitesource
0
3,883
2,694,566,939
IssuesEvent
2015-04-01 21:00:50
oel-mediateam/sbplus
https://api.github.com/repos/oel-mediateam/sbplus
closed
Testing results from Bryan
testing-required
Before you start testing, please perform the following steps. 1. Launch Transmit app. If you have not setup Transmit, please follow the [Transmit documentation]( https://media.uwex.edu/resources/documentation/transmit/). 3. Navigate to your sandbox. 2. Delete the "build" folder in your sandbox. The "build" folder contains the GVP project files. This is not related to SB+, but it is good idea to remove files in your sandbox that we no longer use on the web server. We pay for storage space now. If you have already done the steps above, thank you for taking your time to do it again. **Please check-off each task once you completed or confirmed it.** - [x] Clear browser cache. - [x] Check [jQuery script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/scripts/jquery.min.js). Is the version number 2.1.3? - [x] Check [SB+ JS Script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/scripts/storybookplus.min.js). Is the version number 2.6.0? - [x] Check [VideoJs script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/videoplayer/combined.video.min.js) version. Is the version number 4.11.4? - [x] Go to [SB+ testing page](https://media.uwex.edu/sandbox/ethan/sbplus/build/). - [x] Download bar loads and supplement download button is shown? - [x] Click play. On the table of contents, does "Table of Contents" label shows up and stay fixed at the top? - [x] Try double clicking the table of contents list or the control buttons. The texts should not be able to be selected. The texts on the table of contents can be selected **if you attempt to select them like you would normally do**, but it will not persist. - [x] Click the "Expand" button. Click the "Table of Contents" button. Click the "Contract" button. Does the table of contents stay/show up at where it supposed to be? - [x] Click the "Expand" button again. Play with the "Table of Contents" toggle button and then click the "Contract" button. Does the table of contents still stay/show up at where it supposed to be? - [x] Navigate through each slide. All seems okay? Please pay special attentions to slide 2, 5, 6, 8, 10, 15, and 16. - [x] When navigate to an item on the table of content that is out of view, does the table of content auto scroll down? - [ ] Sign-off for release. Only check this box if you find no issues.
1.0
Testing results from Bryan - Before you start testing, please perform the following steps. 1. Launch Transmit app. If you have not setup Transmit, please follow the [Transmit documentation]( https://media.uwex.edu/resources/documentation/transmit/). 3. Navigate to your sandbox. 2. Delete the "build" folder in your sandbox. The "build" folder contains the GVP project files. This is not related to SB+, but it is good idea to remove files in your sandbox that we no longer use on the web server. We pay for storage space now. If you have already done the steps above, thank you for taking your time to do it again. **Please check-off each task once you completed or confirmed it.** - [x] Clear browser cache. - [x] Check [jQuery script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/scripts/jquery.min.js). Is the version number 2.1.3? - [x] Check [SB+ JS Script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/scripts/storybookplus.min.js). Is the version number 2.6.0? - [x] Check [VideoJs script](https://media.uwex.edu/sandbox/ethan/sbplus/sources/videoplayer/combined.video.min.js) version. Is the version number 4.11.4? - [x] Go to [SB+ testing page](https://media.uwex.edu/sandbox/ethan/sbplus/build/). - [x] Download bar loads and supplement download button is shown? - [x] Click play. On the table of contents, does "Table of Contents" label shows up and stay fixed at the top? - [x] Try double clicking the table of contents list or the control buttons. The texts should not be able to be selected. The texts on the table of contents can be selected **if you attempt to select them like you would normally do**, but it will not persist. - [x] Click the "Expand" button. Click the "Table of Contents" button. Click the "Contract" button. Does the table of contents stay/show up at where it supposed to be? - [x] Click the "Expand" button again. Play with the "Table of Contents" toggle button and then click the "Contract" button. Does the table of contents still stay/show up at where it supposed to be? - [x] Navigate through each slide. All seems okay? Please pay special attentions to slide 2, 5, 6, 8, 10, 15, and 16. - [x] When navigate to an item on the table of content that is out of view, does the table of content auto scroll down? - [ ] Sign-off for release. Only check this box if you find no issues.
non_process
testing results from bryan before you start testing please perform the following steps launch transmit app if you have not setup transmit please follow the navigate to your sandbox delete the build folder in your sandbox the build folder contains the gvp project files this is not related to sb but it is good idea to remove files in your sandbox that we no longer use on the web server we pay for storage space now if you have already done the steps above thank you for taking your time to do it again please check off each task once you completed or confirmed it clear browser cache check is the version number check is the version number check version is the version number go to download bar loads and supplement download button is shown click play on the table of contents does table of contents label shows up and stay fixed at the top try double clicking the table of contents list or the control buttons the texts should not be able to be selected the texts on the table of contents can be selected if you attempt to select them like you would normally do but it will not persist click the expand button click the table of contents button click the contract button does the table of contents stay show up at where it supposed to be click the expand button again play with the table of contents toggle button and then click the contract button does the table of contents still stay show up at where it supposed to be navigate through each slide all seems okay please pay special attentions to slide and when navigate to an item on the table of content that is out of view does the table of content auto scroll down sign off for release only check this box if you find no issues
0
725,531
24,964,997,380
IssuesEvent
2022-11-01 18:36:06
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
[Epic]: MVC Enhancements
Epic Priority:0 Bottom Up Work area-web-frameworks
This epic is to track all the potential work for MVC in NET7. ## Issues categories - [x] Enhance MVC with Minimal API philosophy (Inferred types in the DI, optionality from Nullability, etc.) - #33403 - #29570 ~~[ ] Support generic attributes: #37767~~ - [x] Fix `ProblemDetails` inconsistencies (Breaking): #32957 ~~[ ] Support CORS for 500/404/401/403 response types: #22281~~ - [x] #39425
1.0
[Epic]: MVC Enhancements - This epic is to track all the potential work for MVC in NET7. ## Issues categories - [x] Enhance MVC with Minimal API philosophy (Inferred types in the DI, optionality from Nullability, etc.) - #33403 - #29570 ~~[ ] Support generic attributes: #37767~~ - [x] Fix `ProblemDetails` inconsistencies (Breaking): #32957 ~~[ ] Support CORS for 500/404/401/403 response types: #22281~~ - [x] #39425
non_process
mvc enhancements this epic is to track all the potential work for mvc in issues categories enhance mvc with minimal api philosophy inferred types in the di optionality from nullability etc support generic attributes fix problemdetails inconsistencies breaking support cors for response types
0
274,631
30,087,273,376
IssuesEvent
2023-06-29 09:33:39
slurpcode/slurp
https://api.github.com/repos/slurpcode/slurp
opened
CVE-2020-23064 (Medium) detected in jquery-3.1.1.min.js
Mend: dependency security vulnerability
## CVE-2020-23064 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p>Path to dependency file: /rubycritic/built-in-datatypes/built_in_datatypes.html</p> <p>Path to vulnerable library: /rubycritic/built-in-datatypes/../assets/vendor/javascripts/jquery.min.js,/rubycritic/assets/vendor/javascripts/jquery.min.js,/rubycritic/drivers/../assets/vendor/javascripts/jquery.min.js,/rubycritic/scrapers/ruby/nokogiri/../../../assets/vendor/javascripts/jquery.min.js,/rubycritic/hashcheck/../assets/vendor/javascripts/jquery.min.js,/rubycritic/charts/../assets/vendor/javascripts/jquery.min.js,/rubycritic/well-formed/../assets/vendor/javascripts/jquery.min.js,/rubycritic/seek/../assets/vendor/javascripts/jquery.min.js,/rubycritic/ruby-strip/../assets/vendor/javascripts/jquery.min.js,/rubycritic/ruby-eclipse-cheatsheets-to-dita/../assets/vendor/javascripts/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/slurpcode/slurp/commit/8dbdb4a1170502df116e35d16ab172d26c02609e">8dbdb4a1170502df116e35d16ab172d26c02609e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element. <p>Publish Date: 2023-06-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2023-06-26</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-23064 (Medium) detected in jquery-3.1.1.min.js - ## CVE-2020-23064 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p>Path to dependency file: /rubycritic/built-in-datatypes/built_in_datatypes.html</p> <p>Path to vulnerable library: /rubycritic/built-in-datatypes/../assets/vendor/javascripts/jquery.min.js,/rubycritic/assets/vendor/javascripts/jquery.min.js,/rubycritic/drivers/../assets/vendor/javascripts/jquery.min.js,/rubycritic/scrapers/ruby/nokogiri/../../../assets/vendor/javascripts/jquery.min.js,/rubycritic/hashcheck/../assets/vendor/javascripts/jquery.min.js,/rubycritic/charts/../assets/vendor/javascripts/jquery.min.js,/rubycritic/well-formed/../assets/vendor/javascripts/jquery.min.js,/rubycritic/seek/../assets/vendor/javascripts/jquery.min.js,/rubycritic/ruby-strip/../assets/vendor/javascripts/jquery.min.js,/rubycritic/ruby-eclipse-cheatsheets-to-dita/../assets/vendor/javascripts/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/slurpcode/slurp/commit/8dbdb4a1170502df116e35d16ab172d26c02609e">8dbdb4a1170502df116e35d16ab172d26c02609e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element. <p>Publish Date: 2023-06-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2023-06-26</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file rubycritic built in datatypes built in datatypes html path to vulnerable library rubycritic built in datatypes assets vendor javascripts jquery min js rubycritic assets vendor javascripts jquery min js rubycritic drivers assets vendor javascripts jquery min js rubycritic scrapers ruby nokogiri assets vendor javascripts jquery min js rubycritic hashcheck assets vendor javascripts jquery min js rubycritic charts assets vendor javascripts jquery min js rubycritic well formed assets vendor javascripts jquery min js rubycritic seek assets vendor javascripts jquery min js rubycritic ruby strip assets vendor javascripts jquery min js rubycritic ruby eclipse cheatsheets to dita assets vendor javascripts jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details cross site scripting vulnerability in jquery through x before allows a remote attacker to execute arbitrary code via the element publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
0
654,784
21,662,493,316
IssuesEvent
2022-05-06 21:04:59
operator-framework/rukpak
https://api.github.com/repos/operator-framework/rukpak
closed
Add support for embedding Bundles in the BundleInstance spec
priority/important-longterm
Implement support in the BundleInstance API and the plain manifest bundle provisioner for embedding a Bundle spec as outlined in the [e2e strawman](https://hackmd.io/hppR60SiRGKPcqKQuByvVg?view#Instance-API).
1.0
Add support for embedding Bundles in the BundleInstance spec - Implement support in the BundleInstance API and the plain manifest bundle provisioner for embedding a Bundle spec as outlined in the [e2e strawman](https://hackmd.io/hppR60SiRGKPcqKQuByvVg?view#Instance-API).
non_process
add support for embedding bundles in the bundleinstance spec implement support in the bundleinstance api and the plain manifest bundle provisioner for embedding a bundle spec as outlined in the
0
323,408
23,946,229,151
IssuesEvent
2022-09-12 07:35:18
boto/botocore
https://api.github.com/repos/boto/botocore
opened
Textract / analyze-document / Document parameter has inaccurate documentation
documentation needs-triage
### Describe the issue The public facing Boto3 documentation for the Amazon Textract service give the following description for the Document / Bytes parameter: > **Bytes** (bytes) -- > A blob of base64-encoded document bytes. The maximum size of a document that's provided in a blob of bytes is 5 MB. The document bytes must be in PNG or JPEG format. This description is incorrect. As per the [API documentation](https://docs.aws.amazon.com/textract/latest/dg/API_AnalyzeDocument.html), the document bytes can be PNG, JPEG, PDF, or TIFF. This inaccuracy has been the cause of wasted time in unnecessarily building additional functionality to process PDF documents. --- Suggest replacing description with: > **Bytes** (bytes) -- > A blob of base64-encoded document bytes. The maximum size of a document that's provided in a blob of bytes is 5 MB. The document bytes must be in **PNG, JPEG, PDF, or TIFF format. ** ### Links https://github.com/boto/botocore/blob/develop/botocore/data/textract/2018-06-27/service-2.json#L512 https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/textract.html#Textract.Client.analyze_document
1.0
Textract / analyze-document / Document parameter has inaccurate documentation - ### Describe the issue The public facing Boto3 documentation for the Amazon Textract service give the following description for the Document / Bytes parameter: > **Bytes** (bytes) -- > A blob of base64-encoded document bytes. The maximum size of a document that's provided in a blob of bytes is 5 MB. The document bytes must be in PNG or JPEG format. This description is incorrect. As per the [API documentation](https://docs.aws.amazon.com/textract/latest/dg/API_AnalyzeDocument.html), the document bytes can be PNG, JPEG, PDF, or TIFF. This inaccuracy has been the cause of wasted time in unnecessarily building additional functionality to process PDF documents. --- Suggest replacing description with: > **Bytes** (bytes) -- > A blob of base64-encoded document bytes. The maximum size of a document that's provided in a blob of bytes is 5 MB. The document bytes must be in **PNG, JPEG, PDF, or TIFF format. ** ### Links https://github.com/boto/botocore/blob/develop/botocore/data/textract/2018-06-27/service-2.json#L512 https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/textract.html#Textract.Client.analyze_document
non_process
textract analyze document document parameter has inaccurate documentation describe the issue the public facing documentation for the amazon textract service give the following description for the document bytes parameter bytes bytes a blob of encoded document bytes the maximum size of a document that s provided in a blob of bytes is mb the document bytes must be in png or jpeg format this description is incorrect as per the the document bytes can be png jpeg pdf or tiff this inaccuracy has been the cause of wasted time in unnecessarily building additional functionality to process pdf documents suggest replacing description with bytes bytes a blob of encoded document bytes the maximum size of a document that s provided in a blob of bytes is mb the document bytes must be in png jpeg pdf or tiff format links
0
49,318
10,338,133,510
IssuesEvent
2019-09-03 16:12:36
kabanero-io/roadmap
https://api.github.com/repos/kabanero-io/roadmap
closed
Jane's Kabanero experience should likely involve pointing the Codewind and Appsody CLI to Kabanero Repo
blockedBy-Codewind release-0.1
@stephenkinder commented on [Mon Aug 19 2019](https://github.com/kabanero-io/kabanero-collection/issues/58) The current directions on the kabanero.io website do not show how to change the CLI to the default Kabanero Repository.
1.0
Jane's Kabanero experience should likely involve pointing the Codewind and Appsody CLI to Kabanero Repo - @stephenkinder commented on [Mon Aug 19 2019](https://github.com/kabanero-io/kabanero-collection/issues/58) The current directions on the kabanero.io website do not show how to change the CLI to the default Kabanero Repository.
non_process
jane s kabanero experience should likely involve pointing the codewind and appsody cli to kabanero repo stephenkinder commented on the current directions on the kabanero io website do not show how to change the cli to the default kabanero repository
0
20,659
27,330,098,845
IssuesEvent
2023-02-25 14:15:20
benthosdev/benthos
https://api.github.com/repos/benthosdev/benthos
closed
add support to sqlite driver
enhancement processors inputs outputs
Hello I have to deal with some sqlite files and benthos seems the right tool for this. I don't know what is the criteria to add a new driver, I made a small POC and it works fine (I guess) Enjoy
1.0
add support to sqlite driver - Hello I have to deal with some sqlite files and benthos seems the right tool for this. I don't know what is the criteria to add a new driver, I made a small POC and it works fine (I guess) Enjoy
process
add support to sqlite driver hello i have to deal with some sqlite files and benthos seems the right tool for this i don t know what is the criteria to add a new driver i made a small poc and it works fine i guess enjoy
1
6,566
9,652,816,039
IssuesEvent
2019-05-18 20:42:48
bcoe/c8
https://api.github.com/repos/bcoe/c8
closed
chore(release): proposal for next release
release-candidate type: process
_:robot: Here's what the next release of **c8** would look like._ --- ### [4.1.6](https://www.github.com/bcoe/c8/compare/v4.1.5...v4.1.6) (2019-05-18) ---------------- * [ ] **Should I create this release for you :robot:?**
1.0
chore(release): proposal for next release - _:robot: Here's what the next release of **c8** would look like._ --- ### [4.1.6](https://www.github.com/bcoe/c8/compare/v4.1.5...v4.1.6) (2019-05-18) ---------------- * [ ] **Should I create this release for you :robot:?**
process
chore release proposal for next release robot here s what the next release of would look like should i create this release for you robot
1
11,797
14,624,026,974
IssuesEvent
2020-12-23 05:13:13
rohanchandra30/Spectral-Trajectory-and-Behavior-Prediction
https://api.github.com/repos/rohanchandra30/Spectral-Trajectory-and-Behavior-Prediction
closed
Missing trainSet0.py error during data stream
Data Processing
Hello, I have run the generate_data.py to generate the trainSet0.txt, trainSet0-track.npy and trainSet0-traj.npy by using the Argoverse dataset. In the data_stream.py, the trainSet0.npy should be provided, which is considered as the input of the function data_for_stream1. However, I can not get the trainSet0.npy in the generate_data.py or other related files. How to generate the trainSet0.npy? Thanks~ ![image](https://user-images.githubusercontent.com/48577880/99259764-724e5700-2855-11eb-8bd4-f5e6f419b177.png)
1.0
Missing trainSet0.py error during data stream - Hello, I have run the generate_data.py to generate the trainSet0.txt, trainSet0-track.npy and trainSet0-traj.npy by using the Argoverse dataset. In the data_stream.py, the trainSet0.npy should be provided, which is considered as the input of the function data_for_stream1. However, I can not get the trainSet0.npy in the generate_data.py or other related files. How to generate the trainSet0.npy? Thanks~ ![image](https://user-images.githubusercontent.com/48577880/99259764-724e5700-2855-11eb-8bd4-f5e6f419b177.png)
process
missing py error during data stream hello i have run the generate data py to generate the txt track npy and traj npy by using the argoverse dataset in the data stream py the npy should be provided which is considered as the input of the function data for however i can not get the npy in the generate data py or other related files how to generate the npy thanks
1
13,935
16,702,812,222
IssuesEvent
2021-06-09 06:16:05
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
emmeans support - regrid using an 'exp' transform argument?
feature post-processing
Hello, I'm wanting to regrid an emmGrid object that has `mu` estimates that was modelled as a natural log with an identity link function (using the shifted lognormal distribution). I am wanting these estimates on the response scale and hence want to exponentiate them but can't see a transform = "exp" (or thereofs) in the accepted inputs for the `regrid` function. Thank you!
1.0
emmeans support - regrid using an 'exp' transform argument? - Hello, I'm wanting to regrid an emmGrid object that has `mu` estimates that was modelled as a natural log with an identity link function (using the shifted lognormal distribution). I am wanting these estimates on the response scale and hence want to exponentiate them but can't see a transform = "exp" (or thereofs) in the accepted inputs for the `regrid` function. Thank you!
process
emmeans support regrid using an exp transform argument hello i m wanting to regrid an emmgrid object that has mu estimates that was modelled as a natural log with an identity link function using the shifted lognormal distribution i am wanting these estimates on the response scale and hence want to exponentiate them but can t see a transform exp or thereofs in the accepted inputs for the regrid function thank you
1
8,760
11,879,957,637
IssuesEvent
2020-03-27 09:47:43
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
opened
Replace existing docs content with links to new Prisma 2 docs website
kind/docs process/candidate
See internal discussion: https://prisma-company.slack.com/archives/C5Z9TH6N9/p1585214876001200
1.0
Replace existing docs content with links to new Prisma 2 docs website - See internal discussion: https://prisma-company.slack.com/archives/C5Z9TH6N9/p1585214876001200
process
replace existing docs content with links to new prisma docs website see internal discussion
1
166,770
6,311,246,460
IssuesEvent
2017-07-23 17:46:58
OperationCode/operationcode_frontend
https://api.github.com/repos/OperationCode/operationcode_frontend
closed
Create /media page
in progress Needs: Copy/Content Priority: Medium Status: In Progress
# Feature ## Why is this feature being added? As referenced in #216, this page will be available for media to see how we've been portrayed in articles, photos, and videos. ## What should your feature do? - Implement a page called "Media" with the route being /press - Link the page in a dropdown in the header. - Page should display content described in linked wireframe, excepting the press releases section. Wireframe: https://xd.adobe.com/view/c1e499b5-a59b-430a-8f61-25b9dbcef715/
1.0
Create /media page - # Feature ## Why is this feature being added? As referenced in #216, this page will be available for media to see how we've been portrayed in articles, photos, and videos. ## What should your feature do? - Implement a page called "Media" with the route being /press - Link the page in a dropdown in the header. - Page should display content described in linked wireframe, excepting the press releases section. Wireframe: https://xd.adobe.com/view/c1e499b5-a59b-430a-8f61-25b9dbcef715/
non_process
create media page feature why is this feature being added as referenced in this page will be available for media to see how we ve been portrayed in articles photos and videos what should your feature do implement a page called media with the route being press link the page in a dropdown in the header page should display content described in linked wireframe excepting the press releases section wireframe
0
7,723
10,831,599,285
IssuesEvent
2019-11-11 08:43:03
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Proposal: remove EF Core provider for Spanner
type: process
We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider. The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort. Proposal: - Delist the packages in NuGet - Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back) cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa (Note: unless there are any objections, I plan to implement this on Friday November 8th.)
1.0
Proposal: remove EF Core provider for Spanner - We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider. The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort. Proposal: - Delist the packages in NuGet - Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back) cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa (Note: unless there are any objections, I plan to implement this on Friday November 8th.)
process
proposal remove ef core provider for spanner we currently have an alpha version of an entity framework core provider for spanner along with a ga version of an ado net provider the ef core provider has not been reviewed and maintained as thoroughly as the ado net provider and probably needs significant effort to bring it from alpha to beta and then ga at the moment we have no plans to put in that effort proposal delist the packages in nuget remove the source code from github in a single commit so we can easily revert it later if we decide to bring the project back cc meteatamel justinbeckwith chrisdunelm surferjeffatgoogle jsimonweb amanda tarafa note unless there are any objections i plan to implement this on friday november
1
12,512
14,962,376,246
IssuesEvent
2021-01-27 09:11:26
medic/cht-core
https://api.github.com/repos/medic/cht-core
closed
Release 3.10.2
Type: Internal process
# Planning - [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. - [x] Add all the issues to be worked on to the project. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [x] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` and backported to the release branch release testing can begin. - [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [x] Create a new document in the [release-notes folder](https://github.com/medic/cht-core/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/cht-core/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. - [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [x] Announce the release in #products using this template: ``` @channel *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the release notes for full details: {{url}} ``` - [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [x] Mark this issue "done" and close the project.
1.0
Release 3.10.2 - # Planning - [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. - [x] Add all the issues to be worked on to the project. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [x] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` and backported to the release branch release testing can begin. - [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [x] Create a new document in the [release-notes folder](https://github.com/medic/cht-core/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/cht-core/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. - [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [x] Announce the release in #products using this template: ``` @channel *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the release notes for full details: {{url}} ``` - [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [x] Mark this issue "done" and close the project.
process
release planning create an and add this issue to it add all the issues to be worked on to the project development when development is ready to begin one of the engineers should be nominated as a release manager they will be responsible for making sure the following tasks are completed though not necessarily completing them set the version number in package json and package lock json and submit a pr to the release branch the easiest way to do this is to use npm no git tag version version patch write an update in the weekly product team call agenda summarising development and acceptance testing progress and identifying any blockers the release manager is to update this every week until the version is released releasing once all issues have passed acceptance testing and have been merged into master and backported to the release branch release testing can begin build a beta named beta by pushing a git tag and when ci completes successfully notify the qa team that it s ready for release testing create a new document in the in master ensure all issues are in the gh project that they re correct labelled and have human readable descriptions use to export the issues into our changelog format manually document any known migration steps and known issues until release testing passes make sure regressions are fixed in master cherry pick them into the release branch and release another beta create a release in github from the release branch so it shows up under the with the naming convention this will create the git tag automatically link to the release notes in the description of the release confirm the release build completes successfully and the new release is available on the make sure that the document has new entry with id medic medic announce the release in products using this template channel announcing the release of version this release fixes number of bugs read the release notes for full details url announce the release on the under the product releases category you can use the previous message and omit channel mark this issue done and close the project
1
172,583
6,510,542,539
IssuesEvent
2017-08-25 04:15:10
CanberraOceanRacingClub/namadgi3
https://api.github.com/repos/CanberraOceanRacingClub/namadgi3
closed
Buy some drop sheets for workers to use while working on the boat
priority 1: High Working bee
The saloon cushion covers and berth mattress covers are all near-white colours. They readily show dirt and marks. We need a set of drop-sheets and protective covers that contractors (and CORC members) must use while doing maintenance work on the boat.
1.0
Buy some drop sheets for workers to use while working on the boat - The saloon cushion covers and berth mattress covers are all near-white colours. They readily show dirt and marks. We need a set of drop-sheets and protective covers that contractors (and CORC members) must use while doing maintenance work on the boat.
non_process
buy some drop sheets for workers to use while working on the boat the saloon cushion covers and berth mattress covers are all near white colours they readily show dirt and marks we need a set of drop sheets and protective covers that contractors and corc members must use while doing maintenance work on the boat
0
9,563
12,519,060,409
IssuesEvent
2020-06-03 13:54:21
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Force scaffolding in headless mode
process: tests stage: needs review
We often need to test Cypress, and I have been using `@bahmutov/cly init` to scaffold a quick example using https://github.com/bahmutov/cly But we also need to test Cypress more against Kitchensink examples to make sure we can update the Kitchensink pages safely. For this we need to scaffold full integration examples. Right now this only happens in the interactive mode, but we need to do this in headless mode as well. Looking at the code in `packages/server/src/project.js` it seems we could use an environment variable to force scaffolding all folders (right now there is logic to always scaffold some folders, but integration specs are skipped in headless mode) Will allow https://github.com/cypress-io/cypress-example-kitchensink/issues/420 to work
1.0
Force scaffolding in headless mode - We often need to test Cypress, and I have been using `@bahmutov/cly init` to scaffold a quick example using https://github.com/bahmutov/cly But we also need to test Cypress more against Kitchensink examples to make sure we can update the Kitchensink pages safely. For this we need to scaffold full integration examples. Right now this only happens in the interactive mode, but we need to do this in headless mode as well. Looking at the code in `packages/server/src/project.js` it seems we could use an environment variable to force scaffolding all folders (right now there is logic to always scaffold some folders, but integration specs are skipped in headless mode) Will allow https://github.com/cypress-io/cypress-example-kitchensink/issues/420 to work
process
force scaffolding in headless mode we often need to test cypress and i have been using bahmutov cly init to scaffold a quick example using but we also need to test cypress more against kitchensink examples to make sure we can update the kitchensink pages safely for this we need to scaffold full integration examples right now this only happens in the interactive mode but we need to do this in headless mode as well looking at the code in packages server src project js it seems we could use an environment variable to force scaffolding all folders right now there is logic to always scaffold some folders but integration specs are skipped in headless mode will allow to work
1
12,252
14,785,823,635
IssuesEvent
2021-01-12 03:47:20
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
GRASS/Processing outputs CRS mismatch
Bug Processing Projections/Transformations
Author Name: **Paolo Cavallini** (@pcav) Original Redmine Issue: [10137](https://issues.qgis.org/issues/10137) Affected QGIS version: 3.4.1 Redmine category:processing/grass --- Apparently Processing, during the analyses starting from layers in a CRS with +towgs parameters (e.g. 23030) creates outputs with a similar projection, but without the parameters; a custom CRS is thus created, and the output does not align with the input. This happens at least with GRASS and R backends. --- Related issue(s): #20098 (relates), #25391 (duplicates) Redmine related issue(s): [11884](https://issues.qgis.org/issues/11884), [17494](https://issues.qgis.org/issues/17494) ---
1.0
GRASS/Processing outputs CRS mismatch - Author Name: **Paolo Cavallini** (@pcav) Original Redmine Issue: [10137](https://issues.qgis.org/issues/10137) Affected QGIS version: 3.4.1 Redmine category:processing/grass --- Apparently Processing, during the analyses starting from layers in a CRS with +towgs parameters (e.g. 23030) creates outputs with a similar projection, but without the parameters; a custom CRS is thus created, and the output does not align with the input. This happens at least with GRASS and R backends. --- Related issue(s): #20098 (relates), #25391 (duplicates) Redmine related issue(s): [11884](https://issues.qgis.org/issues/11884), [17494](https://issues.qgis.org/issues/17494) ---
process
grass processing outputs crs mismatch author name paolo cavallini pcav original redmine issue affected qgis version redmine category processing grass apparently processing during the analyses starting from layers in a crs with towgs parameters e g creates outputs with a similar projection but without the parameters a custom crs is thus created and the output does not align with the input this happens at least with grass and r backends related issue s relates duplicates redmine related issue s
1
98,615
12,341,679,788
IssuesEvent
2020-05-14 22:31:14
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
SVG <text> tag flags an error in one case, but not in another.
Needs: Design area-blazor feature-svg investigate
The following razor code flags a compile error that the `<text> `shouldn't have any attributes because it is interpreting `<text>` as the razor literal text tag. ``` @foreach (var designObject in ToolService.DesignObjects) { switch(designObject) { case ImageObjectModel imageObjectModel: break; case TextObjectModel textObjectModel: <text @x="@textObjectModel.Coordinate.X" @y="@textObjectModel.Coordinate.Y" class="small"></text> break; } } ``` On another component's razor code, the following doesn't flag an error: ``` <text class="text" x="@Coordinate.X" y="@Coordinate.Y" >@CustomText</text> @code { [Parameter] public string CustomText { get; set; } = "Custom Text"; public string FontStyle { get; } } ``` Why isn't razor interpreting the `<text> `tag the same way in both cases?
1.0
SVG <text> tag flags an error in one case, but not in another. - The following razor code flags a compile error that the `<text> `shouldn't have any attributes because it is interpreting `<text>` as the razor literal text tag. ``` @foreach (var designObject in ToolService.DesignObjects) { switch(designObject) { case ImageObjectModel imageObjectModel: break; case TextObjectModel textObjectModel: <text @x="@textObjectModel.Coordinate.X" @y="@textObjectModel.Coordinate.Y" class="small"></text> break; } } ``` On another component's razor code, the following doesn't flag an error: ``` <text class="text" x="@Coordinate.X" y="@Coordinate.Y" >@CustomText</text> @code { [Parameter] public string CustomText { get; set; } = "Custom Text"; public string FontStyle { get; } } ``` Why isn't razor interpreting the `<text> `tag the same way in both cases?
non_process
svg tag flags an error in one case but not in another the following razor code flags a compile error that the shouldn t have any attributes because it is interpreting as the razor literal text tag foreach var designobject in toolservice designobjects switch designobject case imageobjectmodel imageobjectmodel break case textobjectmodel textobjectmodel break on another component s razor code the following doesn t flag an error customtext code public string customtext get set custom text public string fontstyle get why isn t razor interpreting the tag the same way in both cases
0
10,709
13,506,463,362
IssuesEvent
2020-09-14 03:04:44
jayspur11/Arbeiterbiene
https://api.github.com/repos/jayspur11/Arbeiterbiene
closed
Create base class for web fetchers
process
Things like `meow` and `aqi` that depend on making a request / having an API key -- those should inherit from a common class so they can share error-handling logic.
1.0
Create base class for web fetchers - Things like `meow` and `aqi` that depend on making a request / having an API key -- those should inherit from a common class so they can share error-handling logic.
process
create base class for web fetchers things like meow and aqi that depend on making a request having an api key those should inherit from a common class so they can share error handling logic
1
20,831
27,591,232,397
IssuesEvent
2023-03-09 00:38:53
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[REMOTO] [QA] Tester / QA na [VÓRTX]
HOME OFFICE REMOTO AZURE TFS QA METODOLOGIAS ÁGEIS HELP WANTED AUTOMAÇÃO DE PROCESSOS Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Estamos em busca de pessoas que amam tecnologia, tem foco no cliente e está sempre em busca de aprendizado contínuo. E sobretudo, engajadas em fazer parte desta incrível jornada que é descomplicar o mercado de capitais. **Responsabilidades:** - Acompanhar os gerentes de produtos nas definições das regras de negócio; - Planejar e executar cenários de testes; - Acompanhar a qualidade das entregas junto às áreas de negócio; - Criação de testes automatizados. ## Local - Home Office ## Benefícios - Plano de saúde e odontológico Bradesco; - Auxílio creche (até R$ 314,00 - para filhos até 5 anos); - Vale refeição (R$ 32,00/dia) e Vale alimentação (R$ 466,00) em cartão flexível para utilização; - Vale transporte ou Ajuda de custo home office (R$ 150,00); - Ambiente colaborativo; - Flexibilidade de horários; - Nada de Dress Code; - PLR – de acordo com política interna; - Subsídio para cursos/treinamento e pós-graduação - de acordo com política interna; - Táxi após as 22h para voltar pra casa - de acordo com política interna. ## Requisitos **Obrigatórios:** - Sólido conhecimento em Testes (Preferencialmente do Mercado financeiro); - Arquitetura / Aplicações web; - Automação e Processos. **Desejáveis:** - Metodologia Ágil; - TFS / Azure DevOps - Inglês ## Contratação - a combinar ## Nossa empresa - A Vórtx® é a primeira fintech B2B de infraestrutura do mercado de capitais, utilizando tecnologia para viabilizar as transações de investimento. Ou seja, somos as pessoas especialistas por trás de toda operação de investimentos de confiança, como um sistema operacional - o iOS/Android do mercado de capitais. ## Como se candidatar - [Clique aqui para se candidatar](https://jobs.kenoby.com/vortx/job/tester-qa/6009981f098e58070440be51?utm_source=website)
1.0
[REMOTO] [QA] Tester / QA na [VÓRTX] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Estamos em busca de pessoas que amam tecnologia, tem foco no cliente e está sempre em busca de aprendizado contínuo. E sobretudo, engajadas em fazer parte desta incrível jornada que é descomplicar o mercado de capitais. **Responsabilidades:** - Acompanhar os gerentes de produtos nas definições das regras de negócio; - Planejar e executar cenários de testes; - Acompanhar a qualidade das entregas junto às áreas de negócio; - Criação de testes automatizados. ## Local - Home Office ## Benefícios - Plano de saúde e odontológico Bradesco; - Auxílio creche (até R$ 314,00 - para filhos até 5 anos); - Vale refeição (R$ 32,00/dia) e Vale alimentação (R$ 466,00) em cartão flexível para utilização; - Vale transporte ou Ajuda de custo home office (R$ 150,00); - Ambiente colaborativo; - Flexibilidade de horários; - Nada de Dress Code; - PLR – de acordo com política interna; - Subsídio para cursos/treinamento e pós-graduação - de acordo com política interna; - Táxi após as 22h para voltar pra casa - de acordo com política interna. ## Requisitos **Obrigatórios:** - Sólido conhecimento em Testes (Preferencialmente do Mercado financeiro); - Arquitetura / Aplicações web; - Automação e Processos. **Desejáveis:** - Metodologia Ágil; - TFS / Azure DevOps - Inglês ## Contratação - a combinar ## Nossa empresa - A Vórtx® é a primeira fintech B2B de infraestrutura do mercado de capitais, utilizando tecnologia para viabilizar as transações de investimento. Ou seja, somos as pessoas especialistas por trás de toda operação de investimentos de confiança, como um sistema operacional - o iOS/Android do mercado de capitais. ## Como se candidatar - [Clique aqui para se candidatar](https://jobs.kenoby.com/vortx/job/tester-qa/6009981f098e58070440be51?utm_source=website)
process
tester qa na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga estamos em busca de pessoas que amam tecnologia tem foco no cliente e está sempre em busca de aprendizado contínuo e sobretudo engajadas em fazer parte desta incrível jornada que é descomplicar o mercado de capitais responsabilidades acompanhar os gerentes de produtos nas definições das regras de negócio planejar e executar cenários de testes acompanhar a qualidade das entregas junto às áreas de negócio criação de testes automatizados local home office benefícios plano de saúde e odontológico bradesco auxílio creche até r para filhos até anos vale refeição r dia e vale alimentação r em cartão flexível para utilização vale transporte ou ajuda de custo home office r ambiente colaborativo flexibilidade de horários nada de dress code plr – de acordo com política interna subsídio para cursos treinamento e pós graduação de acordo com política interna táxi após as para voltar pra casa de acordo com política interna requisitos obrigatórios sólido conhecimento em testes preferencialmente do mercado financeiro arquitetura aplicações web automação e processos desejáveis metodologia ágil tfs azure devops inglês contratação a combinar nossa empresa a vórtx® é a primeira fintech de infraestrutura do mercado de capitais utilizando tecnologia para viabilizar as transações de investimento ou seja somos as pessoas especialistas por trás de toda operação de investimentos de confiança como um sistema operacional o ios android do mercado de capitais como se candidatar
1
276,801
21,001,009,896
IssuesEvent
2022-03-29 17:26:42
napari/napari
https://api.github.com/repos/napari/napari
closed
Windows Bundled App - Command Prompt on Open
documentation bundle
When opening up napari from a bundled install, the user first sees a command prompt window which can hang for a while as the viewer opens. ![image](https://user-images.githubusercontent.com/17995243/96401557-6b650200-121f-11eb-9fab-ff1d818a8d74.png) This is mentioned briefly in the installation tutorial. ![image](https://user-images.githubusercontent.com/17995243/96401581-80419580-121f-11eb-9e1a-ef21905caa25.png) Today I observed a technically inexperienced user go through the installation process. The user did not notice this section in the tutorial, and when the command prompt opened there were two reactions: 1. Concern that she would have to type code into the command prompt 2. Misunderstanding that the command prompt window was actually napari and there was nothing left to load I think we can improve the documentation by including a screenshot of the window in the tutorial, which might be more eye catching than the text alone. I ideally though, the command prompt window itself could say something such as "Loading napari..." or anything that tells the user to wait. I'm not sure whether this is possible, however.
1.0
Windows Bundled App - Command Prompt on Open - When opening up napari from a bundled install, the user first sees a command prompt window which can hang for a while as the viewer opens. ![image](https://user-images.githubusercontent.com/17995243/96401557-6b650200-121f-11eb-9fab-ff1d818a8d74.png) This is mentioned briefly in the installation tutorial. ![image](https://user-images.githubusercontent.com/17995243/96401581-80419580-121f-11eb-9e1a-ef21905caa25.png) Today I observed a technically inexperienced user go through the installation process. The user did not notice this section in the tutorial, and when the command prompt opened there were two reactions: 1. Concern that she would have to type code into the command prompt 2. Misunderstanding that the command prompt window was actually napari and there was nothing left to load I think we can improve the documentation by including a screenshot of the window in the tutorial, which might be more eye catching than the text alone. I ideally though, the command prompt window itself could say something such as "Loading napari..." or anything that tells the user to wait. I'm not sure whether this is possible, however.
non_process
windows bundled app command prompt on open when opening up napari from a bundled install the user first sees a command prompt window which can hang for a while as the viewer opens this is mentioned briefly in the installation tutorial today i observed a technically inexperienced user go through the installation process the user did not notice this section in the tutorial and when the command prompt opened there were two reactions concern that she would have to type code into the command prompt misunderstanding that the command prompt window was actually napari and there was nothing left to load i think we can improve the documentation by including a screenshot of the window in the tutorial which might be more eye catching than the text alone i ideally though the command prompt window itself could say something such as loading napari or anything that tells the user to wait i m not sure whether this is possible however
0
74,531
9,791,024,829
IssuesEvent
2019-06-10 14:09:43
SergeyKurmenev/elama-exam
https://api.github.com/repos/SergeyKurmenev/elama-exam
opened
Добавление документации к API
documentation swagger составная задача
Для создания документации - будет использоваться `swagger` (`flask-restful-swagger`). Задачи: * Добавить `flask-restful-swagger` в `requirements.txt` * Добавить `swagger` обёртку вокруг `api`
1.0
Добавление документации к API - Для создания документации - будет использоваться `swagger` (`flask-restful-swagger`). Задачи: * Добавить `flask-restful-swagger` в `requirements.txt` * Добавить `swagger` обёртку вокруг `api`
non_process
добавление документации к api для создания документации будет использоваться swagger flask restful swagger задачи добавить flask restful swagger в requirements txt добавить swagger обёртку вокруг api
0
23,675
3,851,865,315
IssuesEvent
2016-04-06 05:27:47
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
8bitty trouble + multiplayer future
auto-migrated Priority-Medium Type-Defect
``` ios 6.1, running imame4all v1.11 - have blutrol installed with icade working flawlessly. 8 bitty no luck there is no 8bitty profile in blutrol. Is there an existing work around to get 8bitty controller to work with imame4ios? also is there any future plans to map multiplayer buttons to the screen? if there could be two sets of buttons and joysticks that are mapped on the screen, I could then use blutrol to map two different blutooth controllers to it... and well, that would be a game changer. Connect ipad to TV, two controllers would mean ultimate home gaming emulator machine! Appreciate the support kind regards ``` Original issue reported on code.google.com by `juliocar...@gmail.com` on 2 Mar 2013 at 4:14
1.0
8bitty trouble + multiplayer future - ``` ios 6.1, running imame4all v1.11 - have blutrol installed with icade working flawlessly. 8 bitty no luck there is no 8bitty profile in blutrol. Is there an existing work around to get 8bitty controller to work with imame4ios? also is there any future plans to map multiplayer buttons to the screen? if there could be two sets of buttons and joysticks that are mapped on the screen, I could then use blutrol to map two different blutooth controllers to it... and well, that would be a game changer. Connect ipad to TV, two controllers would mean ultimate home gaming emulator machine! Appreciate the support kind regards ``` Original issue reported on code.google.com by `juliocar...@gmail.com` on 2 Mar 2013 at 4:14
non_process
trouble multiplayer future ios running have blutrol installed with icade working flawlessly bitty no luck there is no profile in blutrol is there an existing work around to get controller to work with also is there any future plans to map multiplayer buttons to the screen if there could be two sets of buttons and joysticks that are mapped on the screen i could then use blutrol to map two different blutooth controllers to it and well that would be a game changer connect ipad to tv two controllers would mean ultimate home gaming emulator machine appreciate the support kind regards original issue reported on code google com by juliocar gmail com on mar at
0
18,089
24,112,332,547
IssuesEvent
2022-09-20 12:23:36
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
how to switch to compiling with g++.exe from microsoft cl.exe?
type: support / not a bug (process) area-Windows team-OSS
hey, there! when i tried to build the tcmalloc/testing:hello_main target, cl.exe was used and an error: D8021: invalid parameter “/Werror” appeared. so, i want to switch to g++, instead of cl. i've installed MinGW-w64, and i could use g++ in terminal or CMD. i tried reinstall bazel but it does't work. bazel still uses cl.exe. if anyone knows how to switch, please tell me or just post a link, thanks very much.
1.0
how to switch to compiling with g++.exe from microsoft cl.exe? - hey, there! when i tried to build the tcmalloc/testing:hello_main target, cl.exe was used and an error: D8021: invalid parameter “/Werror” appeared. so, i want to switch to g++, instead of cl. i've installed MinGW-w64, and i could use g++ in terminal or CMD. i tried reinstall bazel but it does't work. bazel still uses cl.exe. if anyone knows how to switch, please tell me or just post a link, thanks very much.
process
how to switch to compiling with g exe from microsoft cl exe hey there when i tried to build the tcmalloc testing hello main target cl exe was used and an error invalid parameter “ werror” appeared so i want to switch to g instead of cl i ve installed mingw and i could use g in terminal or cmd i tried reinstall bazel but it does t work bazel still uses cl exe if anyone knows how to switch please tell me or just post a link thanks very much
1
5,844
8,671,027,698
IssuesEvent
2018-11-29 17:59:56
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
closed
Incorrect retrieval of computer in `JobCalculation._presubmit`
priority/important topic/JobCalculationAndProcess type/bug
The implementation of `_presubmit` currently tries to retrieve a computer from a given UUID, necessary for the remote copy lists, through the backend, like `self.backend.computers.get`, except that backend `ComputerCollection` does not have a `get` method.
1.0
Incorrect retrieval of computer in `JobCalculation._presubmit` - The implementation of `_presubmit` currently tries to retrieve a computer from a given UUID, necessary for the remote copy lists, through the backend, like `self.backend.computers.get`, except that backend `ComputerCollection` does not have a `get` method.
process
incorrect retrieval of computer in jobcalculation presubmit the implementation of presubmit currently tries to retrieve a computer from a given uuid necessary for the remote copy lists through the backend like self backend computers get except that backend computercollection does not have a get method
1
102,384
4,155,187,312
IssuesEvent
2016-06-16 14:13:52
jpppina/migracion-galeno-art-forms11g
https://api.github.com/repos/jpppina/migracion-galeno-art-forms11g
opened
ART - NO aparece la pantalla ARLPI029
Aplicación-ART Error Priority-Low
No aparece la pantalla Usuario: H20348 pass: desaa002 Ruta: ARTManager --> Cuentas a Pagar --> Tablas de cuentas a pagar --> Motivos y Soluciones de Anulaciones
1.0
ART - NO aparece la pantalla ARLPI029 - No aparece la pantalla Usuario: H20348 pass: desaa002 Ruta: ARTManager --> Cuentas a Pagar --> Tablas de cuentas a pagar --> Motivos y Soluciones de Anulaciones
non_process
art no aparece la pantalla no aparece la pantalla usuario pass ruta artmanager cuentas a pagar tablas de cuentas a pagar motivos y soluciones de anulaciones
0
19,672
26,031,311,248
IssuesEvent
2022-12-21 21:36:53
phytooracle/automation
https://api.github.com/repos/phytooracle/automation
closed
S11 postprocessing pipeling container #4 3d_individual_plant_registration.simg
wontfix Season 11 S11_postprocessing
3d_individual_plant_registration.simg
1.0
S11 postprocessing pipeling container #4 3d_individual_plant_registration.simg - 3d_individual_plant_registration.simg
process
postprocessing pipeling container individual plant registration simg individual plant registration simg
1
13,934
4,790,660,845
IssuesEvent
2016-10-31 09:33:51
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Module Position parameter not working with admin template Hathor
No Code Attached Yet
### Steps to reproduce the issue Set Hathor as admin template. Edit or create a module and try to set a module position. ### Expected result Opens a drop down menu to select the position ### Actual result The drop down of the parameter Position doeasn't open. If you click on Select a list of module positions appears, but if you click on the position links nothing happens.
1.0
Module Position parameter not working with admin template Hathor - ### Steps to reproduce the issue Set Hathor as admin template. Edit or create a module and try to set a module position. ### Expected result Opens a drop down menu to select the position ### Actual result The drop down of the parameter Position doeasn't open. If you click on Select a list of module positions appears, but if you click on the position links nothing happens.
non_process
module position parameter not working with admin template hathor steps to reproduce the issue set hathor as admin template edit or create a module and try to set a module position expected result opens a drop down menu to select the position actual result the drop down of the parameter position doeasn t open if you click on select a list of module positions appears but if you click on the position links nothing happens
0
234,983
18,024,910,805
IssuesEvent
2021-09-17 02:18:37
JDAI-CV/fast-reid
https://api.github.com/repos/JDAI-CV/fast-reid
closed
Distillation Tutorial
documentation stale
This guide explains how to apply the model distillation technique to a fastreid model. ## Before You Start Following **Getting Started** to setup the environment and install requirements.txt dependencies. ## Train Teacher Model Go to [projects/FastDistill](https://github.com/JDAI-CV/fast-reid/tree/master/projects/FastDistill), then you can find the configuration example for `DukeMTMC` dataset in `configs/Base-kd`. You can train `ibn101` as the teacher model with 4 gpus by command line below ```bash # in the root directory of projects python3 projects/FastDisill/train_net.py --config-file projects/FastDisillDistill/configs/sbs_r101ibn.yml --num-gpus 4 ``` The `OUTPUT_DIR` in config file shows the path of the trained model. Here we set it `projects/FastDistill/logs/dukemtmc/r101-ibn` as default. ## Train Distilled Model After finishing the teacher model training, you can perform model distillation. In the config file, you need to change `MODEL.META_ARCHITECTURE` from `Baseline` to `Distiller` to turn on distillation mode. Then you should set the teacher models config file and weights sperately. YOu can set `KD.MODEL_CONFIG` to the path of trained teacher model configuration file and `KD.MODEL_WEIGHTS` to the path of trained teacher model weights. If you want to use `ibn101` as teacher model and `r34` as a student model, you can run the command below ```bash python3 projects/FastDistill/train_net.py --config-file projects/DistillReID/configs/kd-sbs_r101-ibn-sbs_r34.yml ``` You can find the distilled model in `OUTPUT_DIR` in kd config file. If you want to train other datasets or use other model architectures, you can refer to this example and change the dataset and model names.
1.0
Distillation Tutorial - This guide explains how to apply the model distillation technique to a fastreid model. ## Before You Start Following **Getting Started** to setup the environment and install requirements.txt dependencies. ## Train Teacher Model Go to [projects/FastDistill](https://github.com/JDAI-CV/fast-reid/tree/master/projects/FastDistill), then you can find the configuration example for `DukeMTMC` dataset in `configs/Base-kd`. You can train `ibn101` as the teacher model with 4 gpus by command line below ```bash # in the root directory of projects python3 projects/FastDisill/train_net.py --config-file projects/FastDisillDistill/configs/sbs_r101ibn.yml --num-gpus 4 ``` The `OUTPUT_DIR` in config file shows the path of the trained model. Here we set it `projects/FastDistill/logs/dukemtmc/r101-ibn` as default. ## Train Distilled Model After finishing the teacher model training, you can perform model distillation. In the config file, you need to change `MODEL.META_ARCHITECTURE` from `Baseline` to `Distiller` to turn on distillation mode. Then you should set the teacher models config file and weights sperately. YOu can set `KD.MODEL_CONFIG` to the path of trained teacher model configuration file and `KD.MODEL_WEIGHTS` to the path of trained teacher model weights. If you want to use `ibn101` as teacher model and `r34` as a student model, you can run the command below ```bash python3 projects/FastDistill/train_net.py --config-file projects/DistillReID/configs/kd-sbs_r101-ibn-sbs_r34.yml ``` You can find the distilled model in `OUTPUT_DIR` in kd config file. If you want to train other datasets or use other model architectures, you can refer to this example and change the dataset and model names.
non_process
distillation tutorial this guide explains how to apply the model distillation technique to a fastreid model before you start following getting started to setup the environment and install requirements txt dependencies train teacher model go to then you can find the configuration example for dukemtmc dataset in configs base kd you can train as the teacher model with gpus by command line below bash in the root directory of projects projects fastdisill train net py config file projects fastdisilldistill configs sbs yml num gpus the output dir in config file shows the path of the trained model here we set it projects fastdistill logs dukemtmc ibn as default train distilled model after finishing the teacher model training you can perform model distillation in the config file you need to change model meta architecture from baseline to distiller to turn on distillation mode then you should set the teacher models config file and weights sperately you can set kd model config to the path of trained teacher model configuration file and kd model weights to the path of trained teacher model weights if you want to use as teacher model and as a student model you can run the command below bash projects fastdistill train net py config file projects distillreid configs kd sbs ibn sbs yml you can find the distilled model in output dir in kd config file if you want to train other datasets or use other model architectures you can refer to this example and change the dataset and model names
0
19,810
26,201,171,756
IssuesEvent
2023-01-03 17:36:16
plazi/treatmentBank
https://api.github.com/repos/plazi/treatmentBank
opened
BigBatch processing: issues to cover
processing
Batch processing of all the treatments should include the following issues * Accession numbers * SubSubSection types normalization * vernacularName * COL taxon names ID and cited taxa
1.0
BigBatch processing: issues to cover - Batch processing of all the treatments should include the following issues * Accession numbers * SubSubSection types normalization * vernacularName * COL taxon names ID and cited taxa
process
bigbatch processing issues to cover batch processing of all the treatments should include the following issues accession numbers subsubsection types normalization vernacularname col taxon names id and cited taxa
1
4,461
7,329,952,690
IssuesEvent
2018-03-05 08:04:34
UKHomeOffice/dq-aws-transition
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
closed
Add ACL FTP crontab for Data Ingest Linux
DQ Data Ingest DQ Tranche 1 Production SSM processing
# ACL from ACL FTP server - [x] */10 0 * * * /ADT/scripts/ftp_acl_web02.py
1.0
Add ACL FTP crontab for Data Ingest Linux - # ACL from ACL FTP server - [x] */10 0 * * * /ADT/scripts/ftp_acl_web02.py
process
add acl ftp crontab for data ingest linux acl from acl ftp server adt scripts ftp acl py
1
2,461
5,241,274,330
IssuesEvent
2017-01-31 15:20:46
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Problem in hra processor with isrctn id
bug Launch? Processors
1-2 cases per 100 ``` Processing error: TypeError('expected string or buffer',) [8] Traceback (most recent call last): File "processors/base/processors/trial.py", line 38, in process_trial trial = extractors['extract_trial'](record) File "processors/hra/extractors.py", line 27, in extract_trial 'isrctn': _clean_identifier(record['isrctn_id'], prefix='ISRCTN'), File "processors/hra/extractors.py", line 93, in _clean_identifier if re.match(r'%s\d{3,}' % prefix, ident): File "/usr/local/lib/python2.7/re.py", line 141, in match return _compile(pattern, flags).match(string) TypeError: expected string or buffer ```
1.0
Problem in hra processor with isrctn id - 1-2 cases per 100 ``` Processing error: TypeError('expected string or buffer',) [8] Traceback (most recent call last): File "processors/base/processors/trial.py", line 38, in process_trial trial = extractors['extract_trial'](record) File "processors/hra/extractors.py", line 27, in extract_trial 'isrctn': _clean_identifier(record['isrctn_id'], prefix='ISRCTN'), File "processors/hra/extractors.py", line 93, in _clean_identifier if re.match(r'%s\d{3,}' % prefix, ident): File "/usr/local/lib/python2.7/re.py", line 141, in match return _compile(pattern, flags).match(string) TypeError: expected string or buffer ```
process
problem in hra processor with isrctn id cases per processing error typeerror expected string or buffer traceback most recent call last file processors base processors trial py line in process trial trial extractors record file processors hra extractors py line in extract trial isrctn clean identifier record prefix isrctn file processors hra extractors py line in clean identifier if re match r s d prefix ident file usr local lib re py line in match return compile pattern flags match string typeerror expected string or buffer
1
12,197
19,286,159,761
IssuesEvent
2021-12-11 01:50:28
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Renovate specifying non-existing PyPI version
type:bug datasource:pypi reproduction:needed status:requirements priority-5-triage stale
### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 27.19.2 ### Please select which platform you are using if self-hosting. github.com ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Describe the bug We are self-hosting Renovate and found an issue with one of the dependency updates. Renovate wanted to update the package `PyMuPDF==1.19.1` to `PyMuPDF==1.19.1.1`, which does not [exist](https://pypi.org/project/PyMuPDF/#history). We are using a private AWS Codeartifact Python Index with the public PyPI as upstream repository. All those repositories did not contain the unknown version. The logs below show where renovate wanted to increase the version. The bug happens again after specifying the correct version, so its not some kind of repository fluke. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: packageFiles with updates (repository=repo123) "config": { "pip_requirements": [ { "packageFile": "requirements.txt", "deps": [ { "depName": "PyMuPDF", "currentValue": "==1.19.1", "datasource": "pypi", "currentVersion": "1.19.1", "depIndex": 8, "updates": [ { "bucket": "non-major", "newVersion": "1.19.1.1", "newValue": "==1.19.1.1", "newMajor": 1, "newMinor": 19, "updateType": "patch", "isRange": true, "branchName": "renovate/pymupdf-1.x" } ], "warnings": [], "versioning": "pep440", "isSingleVersion": true, "fixedVersion": "1.19.1" } ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
1.0
Renovate specifying non-existing PyPI version - ### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 27.19.2 ### Please select which platform you are using if self-hosting. github.com ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Describe the bug We are self-hosting Renovate and found an issue with one of the dependency updates. Renovate wanted to update the package `PyMuPDF==1.19.1` to `PyMuPDF==1.19.1.1`, which does not [exist](https://pypi.org/project/PyMuPDF/#history). We are using a private AWS Codeartifact Python Index with the public PyPI as upstream repository. All those repositories did not contain the unknown version. The logs below show where renovate wanted to increase the version. The bug happens again after specifying the correct version, so its not some kind of repository fluke. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: packageFiles with updates (repository=repo123) "config": { "pip_requirements": [ { "packageFile": "requirements.txt", "deps": [ { "depName": "PyMuPDF", "currentValue": "==1.19.1", "datasource": "pypi", "currentVersion": "1.19.1", "depIndex": 8, "updates": [ { "bucket": "non-major", "newVersion": "1.19.1.1", "newValue": "==1.19.1.1", "newMajor": 1, "newMinor": 19, "updateType": "patch", "isRange": true, "branchName": "renovate/pymupdf-1.x" } ], "warnings": [], "versioning": "pep440", "isSingleVersion": true, "fixedVersion": "1.19.1" } ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
non_process
renovate specifying non existing pypi version how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run please select which platform you are using if self hosting github com if you re self hosting renovate tell us what version of the platform you run no response describe the bug we are self hosting renovate and found an issue with one of the dependency updates renovate wanted to update the package pymupdf to pymupdf which does not we are using a private aws codeartifact python index with the public pypi as upstream repository all those repositories did not contain the unknown version the logs below show where renovate wanted to increase the version the bug happens again after specifying the correct version so its not some kind of repository fluke relevant debug logs logs debug packagefiles with updates repository config pip requirements packagefile requirements txt deps depname pymupdf currentvalue datasource pypi currentversion depindex updates bucket non major newversion newvalue newmajor newminor updatetype patch isrange true branchname renovate pymupdf x warnings versioning issingleversion true fixedversion have you created a minimal reproduction repository no reproduction repository
0
595,601
18,070,134,152
IssuesEvent
2021-09-21 01:15:14
woocommerce/woocommerce-gateway-stripe
https://api.github.com/repos/woocommerce/woocommerce-gateway-stripe
closed
[old settings] "Requires currency" subtext
priority: low component: UPE settings
## Description Display message next to payment methods that require a different currency than the one set in the main settings. ![image (1)](https://user-images.githubusercontent.com/273592/133826842-f6422c3d-419a-47a0-9a05-94d5b42d47dc.png)
1.0
[old settings] "Requires currency" subtext - ## Description Display message next to payment methods that require a different currency than the one set in the main settings. ![image (1)](https://user-images.githubusercontent.com/273592/133826842-f6422c3d-419a-47a0-9a05-94d5b42d47dc.png)
non_process
requires currency subtext description display message next to payment methods that require a different currency than the one set in the main settings
0
15,702
19,848,397,889
IssuesEvent
2022-01-21 09:31:07
ooi-data/RS03ASHS-MJ03B-09-BOTPTA304-streamed-botpt_nano_sample
https://api.github.com/repos/ooi-data/RS03ASHS-MJ03B-09-BOTPTA304-streamed-botpt_nano_sample
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:31:06.652844. ## Details Flow name: `RS03ASHS-MJ03B-09-BOTPTA304-streamed-botpt_nano_sample` Task name: `processing_task` Error type: `ValueError` Error message: cannot reshape array of size 1209600 into shape (25000000,) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2305, in append return self._write_op(self._append_nosync, data, axis=axis) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2211, in _write_op return self._synchronized_op(f, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2201, in _synchronized_op result = f(*args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2341, in _append_nosync self[append_selection] = data File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1224, in __setitem__ self.set_basic_selection(selection, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1319, in set_basic_selection return self._set_basic_selection_nd(selection, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1610, in _set_basic_selection_nd self._set_selection(indexer, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1682, in _set_selection self._chunk_setitems(lchunk_coords, lchunk_selection, chunk_values, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1871, in _chunk_setitems cdatas = [self._process_for_setitem(key, sel, val, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1871, in <listcomp> cdatas = [self._process_for_setitem(key, sel, val, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1950, in _process_for_setitem chunk = self._decode_chunk(cdata) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2003, in _decode_chunk chunk = chunk.reshape(expected_shape or self._chunks, order=self._order) ValueError: cannot reshape array of size 1209600 into shape (25000000,) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:31:06.652844. ## Details Flow name: `RS03ASHS-MJ03B-09-BOTPTA304-streamed-botpt_nano_sample` Task name: `processing_task` Error type: `ValueError` Error message: cannot reshape array of size 1209600 into shape (25000000,) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2305, in append return self._write_op(self._append_nosync, data, axis=axis) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2211, in _write_op return self._synchronized_op(f, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2201, in _synchronized_op result = f(*args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2341, in _append_nosync self[append_selection] = data File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1224, in __setitem__ self.set_basic_selection(selection, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1319, in set_basic_selection return self._set_basic_selection_nd(selection, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1610, in _set_basic_selection_nd self._set_selection(indexer, value, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1682, in _set_selection self._chunk_setitems(lchunk_coords, lchunk_selection, chunk_values, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1871, in _chunk_setitems cdatas = [self._process_for_setitem(key, sel, val, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1871, in <listcomp> cdatas = [self._process_for_setitem(key, sel, val, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1950, in _process_for_setitem chunk = self._decode_chunk(cdata) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 2003, in _decode_chunk chunk = chunk.reshape(expected_shape or self._chunks, order=self._order) ValueError: cannot reshape array of size 1209600 into shape (25000000,) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name streamed botpt nano sample task name processing task error type valueerror error message cannot reshape array of size into shape traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages zarr core py line in append return self write op self append nosync data axis axis file srv conda envs notebook lib site packages zarr core py line in write op return self synchronized op f args kwargs file srv conda envs notebook lib site packages zarr core py line in synchronized op result f args kwargs file srv conda envs notebook lib site packages zarr core py line in append nosync self data file srv conda envs notebook lib site packages zarr core py line in setitem self set basic selection selection value fields fields file srv conda envs notebook lib site packages zarr core py line in set basic selection return self set basic selection nd selection value fields fields file srv conda envs notebook lib site packages zarr core py line in set basic selection nd self set selection indexer value fields fields file srv conda envs notebook lib site packages zarr core py line in set selection self chunk setitems lchunk coords lchunk selection chunk values file srv conda envs notebook lib site packages zarr core py line in chunk setitems cdatas self process for setitem key sel val fields fields file srv conda envs notebook lib site packages zarr core py line in cdatas self process for setitem key sel val fields fields file srv conda envs notebook lib site packages zarr core py line in process for setitem chunk self decode chunk cdata file srv conda envs notebook lib site packages zarr core py line in decode chunk chunk chunk reshape expected shape or self chunks order self order valueerror cannot reshape array of size into shape
1
17,445
23,267,451,505
IssuesEvent
2022-08-04 18:52:17
mdsreq-fga-unb/2022.1-Meio-a-Meio
https://api.github.com/repos/mdsreq-fga-unb/2022.1-Meio-a-Meio
closed
ABORDAGEM DE DESENVOLVIMENTO DE SOFTWARE
Processo de Desenvolvimento Planejamento e Organização
**Decrição** o planejamento do projeto e a abordagem informada **estão em desacordo**.
1.0
ABORDAGEM DE DESENVOLVIMENTO DE SOFTWARE - **Decrição** o planejamento do projeto e a abordagem informada **estão em desacordo**.
process
abordagem de desenvolvimento de software decrição o planejamento do projeto e a abordagem informada estão em desacordo
1
33,961
7,312,980,387
IssuesEvent
2018-02-28 22:54:11
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Log Scoping Needed with recoverable exceptions
C: MOOSE P: normal T: defect
## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> MOOSE currently fails in debug mode when using logging and a recoverable exception is throw. The behavior is undefined in optimized mode, but the program is likely to fail. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> Right now if you run MOOSE with performance logging turned on and your application throws an exception your performance log will be corrupted and may cause your program to crash. In debug mode, it'll certainly assert. The reason is that log "pops" are missed when exceptions are thrown. This needs to be fixed to keep the framework working robustly and properly. LibMesh already has a solution to this problem using a MACRO to create a sentinel object that properly pops the log when that object goes out of scope. One solution would be to implement something similar for MOOSE. ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> Framework robustness
1.0
Log Scoping Needed with recoverable exceptions - ## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> MOOSE currently fails in debug mode when using logging and a recoverable exception is throw. The behavior is undefined in optimized mode, but the program is likely to fail. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> Right now if you run MOOSE with performance logging turned on and your application throws an exception your performance log will be corrupted and may cause your program to crash. In debug mode, it'll certainly assert. The reason is that log "pops" are missed when exceptions are thrown. This needs to be fixed to keep the framework working robustly and properly. LibMesh already has a solution to this problem using a MACRO to create a sentinel object that properly pops the log when that object goes out of scope. One solution would be to implement something similar for MOOSE. ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> Framework robustness
non_process
log scoping needed with recoverable exceptions rationale moose currently fails in debug mode when using logging and a recoverable exception is throw the behavior is undefined in optimized mode but the program is likely to fail description right now if you run moose with performance logging turned on and your application throws an exception your performance log will be corrupted and may cause your program to crash in debug mode it ll certainly assert the reason is that log pops are missed when exceptions are thrown this needs to be fixed to keep the framework working robustly and properly libmesh already has a solution to this problem using a macro to create a sentinel object that properly pops the log when that object goes out of scope one solution would be to implement something similar for moose impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api framework robustness
0
8,201
11,396,207,932
IssuesEvent
2020-01-30 13:07:47
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
opened
.raw(..) query
kind/feature process/candidate
We want to offer a fallback for all queries that are not supported by Prisma Client itself. We will achieve that by offering a `.raw` method that you can use to send any query to the database and get back its result.
1.0
.raw(..) query - We want to offer a fallback for all queries that are not supported by Prisma Client itself. We will achieve that by offering a `.raw` method that you can use to send any query to the database and get back its result.
process
raw query we want to offer a fallback for all queries that are not supported by prisma client itself we will achieve that by offering a raw method that you can use to send any query to the database and get back its result
1
15,626
19,782,786,931
IssuesEvent
2022-01-18 00:14:59
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
closed
poem file creator tool
script migration processing
similar to #8, a tool that creates the poem file. very simple, it just needs: - [x] poem id - [x] poem title - [x] my name - [x] date - [x] poem lines
1.0
poem file creator tool - similar to #8, a tool that creates the poem file. very simple, it just needs: - [x] poem id - [x] poem title - [x] my name - [x] date - [x] poem lines
process
poem file creator tool similar to a tool that creates the poem file very simple it just needs poem id poem title my name date poem lines
1
46,235
2,951,657,491
IssuesEvent
2015-07-07 01:22:40
Co0sh/BetonQuest
https://api.github.com/repos/Co0sh/BetonQuest
opened
Translations
High Priority Other
We are getting closer and closer to 1.7 release. There were some additions to the _messages.yml_, so it needs to be updated with new translations. Also, the plugin now has translation system for all quests, so it would be nice to have the default one translated as well. Remember, in [_messages.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/messages.yml) there cannot be any special characters: they go to [_advanced-messages.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/advanced-messages.yml) file. The strings that need translation are on the bottom of each language. The only exception is "specify_objective" string, where "instruction" should be renamed to "name" (see English version). If you want to translate the default quest as well, the details can be found [here](http://betonquest.betoncraft.pl/BetonQuestDevDocumentation.pdf). The things that needs translating are: NPC/player options, "quester" name and "unknown" message in [_defaultConversation.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/defaultConversation.yml), names of quest canceler in [_main.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/main.yml) and journal entries in [_journal.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/journal.yml). And a notification for you guys ^^ * Chinese: @SmilingTrickster * German: @Bastikeks * French: @fastlockel * Spanish: @adolfotupo I would be really happy if you could help me with this :grinning:
1.0
Translations - We are getting closer and closer to 1.7 release. There were some additions to the _messages.yml_, so it needs to be updated with new translations. Also, the plugin now has translation system for all quests, so it would be nice to have the default one translated as well. Remember, in [_messages.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/messages.yml) there cannot be any special characters: they go to [_advanced-messages.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/advanced-messages.yml) file. The strings that need translation are on the bottom of each language. The only exception is "specify_objective" string, where "instruction" should be renamed to "name" (see English version). If you want to translate the default quest as well, the details can be found [here](http://betonquest.betoncraft.pl/BetonQuestDevDocumentation.pdf). The things that needs translating are: NPC/player options, "quester" name and "unknown" message in [_defaultConversation.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/defaultConversation.yml), names of quest canceler in [_main.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/main.yml) and journal entries in [_journal.yml_](https://github.com/Co0sh/BetonQuest/blob/master/src/main/resources/journal.yml). And a notification for you guys ^^ * Chinese: @SmilingTrickster * German: @Bastikeks * French: @fastlockel * Spanish: @adolfotupo I would be really happy if you could help me with this :grinning:
non_process
translations we are getting closer and closer to release there were some additions to the messages yml so it needs to be updated with new translations also the plugin now has translation system for all quests so it would be nice to have the default one translated as well remember in there cannot be any special characters they go to file the strings that need translation are on the bottom of each language the only exception is specify objective string where instruction should be renamed to name see english version if you want to translate the default quest as well the details can be found the things that needs translating are npc player options quester name and unknown message in names of quest canceler in and journal entries in and a notification for you guys chinese smilingtrickster german bastikeks french fastlockel spanish adolfotupo i would be really happy if you could help me with this grinning
0
111,014
4,448,623,911
IssuesEvent
2016-08-22 01:01:17
raxod502/dotfiles
https://api.github.com/repos/raxod502/dotfiles
closed
ensure-gitconfig-local.sh should copy existing info from ~/.gitconfig
enhancement git priority scripts
Currently, you always have to enter your name and email when running `setup.sh` for the first time. We can get this information from the `git config` command if there is already a `~/.gitconfig` file, so this should be copied automatically (if available).
1.0
ensure-gitconfig-local.sh should copy existing info from ~/.gitconfig - Currently, you always have to enter your name and email when running `setup.sh` for the first time. We can get this information from the `git config` command if there is already a `~/.gitconfig` file, so this should be copied automatically (if available).
non_process
ensure gitconfig local sh should copy existing info from gitconfig currently you always have to enter your name and email when running setup sh for the first time we can get this information from the git config command if there is already a gitconfig file so this should be copied automatically if available
0
791
3,274,035,580
IssuesEvent
2015-10-26 08:34:19
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Move same topic fragment processing before conref
bug DITA 1.3 P2 preprocess preprocess/conref ready
Move same topic fragment processing before conref processing to enable ```xml <p conref="#./foo"/> ``` As reported in #1649
2.0
Move same topic fragment processing before conref - Move same topic fragment processing before conref processing to enable ```xml <p conref="#./foo"/> ``` As reported in #1649
process
move same topic fragment processing before conref move same topic fragment processing before conref processing to enable xml as reported in
1
132,305
5,176,356,011
IssuesEvent
2017-01-19 00:21:52
yairodriguez/es6.today
https://api.github.com/repos/yairodriguez/es6.today
opened
Home
[priority] critical [status] accepted [type] feature
### Description Create the *Home* description. --- ### Issue Checklist - [ ] Write the needed tests. - [ ] Create the markup for the component. - [ ] Add the correct styles. - [ ] Use the lint to clean `CSS` code. - [ ] Document all things. --- ### Assignees - [ ] Final assign @yairodriguez
1.0
Home - ### Description Create the *Home* description. --- ### Issue Checklist - [ ] Write the needed tests. - [ ] Create the markup for the component. - [ ] Add the correct styles. - [ ] Use the lint to clean `CSS` code. - [ ] Document all things. --- ### Assignees - [ ] Final assign @yairodriguez
non_process
home description create the home description issue checklist write the needed tests create the markup for the component add the correct styles use the lint to clean css code document all things assignees final assign yairodriguez
0
12,875
15,264,988,082
IssuesEvent
2021-02-22 06:30:47
tdwg/chrono
https://api.github.com/repos/tdwg/chrono
closed
Change term - ChronometricAge
Process - under public review Term - change
## Change term * Submitter: Laura Brenskelle * Justification (why is this change necessary?): After discussion (that can be found here https://github.com/tdwg/chrono/issues/20), this definition requires clarification. This will make the intended uses of this extension more clear for data users and providers. We have also changed it to provide examples and a usage comment of when it is and is not appropriate to use the ChronometricAge extension. * Proponents (who needs this change): Public review. Proposed new attributes of the term: * Organized in Class: ChronometricAge * Term name (in lowerCamelCase): chrono:ChronometricAge * Definition of the term: "The age of a specimen and how this age is known, whether by a dating assay, a relative association with dated material, or legacy collections information." * Usage comments (recommendations regarding content, etc.): The ChronometricAge extension is to be used only in cases where the collection event is not contemporaneous with the time when the organism was alive in its context. Collection event information can be reported in dwc:eventDate. * Examples: "An age range associated with a specimen derived from an AMS dating assay applied to an oyster shell in the same stratum"; "An age range associated with a specimen derived from a ceramics analysis based on other materials found in the same stratum"; “A maximum age associated with a specimen derived from K-Ar dating applied to a proximal volcanic tuff found stratigraphically below the specimen”; “An age range of a specimen based on its biostratigraphic context”; “An age of a specimen based on what is reported in legacy collections data”. * Refines (identifier of the broader term this term refines, if applicable): N/A * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): N/A * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): N/A
1.0
Change term - ChronometricAge - ## Change term * Submitter: Laura Brenskelle * Justification (why is this change necessary?): After discussion (that can be found here https://github.com/tdwg/chrono/issues/20), this definition requires clarification. This will make the intended uses of this extension more clear for data users and providers. We have also changed it to provide examples and a usage comment of when it is and is not appropriate to use the ChronometricAge extension. * Proponents (who needs this change): Public review. Proposed new attributes of the term: * Organized in Class: ChronometricAge * Term name (in lowerCamelCase): chrono:ChronometricAge * Definition of the term: "The age of a specimen and how this age is known, whether by a dating assay, a relative association with dated material, or legacy collections information." * Usage comments (recommendations regarding content, etc.): The ChronometricAge extension is to be used only in cases where the collection event is not contemporaneous with the time when the organism was alive in its context. Collection event information can be reported in dwc:eventDate. * Examples: "An age range associated with a specimen derived from an AMS dating assay applied to an oyster shell in the same stratum"; "An age range associated with a specimen derived from a ceramics analysis based on other materials found in the same stratum"; “A maximum age associated with a specimen derived from K-Ar dating applied to a proximal volcanic tuff found stratigraphically below the specimen”; “An age range of a specimen based on its biostratigraphic context”; “An age of a specimen based on what is reported in legacy collections data”. * Refines (identifier of the broader term this term refines, if applicable): N/A * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): N/A * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): N/A
process
change term chronometricage change term submitter laura brenskelle justification why is this change necessary after discussion that can be found here this definition requires clarification this will make the intended uses of this extension more clear for data users and providers we have also changed it to provide examples and a usage comment of when it is and is not appropriate to use the chronometricage extension proponents who needs this change public review proposed new attributes of the term organized in class chronometricage term name in lowercamelcase chrono chronometricage definition of the term the age of a specimen and how this age is known whether by a dating assay a relative association with dated material or legacy collections information usage comments recommendations regarding content etc the chronometricage extension is to be used only in cases where the collection event is not contemporaneous with the time when the organism was alive in its context collection event information can be reported in dwc eventdate examples an age range associated with a specimen derived from an ams dating assay applied to an oyster shell in the same stratum an age range associated with a specimen derived from a ceramics analysis based on other materials found in the same stratum “a maximum age associated with a specimen derived from k ar dating applied to a proximal volcanic tuff found stratigraphically below the specimen” “an age range of a specimen based on its biostratigraphic context” “an age of a specimen based on what is reported in legacy collections data” refines identifier of the broader term this term refines if applicable n a replaces identifier of the existing term that would be deprecated and replaced by this term if applicable n a abcd xpath of the equivalent term in abcd or efg if applicable n a
1
20,747
27,451,656,142
IssuesEvent
2023-03-02 17:50:25
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Missing parent: [GO term] should be descendants of nuclear mRNA surveillance
RNA processes ready missing parentage
* **GO term ID and label for which you request a new superclass** <img width="892" alt="Screenshot 2022-10-26 at 12 05 16" src="https://user-images.githubusercontent.com/7359272/198010750-bd1ff69c-cddd-4e22-94b5-6f7ce5862428.png"> * **New superclass (parent) suggested** nuclear mRNA surveillance * **Reference(s), if appropriate** PMID:nnnnnnn Once this is done there will be terms under " nuclear mRNA surveillance" with representing the same process. To deal with later.
1.0
Missing parent: [GO term] should be descendants of nuclear mRNA surveillance - * **GO term ID and label for which you request a new superclass** <img width="892" alt="Screenshot 2022-10-26 at 12 05 16" src="https://user-images.githubusercontent.com/7359272/198010750-bd1ff69c-cddd-4e22-94b5-6f7ce5862428.png"> * **New superclass (parent) suggested** nuclear mRNA surveillance * **Reference(s), if appropriate** PMID:nnnnnnn Once this is done there will be terms under " nuclear mRNA surveillance" with representing the same process. To deal with later.
process
missing parent should be descendants of nuclear mrna surveillance go term id and label for which you request a new superclass img width alt screenshot at src new superclass parent suggested nuclear mrna surveillance reference s if appropriate pmid nnnnnnn once this is done there will be terms under nuclear mrna surveillance with representing the same process to deal with later
1
123,728
17,772,307,158
IssuesEvent
2021-08-30 14:57:12
kapseliboi/platform-status
https://api.github.com/repos/kapseliboi/platform-status
opened
CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz, cryptiles-3.1.2.tgz
security vulnerability
## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>cryptiles-2.0.5.tgz</b>, <b>cryptiles-3.1.2.tgz</b></p></summary> <p> <details><summary><b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> <p>Path to dependency file: platform-status/package.json</p> <p>Path to vulnerable library: platform-status/node_modules/browser-sync/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.18.13.tgz (Root Library) - localtunnel-1.8.3.tgz - request-2.81.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </details> <details><summary><b>cryptiles-3.1.2.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-3.1.2.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-3.1.2.tgz</a></p> <p>Path to dependency file: platform-status/package.json</p> <p>Path to vulnerable library: platform-status/node_modules/eu/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - eu-1.3.2.tgz (Root Library) - request-2.83.0.tgz - hawk-6.0.2.tgz - :x: **cryptiles-3.1.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/platform-status/commit/f1843ccb4f9fa8cac219c196c9bcceb734286e98">f1843ccb4f9fa8cac219c196c9bcceb734286e98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p> <p>Release Date: 2018-07-09</p> <p>Fix Resolution: v4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz, cryptiles-3.1.2.tgz - ## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>cryptiles-2.0.5.tgz</b>, <b>cryptiles-3.1.2.tgz</b></p></summary> <p> <details><summary><b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> <p>Path to dependency file: platform-status/package.json</p> <p>Path to vulnerable library: platform-status/node_modules/browser-sync/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.18.13.tgz (Root Library) - localtunnel-1.8.3.tgz - request-2.81.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </details> <details><summary><b>cryptiles-3.1.2.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-3.1.2.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-3.1.2.tgz</a></p> <p>Path to dependency file: platform-status/package.json</p> <p>Path to vulnerable library: platform-status/node_modules/eu/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - eu-1.3.2.tgz (Root Library) - request-2.83.0.tgz - hawk-6.0.2.tgz - :x: **cryptiles-3.1.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/platform-status/commit/f1843ccb4f9fa8cac219c196c9bcceb734286e98">f1843ccb4f9fa8cac219c196c9bcceb734286e98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p> <p>Release Date: 2018-07-09</p> <p>Fix Resolution: v4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in cryptiles tgz cryptiles tgz cve high severity vulnerability vulnerable libraries cryptiles tgz cryptiles tgz cryptiles tgz general purpose crypto utilities library home page a href path to dependency file platform status package json path to vulnerable library platform status node modules browser sync node modules cryptiles package json dependency hierarchy browser sync tgz root library localtunnel tgz request tgz hawk tgz x cryptiles tgz vulnerable library cryptiles tgz general purpose crypto utilities library home page a href path to dependency file platform status package json path to vulnerable library platform status node modules eu node modules cryptiles package json dependency hierarchy eu tgz root library request tgz hawk tgz x cryptiles tgz vulnerable library found in head commit a href found in base branch master vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
9,454
12,429,905,385
IssuesEvent
2020-05-25 09:20:48
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
closed
Cannot use prisma.create() with nullable field set to null
bug/2-confirmed kind/regression process/candidate
## Bug description <!-- A clear and concise description of what the bug is. --> When i give the value "null" to a nullable field on prisma.mymodel.create(...) i receive an error : Invalid `prisma.mymodel.create()` invocation. Missing a required value at... ## How to reproduce Schema: ``` model Sport { name String @id } model NameReconciliation { id String @default(uuid()) @id target String name String sport Sport? @relation(fields: [sportId], references: [name]) sportId String? } ``` So the relation Sport of NameReconciliation is nullable as in the generated prisma client : ``` export type NameReconciliationCreateInput = { id?: string | null target: string name: string sport?: SportCreateOneWithoutNameReconciliationInput | null } ``` This is my mutation : ``` createNameReconciliation(root: any, { name, target, sportId }: any, ctx: Context): Promise<NameReconciliation | null> { return ctx.prisma.nameReconciliation.create({ data: { name, sport: null, target } }) } ``` So when i run this mutation in the playground : ``` mutation { createNameReconciliation(name: "soccer2", target: "football") {id, name} } ``` I receive : ``` "message": "\nInvalid prisma.nameReconciliation.create() invocation in\n/home/****/src/resolvers/reconciliate.ts:14:50 Missing a required value at Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport" "extensions": { "code": "INTERNAL_SERVER_ERROR", "exception": { "code": "P2012", "meta": { "path": "Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport" }, "stacktrace": [ "Error: ", "Invalid `prisma.nameReconciliation.create()` invocation in", "/home/****/src/resolvers/reconciliate.ts:14:50", "", "Missing a required value at `Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport`", " at PrismaClientFetcher.request (/home/****/node_modules/@prisma/client/src/runtime/getPrismaClient.ts:647:15)", " at processTicksAndRejections (internal/process/task_queues.js:97:5)" ] } } ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> I expect to just create a NameReconciliation object with the relation Sport set to null. ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... Do not include your database credentials when sharing your Prisma schema! --> It was working before i made the update to 2.0.0-beta.5 (i was in 2.0.0-beta.1) ## Environment & setup <!-- In which environment does the problem occur --> @prisma/cli : 2.0.0-beta.5 Current platform : debian-openssl-1.1.x OS: ArchLinux Database: PostgreSQL Node.js: v12.16.2
1.0
Cannot use prisma.create() with nullable field set to null - ## Bug description <!-- A clear and concise description of what the bug is. --> When i give the value "null" to a nullable field on prisma.mymodel.create(...) i receive an error : Invalid `prisma.mymodel.create()` invocation. Missing a required value at... ## How to reproduce Schema: ``` model Sport { name String @id } model NameReconciliation { id String @default(uuid()) @id target String name String sport Sport? @relation(fields: [sportId], references: [name]) sportId String? } ``` So the relation Sport of NameReconciliation is nullable as in the generated prisma client : ``` export type NameReconciliationCreateInput = { id?: string | null target: string name: string sport?: SportCreateOneWithoutNameReconciliationInput | null } ``` This is my mutation : ``` createNameReconciliation(root: any, { name, target, sportId }: any, ctx: Context): Promise<NameReconciliation | null> { return ctx.prisma.nameReconciliation.create({ data: { name, sport: null, target } }) } ``` So when i run this mutation in the playground : ``` mutation { createNameReconciliation(name: "soccer2", target: "football") {id, name} } ``` I receive : ``` "message": "\nInvalid prisma.nameReconciliation.create() invocation in\n/home/****/src/resolvers/reconciliate.ts:14:50 Missing a required value at Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport" "extensions": { "code": "INTERNAL_SERVER_ERROR", "exception": { "code": "P2012", "meta": { "path": "Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport" }, "stacktrace": [ "Error: ", "Invalid `prisma.nameReconciliation.create()` invocation in", "/home/****/src/resolvers/reconciliate.ts:14:50", "", "Missing a required value at `Mutation.createOneNameReconciliation.data.NameReconciliationCreateInput.sport`", " at PrismaClientFetcher.request (/home/****/node_modules/@prisma/client/src/runtime/getPrismaClient.ts:647:15)", " at processTicksAndRejections (internal/process/task_queues.js:97:5)" ] } } ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> I expect to just create a NameReconciliation object with the relation Sport set to null. ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... Do not include your database credentials when sharing your Prisma schema! --> It was working before i made the update to 2.0.0-beta.5 (i was in 2.0.0-beta.1) ## Environment & setup <!-- In which environment does the problem occur --> @prisma/cli : 2.0.0-beta.5 Current platform : debian-openssl-1.1.x OS: ArchLinux Database: PostgreSQL Node.js: v12.16.2
process
cannot use prisma create with nullable field set to null bug description when i give the value null to a nullable field on prisma mymodel create i receive an error invalid prisma mymodel create invocation missing a required value at how to reproduce schema model sport name string id model namereconciliation id string default uuid id target string name string sport sport relation fields references sportid string so the relation sport of namereconciliation is nullable as in the generated prisma client export type namereconciliationcreateinput id string null target string name string sport sportcreateonewithoutnamereconciliationinput null this is my mutation createnamereconciliation root any name target sportid any ctx context promise return ctx prisma namereconciliation create data name sport null target so when i run this mutation in the playground mutation createnamereconciliation name target football id name i receive message ninvalid prisma namereconciliation create invocation in n home src resolvers reconciliate ts missing a required value at mutation createonenamereconciliation data namereconciliationcreateinput sport extensions code internal server error exception code meta path mutation createonenamereconciliation data namereconciliationcreateinput sport stacktrace error invalid prisma namereconciliation create invocation in home src resolvers reconciliate ts missing a required value at mutation createonenamereconciliation data namereconciliationcreateinput sport at prismaclientfetcher request home node modules prisma client src runtime getprismaclient ts at processticksandrejections internal process task queues js expected behavior i expect to just create a namereconciliation object with the relation sport set to null prisma information your prisma schema prisma client queries do not include your database credentials when sharing your prisma schema it was working before i made the update to beta i was in beta environment setup prisma cli beta current platform debian openssl x os archlinux database postgresql node js
1
182,091
14,102,244,323
IssuesEvent
2020-11-06 08:26:15
input-output-hk/cardano-wallet
https://api.github.com/repos/input-output-hk/cardano-wallet
closed
Flaky TRANS_CREATE_03 - 0 balance after transaction
BUG Test failure
# Context <!-- WHEN CREATED Any information that is useful to understand the bug and the subsystem it evolves in. References to documentation and or other tickets are welcome. --> # Test Case https://github.com/input-output-hk/cardano-wallet/blob/0bf98c9e65af84702139dacc5ed0165777ff59e7/lib/core-integration/src/Test/Integration/Scenario/API/Shelley/Transactions.hs#L429-L439 # Failure / Counter-example ``` src/Test/Integration/Scenario/API/Shelley/Transactions.hs:435:26: 1) API Specifications, SHELLEY_TRANSACTIONS, TRANS_CREATE_03 - 0 balance after transaction While verifying (Status {statusCode = 200, statusMessage = "OK"},Right (ApiWallet {id = ApiT {getApiT = WalletId {getWalletId = 278f9ebf2dcf1cb23e1af27281daed107333fd16}}, addressPoolGap = ApiT {getApiT = AddressPoolGap {getAddressPoolGap = 20}}, balance = ApiT {getApiT = WalletBalance {available = Quantity {getQuantity = 2000000}, total = Quantity {getQuantity = 2000000}, reward = Quantity {getQuantity = 0}}}, delegation = ApiWalletDelegation {active = ApiWalletDelegationNext {status = NotDelegating, target = Nothing, changesAt = Nothing}, next = []}, name = ApiT {getApiT = WalletName {getWalletName = "Empty Wallet"}}, passphrase = Just (ApiWalletPassphraseInfo {lastUpdatedAt = 2020-11-04 18:51:44.994771905 UTC}), state = ApiT {getApiT = Ready}, tip = ApiBlockReference {absoluteSlotNumber = ApiT {getApiT = SlotNo 2113}, slotId = ApiSlotId {epochNumber = ApiT {getApiT = EpochNo {unEpochNo = 11}}, slotNumber = ApiT {getApiT = SlotInEpoch {unSlotInEpoch = 33}}}, time = 2020-11-04 18:53:51.6 UTC, block = ApiBlockInfo {height = Quantity {getQuantity = 946}}}})) Waited longer than 90s to resolve action: "Wallet balance is as expected". expected: Quantity {getQuantity = 1000000} but got: Quantity {getQuantity = 2000000} To rerun use: --match "/API Specifications/SHELLEY_TRANSACTIONS/TRANS_CREATE_03 - 0 balance after transaction/" ``` From custom build: https://buildkite.com/input-output-hk/cardano-wallet/builds/11719#0c39cbfa-8eac-4930-a3f1-016f08db8500, but I think I've seen it elsewhere too. <!-- WHEN CREATED For failing unit or integration tests, put here the output of the test runner. For properties, also provide the counter example given by QuickCheck. --> --- # Resolution <!-- WHEN IN PROGRESS What is happening? How is this going to be fixed? Detail the approach and give, in the form of a TODO list steps toward the resolution of the bug. Attach a PR to each item in the list. This may be refined as the investigation progresses. --> --- # QA <!-- WHEN IN PROGRESS How do we make sure the bug has been fixed? Give here manual steps or tests to verify the fix. How/why could this bug slip through testing? -->
1.0
Flaky TRANS_CREATE_03 - 0 balance after transaction - # Context <!-- WHEN CREATED Any information that is useful to understand the bug and the subsystem it evolves in. References to documentation and or other tickets are welcome. --> # Test Case https://github.com/input-output-hk/cardano-wallet/blob/0bf98c9e65af84702139dacc5ed0165777ff59e7/lib/core-integration/src/Test/Integration/Scenario/API/Shelley/Transactions.hs#L429-L439 # Failure / Counter-example ``` src/Test/Integration/Scenario/API/Shelley/Transactions.hs:435:26: 1) API Specifications, SHELLEY_TRANSACTIONS, TRANS_CREATE_03 - 0 balance after transaction While verifying (Status {statusCode = 200, statusMessage = "OK"},Right (ApiWallet {id = ApiT {getApiT = WalletId {getWalletId = 278f9ebf2dcf1cb23e1af27281daed107333fd16}}, addressPoolGap = ApiT {getApiT = AddressPoolGap {getAddressPoolGap = 20}}, balance = ApiT {getApiT = WalletBalance {available = Quantity {getQuantity = 2000000}, total = Quantity {getQuantity = 2000000}, reward = Quantity {getQuantity = 0}}}, delegation = ApiWalletDelegation {active = ApiWalletDelegationNext {status = NotDelegating, target = Nothing, changesAt = Nothing}, next = []}, name = ApiT {getApiT = WalletName {getWalletName = "Empty Wallet"}}, passphrase = Just (ApiWalletPassphraseInfo {lastUpdatedAt = 2020-11-04 18:51:44.994771905 UTC}), state = ApiT {getApiT = Ready}, tip = ApiBlockReference {absoluteSlotNumber = ApiT {getApiT = SlotNo 2113}, slotId = ApiSlotId {epochNumber = ApiT {getApiT = EpochNo {unEpochNo = 11}}, slotNumber = ApiT {getApiT = SlotInEpoch {unSlotInEpoch = 33}}}, time = 2020-11-04 18:53:51.6 UTC, block = ApiBlockInfo {height = Quantity {getQuantity = 946}}}})) Waited longer than 90s to resolve action: "Wallet balance is as expected". expected: Quantity {getQuantity = 1000000} but got: Quantity {getQuantity = 2000000} To rerun use: --match "/API Specifications/SHELLEY_TRANSACTIONS/TRANS_CREATE_03 - 0 balance after transaction/" ``` From custom build: https://buildkite.com/input-output-hk/cardano-wallet/builds/11719#0c39cbfa-8eac-4930-a3f1-016f08db8500, but I think I've seen it elsewhere too. <!-- WHEN CREATED For failing unit or integration tests, put here the output of the test runner. For properties, also provide the counter example given by QuickCheck. --> --- # Resolution <!-- WHEN IN PROGRESS What is happening? How is this going to be fixed? Detail the approach and give, in the form of a TODO list steps toward the resolution of the bug. Attach a PR to each item in the list. This may be refined as the investigation progresses. --> --- # QA <!-- WHEN IN PROGRESS How do we make sure the bug has been fixed? Give here manual steps or tests to verify the fix. How/why could this bug slip through testing? -->
non_process
flaky trans create balance after transaction context when created any information that is useful to understand the bug and the subsystem it evolves in references to documentation and or other tickets are welcome test case failure counter example src test integration scenario api shelley transactions hs api specifications shelley transactions trans create balance after transaction while verifying status statuscode statusmessage ok right apiwallet id apit getapit walletid getwalletid addresspoolgap apit getapit addresspoolgap getaddresspoolgap balance apit getapit walletbalance available quantity getquantity total quantity getquantity reward quantity getquantity delegation apiwalletdelegation active apiwalletdelegationnext status notdelegating target nothing changesat nothing next name apit getapit walletname getwalletname empty wallet passphrase just apiwalletpassphraseinfo lastupdatedat utc state apit getapit ready tip apiblockreference absoluteslotnumber apit getapit slotno slotid apislotid epochnumber apit getapit epochno unepochno slotnumber apit getapit slotinepoch unslotinepoch time utc block apiblockinfo height quantity getquantity waited longer than to resolve action wallet balance is as expected expected quantity getquantity but got quantity getquantity to rerun use match api specifications shelley transactions trans create balance after transaction from custom build but i think i ve seen it elsewhere too when created for failing unit or integration tests put here the output of the test runner for properties also provide the counter example given by quickcheck resolution when in progress what is happening how is this going to be fixed detail the approach and give in the form of a todo list steps toward the resolution of the bug attach a pr to each item in the list this may be refined as the investigation progresses qa when in progress how do we make sure the bug has been fixed give here manual steps or tests to verify the fix how why could this bug slip through testing
0
789,250
27,784,608,813
IssuesEvent
2023-03-17 01:22:03
NireaE/github-issues-template
https://api.github.com/repos/NireaE/github-issues-template
opened
index.html does not pass CSS validation
bug Severity 3 Priority 3
**Describe the bug** One CSS error comes up when running index.html page against CSS validator **To Reproduce** Steps to reproduce the behavior: 1. Go to 'https://jigsaw.w3.org/css-validator/' 2. Paste in the index page URL 3. Scroll down to '....' 4. View the error **Expected behavior** Page should be error free and valid, however an error shows up **Additional context** Click the link to view the CSS error https://jigsaw.w3.org/css-validator/validator?uri=https%3A%2F%2Fnireae.github.io%2Fgithub-issues-template%2Fresources%2Ffaq.html&profile=css3svg&usermedium=all&warning=1&vextwarning=&lang=en
1.0
index.html does not pass CSS validation - **Describe the bug** One CSS error comes up when running index.html page against CSS validator **To Reproduce** Steps to reproduce the behavior: 1. Go to 'https://jigsaw.w3.org/css-validator/' 2. Paste in the index page URL 3. Scroll down to '....' 4. View the error **Expected behavior** Page should be error free and valid, however an error shows up **Additional context** Click the link to view the CSS error https://jigsaw.w3.org/css-validator/validator?uri=https%3A%2F%2Fnireae.github.io%2Fgithub-issues-template%2Fresources%2Ffaq.html&profile=css3svg&usermedium=all&warning=1&vextwarning=&lang=en
non_process
index html does not pass css validation describe the bug one css error comes up when running index html page against css validator to reproduce steps to reproduce the behavior go to paste in the index page url scroll down to view the error expected behavior page should be error free and valid however an error shows up additional context click the link to view the css error
0
183,682
21,778,191,146
IssuesEvent
2022-05-13 15:45:43
capitalone/DataProfiler
https://api.github.com/repos/capitalone/DataProfiler
closed
CVE-2021-41496 (High) detected in numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl - autoclosed
security vulnerability
## CVE-2021-41496 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p></summary> <p>NumPy is the fundamental package for array computing with Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/50/46/292cff79f5b30151b027400efdb3f740ea03271b600751b6696cf550c10d/numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/50/46/292cff79f5b30151b027400efdb3f740ea03271b600751b6696cf550c10d/numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/requirements.txt</p> <p> Dependency Hierarchy: - :x: **numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** Buffer overflow in the array_from_pyobj function of fortranobject.c in NumPy < 1.19, which allows attackers to conduct a Denial of Service attacks by carefully constructing an array with negative values. NOTE: The vendor does not agree this is a vulnerability; the negative dimensions can only be created by an already privileged user (or internally). <p>Publish Date: 2021-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41496>CVE-2021-41496</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-41496">https://nvd.nist.gov/vuln/detail/CVE-2021-41496</a></p> <p>Release Date: 2021-12-17</p> <p>Fix Resolution: autovizwidget - 0.12.7;numpy - 1.22.0rc1;numcodecs - 0.6.2;numpy-base - 1.11.3;numpy - 1.17.4</p> </p> </details> <p></p>
True
CVE-2021-41496 (High) detected in numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl - autoclosed - ## CVE-2021-41496 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p></summary> <p>NumPy is the fundamental package for array computing with Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/50/46/292cff79f5b30151b027400efdb3f740ea03271b600751b6696cf550c10d/numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/50/46/292cff79f5b30151b027400efdb3f740ea03271b600751b6696cf550c10d/numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/requirements.txt</p> <p> Dependency Hierarchy: - :x: **numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** Buffer overflow in the array_from_pyobj function of fortranobject.c in NumPy < 1.19, which allows attackers to conduct a Denial of Service attacks by carefully constructing an array with negative values. NOTE: The vendor does not agree this is a vulnerability; the negative dimensions can only be created by an already privileged user (or internally). <p>Publish Date: 2021-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41496>CVE-2021-41496</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-41496">https://nvd.nist.gov/vuln/detail/CVE-2021-41496</a></p> <p>Release Date: 2021-12-17</p> <p>Fix Resolution: autovizwidget - 0.12.7;numpy - 1.22.0rc1;numcodecs - 0.6.2;numpy-base - 1.11.3;numpy - 1.17.4</p> </p> </details> <p></p>
non_process
cve high detected in numpy manylinux whl autoclosed cve high severity vulnerability vulnerable library numpy manylinux whl numpy is the fundamental package for array computing with python library home page a href path to dependency file requirements txt path to vulnerable library requirements txt requirements txt dependency hierarchy x numpy manylinux whl vulnerable library vulnerability details disputed buffer overflow in the array from pyobj function of fortranobject c in numpy which allows attackers to conduct a denial of service attacks by carefully constructing an array with negative values note the vendor does not agree this is a vulnerability the negative dimensions can only be created by an already privileged user or internally publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution autovizwidget numpy numcodecs numpy base numpy
0
175,665
13,578,474,583
IssuesEvent
2020-09-20 07:57:35
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
closed
Bind HTML to input boxes so that we can test the functionality of clearing the text boxes
a-Testing p.Low
Related to #3235. ![screen shot 2015-04-05 at 1 06 17 am](https://cloud.githubusercontent.com/assets/6291947/6993846/f82545ce-db34-11e4-8a9d-7e8fbf687acf.png) ![screen shot 2015-04-05 at 1 06 01 am](https://cloud.githubusercontent.com/assets/6291947/6993847/f827e612-db34-11e4-9258-6836fa0d614e.png) Currently, when changing the text boxes for Rubric options, MCQ and MSQ options, the HTML remains as the default initialized values and do not change - thus rendering us unable to test for the functionality of clearing text boxes.
1.0
Bind HTML to input boxes so that we can test the functionality of clearing the text boxes - Related to #3235. ![screen shot 2015-04-05 at 1 06 17 am](https://cloud.githubusercontent.com/assets/6291947/6993846/f82545ce-db34-11e4-8a9d-7e8fbf687acf.png) ![screen shot 2015-04-05 at 1 06 01 am](https://cloud.githubusercontent.com/assets/6291947/6993847/f827e612-db34-11e4-9258-6836fa0d614e.png) Currently, when changing the text boxes for Rubric options, MCQ and MSQ options, the HTML remains as the default initialized values and do not change - thus rendering us unable to test for the functionality of clearing text boxes.
non_process
bind html to input boxes so that we can test the functionality of clearing the text boxes related to currently when changing the text boxes for rubric options mcq and msq options the html remains as the default initialized values and do not change thus rendering us unable to test for the functionality of clearing text boxes
0
22,369
31,141,081,475
IssuesEvent
2023-08-16 00:04:30
Warzone2100/map-submission
https://api.github.com/repos/Warzone2100/map-submission
opened
[MAP]: Fortresses
map unprocessed
### Upload Map [10c-Fortresses.zip](https://github.com/Warzone2100/map-submission/files/12353843/10c-Fortresses.zip) ### Authorship Mine: I am the author of this map ### Map Description (optional) ```text 10 Player Skirmish map ``` ### Notes for Reviewers (optional) _No response_
1.0
[MAP]: Fortresses - ### Upload Map [10c-Fortresses.zip](https://github.com/Warzone2100/map-submission/files/12353843/10c-Fortresses.zip) ### Authorship Mine: I am the author of this map ### Map Description (optional) ```text 10 Player Skirmish map ``` ### Notes for Reviewers (optional) _No response_
process
fortresses upload map authorship mine i am the author of this map map description optional text player skirmish map notes for reviewers optional no response
1
13,554
16,099,085,334
IssuesEvent
2021-04-27 06:54:22
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Unable to verify S/MIME mail signature signed with certificate chain
bug mail processing verified
### Infos: * Used Zammad version: 4.0 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any ### Expected behavior: An S/MIME email that is signed with a certificate from a certificate chain should be successfully verifiable if Zammad has each certificate of the chain available in its certificate store. ### Actual behavior: An S/MIME email that is signed with a certificate from a certificate chain can't be verified by Zammad even if Zammad has each certificate of the chain available in its certificate store. ### Steps to reproduce the behavior: * Upload all the certificates from a certificate into Zammads certificate store * Sign an email with that certificate chain * Send that mail to Zammad * See that Zammad can't verify the signature Yes I'm sure this is a bug and no feature request or a general question.
1.0
Unable to verify S/MIME mail signature signed with certificate chain - ### Infos: * Used Zammad version: 4.0 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any ### Expected behavior: An S/MIME email that is signed with a certificate from a certificate chain should be successfully verifiable if Zammad has each certificate of the chain available in its certificate store. ### Actual behavior: An S/MIME email that is signed with a certificate from a certificate chain can't be verified by Zammad even if Zammad has each certificate of the chain available in its certificate store. ### Steps to reproduce the behavior: * Upload all the certificates from a certificate into Zammads certificate store * Sign an email with that certificate chain * Send that mail to Zammad * See that Zammad can't verify the signature Yes I'm sure this is a bug and no feature request or a general question.
process
unable to verify s mime mail signature signed with certificate chain infos used zammad version installation method source package any operating system any database version any elasticsearch version any browser version any expected behavior an s mime email that is signed with a certificate from a certificate chain should be successfully verifiable if zammad has each certificate of the chain available in its certificate store actual behavior an s mime email that is signed with a certificate from a certificate chain can t be verified by zammad even if zammad has each certificate of the chain available in its certificate store steps to reproduce the behavior upload all the certificates from a certificate into zammads certificate store sign an email with that certificate chain send that mail to zammad see that zammad can t verify the signature yes i m sure this is a bug and no feature request or a general question
1
63,819
3,201,092,183
IssuesEvent
2015-10-02 02:46:31
cs2103aug2015-f10-3j/main
https://api.github.com/repos/cs2103aug2015-f10-3j/main
closed
Parser:API support for adding deadline task
priority.high type.task
By specifying description or By specifying description, time and date.
1.0
Parser:API support for adding deadline task - By specifying description or By specifying description, time and date.
non_process
parser api support for adding deadline task by specifying description or by specifying description time and date
0
6,963
10,116,266,756
IssuesEvent
2019-07-31 01:11:22
tokio-rs/tokio
https://api.github.com/repos/tokio-rs/tokio
opened
process: misc improvements
meta tokio-process
Tracking misc smaller improvements to `tokio-process`. This issue is to be closed once there are no remaining breaking changes to make **or** remaining breaking changes are tracked in separate issues. ### Update to `std::future` - [ ] follow the future / stream helper strategy from #1225. ### Polish - [ ] revisit `CommandExt` trait and default impls (is this the optimal way of exposing functionality? can we update the trait without technically "breaking" consumers?) - [ ] Update CHANGELOG with all changes since the previous version ### Re-export - [ ] consider re-exporting `tokio_process::CommandExt` as `tokio::process::CommandExt`
1.0
process: misc improvements - Tracking misc smaller improvements to `tokio-process`. This issue is to be closed once there are no remaining breaking changes to make **or** remaining breaking changes are tracked in separate issues. ### Update to `std::future` - [ ] follow the future / stream helper strategy from #1225. ### Polish - [ ] revisit `CommandExt` trait and default impls (is this the optimal way of exposing functionality? can we update the trait without technically "breaking" consumers?) - [ ] Update CHANGELOG with all changes since the previous version ### Re-export - [ ] consider re-exporting `tokio_process::CommandExt` as `tokio::process::CommandExt`
process
process misc improvements tracking misc smaller improvements to tokio process this issue is to be closed once there are no remaining breaking changes to make or remaining breaking changes are tracked in separate issues update to std future follow the future stream helper strategy from polish revisit commandext trait and default impls is this the optimal way of exposing functionality can we update the trait without technically breaking consumers update changelog with all changes since the previous version re export consider re exporting tokio process commandext as tokio process commandext
1
3,121
6,151,408,080
IssuesEvent
2017-06-28 02:25:43
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Test: System.Diagnostics.Tests.ProcessModuleTests/Modules_Get_ContainsHostFileName failed with "Xunit.Sdk.ContainsException"
area-System.Diagnostics.Process os-windows-uwp test-run-uwp-coreclr
Opened on behalf of @Jiayili1 The test `System.Diagnostics.Tests.ProcessModuleTests/Modules_Get_ContainsHostFileName` has failed. Assert.Contains() Failure\r Not found: (filter expression)\r In value: <CastIterator>d__34<ProcessModule> [System.Diagnostics.ProcessModule (XUnit.Runner.Uap.exe), System.Diagnostics.ProcessModule (ntdll.dll), System.Diagnostics.ProcessModule (KERNEL32.DLL), System.Diagnostics.ProcessModule (KERNELBASE.dll), System.Diagnostics.ProcessModule (ucrtbase.dll), ...] Stack Trace: at System.Diagnostics.Tests.ProcessModuleTests.Modules_Get_ContainsHostFileName() Build : Master - 20170510.01 (UWP F5 Tests) Failing configurations: - Windows.10.Amd64-x64 - Debug - Release Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fuwp~2F/build/20170510.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessModuleTests~2FModules_Get_ContainsHostFileName
1.0
Test: System.Diagnostics.Tests.ProcessModuleTests/Modules_Get_ContainsHostFileName failed with "Xunit.Sdk.ContainsException" - Opened on behalf of @Jiayili1 The test `System.Diagnostics.Tests.ProcessModuleTests/Modules_Get_ContainsHostFileName` has failed. Assert.Contains() Failure\r Not found: (filter expression)\r In value: <CastIterator>d__34<ProcessModule> [System.Diagnostics.ProcessModule (XUnit.Runner.Uap.exe), System.Diagnostics.ProcessModule (ntdll.dll), System.Diagnostics.ProcessModule (KERNEL32.DLL), System.Diagnostics.ProcessModule (KERNELBASE.dll), System.Diagnostics.ProcessModule (ucrtbase.dll), ...] Stack Trace: at System.Diagnostics.Tests.ProcessModuleTests.Modules_Get_ContainsHostFileName() Build : Master - 20170510.01 (UWP F5 Tests) Failing configurations: - Windows.10.Amd64-x64 - Debug - Release Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fuwp~2F/build/20170510.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessModuleTests~2FModules_Get_ContainsHostFileName
process
test system diagnostics tests processmoduletests modules get containshostfilename failed with xunit sdk containsexception opened on behalf of the test system diagnostics tests processmoduletests modules get containshostfilename has failed assert contains failure r not found filter expression r in value d stack trace at system diagnostics tests processmoduletests modules get containshostfilename build master uwp tests failing configurations windows debug release detail
1
2,151
7,345,448,378
IssuesEvent
2018-03-07 17:26:31
department-of-veterans-affairs/compliance
https://api.github.com/repos/department-of-veterans-affairs/compliance
opened
Human Interface
Design Engineering Architecture (DEA) In Scope
**Description** Interfaces with Common Look and Feel, that are compliant with Section 508 of the Rehabilitation Act of 1998 **Success Criteria**
1.0
Human Interface - **Description** Interfaces with Common Look and Feel, that are compliant with Section 508 of the Rehabilitation Act of 1998 **Success Criteria**
non_process
human interface description interfaces with common look and feel that are compliant with section of the rehabilitation act of success criteria
0
726,355
24,996,011,899
IssuesEvent
2022-11-03 00:18:42
EmulatorNexus/VeniceUnleashed
https://api.github.com/repos/EmulatorNexus/VeniceUnleashed
closed
[REQ] rcon var to disable preround
enhancement low priority
Since VU doesn't feature a progression system, there should be an option to disable preround entirely, much like game.disablepreround
1.0
[REQ] rcon var to disable preround - Since VU doesn't feature a progression system, there should be an option to disable preround entirely, much like game.disablepreround
non_process
rcon var to disable preround since vu doesn t feature a progression system there should be an option to disable preround entirely much like game disablepreround
0
6,064
8,494,137,726
IssuesEvent
2018-10-28 18:44:46
Thutmose/Thut
https://api.github.com/repos/Thutmose/Thut
closed
World breaking Crash if putting ProjectRed Lamps into an elevator i.e. as top or bottom
Compatiblity Issue bug
#### What happens: If you have built your elevator and want to create it with the device linker and have Lamp - blocks from ProjectRed in them, the server crashes and can't be started any more until you delete the elevator via MC-Edit or something. I'll also open an issue on PR Repo. #### What you expected to happen: A perfectly fine and fluid gaming ex...perie.. Well it's Minecraft. #### Steps to reproduce: 1. Same 2. as 3. What happens ____ Crash Report, including Mod List ``` ---- Minecraft Crash Report ---- WARNING: coremods are present: FarseekCoreMod (Farseek-1.12-2.3.1.jar) Inventory Tweaks Coremod (InventoryTweaks-1.63.jar) CTMCorePlugin (CTM-MC1.12.2-0.3.2.18.jar) AppleCore (AppleCore-mc1.12.2-3.1.4.jar) BetterFoliageLoader (BetterFoliage-MC1.12-2.1.10.jar) Contact their authors BEFORE contacting forge // Everything's going to plan. No, really, that was supposed to happen. Time: 9/20/18 7:20 PM Description: Ticking entity java.lang.NoClassDefFoundError: codechicken/lib/model/bakery/ModelBakery at mrtjp.projectred.illumination.BlockLamp.getExtendedState(blocks.scala:75) at thut.api.entity.blockentity.BlockEntityUpdater.applyEntityCollision(BlockEntityUpdater.java:157) at thut.api.entity.blockentity.BlockEntityBase.func_70108_f(BlockEntityBase.java:153) at thut.api.entity.blockentity.BlockEntityBase.checkCollision(BlockEntityBase.java:199) at thut.api.entity.blockentity.BlockEntityBase.func_70071_h_(BlockEntityBase.java:314) at net.minecraft.world.World.func_72866_a(World.java:1990) at net.minecraft.world.WorldServer.func_72866_a(WorldServer.java:832) at net.minecraft.world.World.func_72870_g(World.java:1952) at net.minecraft.world.World.func_72939_s(World.java:1756) at net.minecraft.world.WorldServer.func_72939_s(WorldServer.java:613) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:767) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.ClassNotFoundException: codechicken.lib.model.bakery.ModelBakery at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:191) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 15 more Caused by: net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerException: Exception in class transformer net.minecraftforge.fml.common.asm.transformers.SideTransformer@d02f8d from coremod FMLCorePlugin at net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerWrapper.transform(ASMTransformerWrapper.java:260) at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:279) at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:176) ... 17 more Caused by: java.lang.RuntimeException: Attempted to load class codechicken/lib/model/bakery/ModelBakery for invalid side SERVER at net.minecraftforge.fml.common.asm.transformers.SideTransformer.transform(SideTransformer.java:62) at net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerWrapper.transform(ASMTransformerWrapper.java:256) ... 19 more A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Thread: Server thread Stacktrace: at mrtjp.projectred.illumination.BlockLamp.getExtendedState(blocks.scala:75) at thut.api.entity.blockentity.BlockEntityUpdater.applyEntityCollision(BlockEntityUpdater.java:157) at thut.api.entity.blockentity.BlockEntityBase.func_70108_f(BlockEntityBase.java:153) at thut.api.entity.blockentity.BlockEntityBase.checkCollision(BlockEntityBase.java:199) at thut.api.entity.blockentity.BlockEntityBase.func_70071_h_(BlockEntityBase.java:314) at net.minecraft.world.World.func_72866_a(World.java:1990) at net.minecraft.world.WorldServer.func_72866_a(WorldServer.java:832) at net.minecraft.world.World.func_72870_g(World.java:1952) -- Entity being ticked -- Details: Entity Type: thuttech:lift (thut.tech.common.entity.EntityLift) Entity ID: 2022 Entity Name: Elevator Entity's Exact location: 258.50, 4.00, 2053.50 Entity's Block location: World: (258,4,2053), Chunk: (at 2,0,5 in 16,128; contains blocks 256,0,2048 to 271,255,2063), Region: (0,4; contains chunks 0,128 to 31,159, blocks 0,0,2048 to 511,255,2559) Entity's Momentum: 0.00, 0.00, 0.00 Entity's Passengers: [] Entity's Vehicle: ~~ERROR~~ NullPointerException: null Stacktrace: at net.minecraft.world.World.func_72939_s(World.java:1756) at net.minecraft.world.WorldServer.func_72939_s(WorldServer.java:613) -- Affected level -- Details: Level name: world All players: 1 total; [EntityPlayerMP['PsyVirus'/2214, l='world', x=259.81, y=9.00, z=2053.50]] Chunk stats: ServerChunkCache: 817 Drop: 0 Level seed: 4209520796579142338 Level generator: ID 06 - BIOMESOP, ver 0. Features enabled: true Level generator options: Level spawn location: World: (87,73,2035), Chunk: (at 7,4,3 in 5,127; contains blocks 80,0,2032 to 95,255,2047), Region: (0,3; contains chunks 0,96 to 31,127, blocks 0,0,1536 to 511,255,2047) Level time: 13471423 game time, 14130359 day time Level dimension: 0 Level storage version: 0x04ABD - Anvil Level weather: Rain time: 6353 (now: true), thunder time: 53727 (now: false) Level game mode: Game mode: survival (ID 0). Hardcore: false. Cheats: false Stacktrace: at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:767) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526) at java.lang.Thread.run(Thread.java:748) -- System Details -- Details: Minecraft Version: 1.12.2 Operating System: Linux (amd64) version 4.4.0-135-generic Java Version: 1.8.0_181, Oracle Corporation Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 580625768 bytes (553 MB) / 2359259136 bytes (2249 MB) up to 10468982784 bytes (9984 MB) JVM Flags: 8 total; -Xms512M -Xmx10G -XX:NewSize=512M -XX:MaxNewSize=1G -XX:SurvivorRatio=2 -XX:+DisableExplicitGC -XX:+UseConcMarkSweepGC -XX:+AggressiveOpts IntCache: cache: 0, tcache: 0, allocated: 4, tallocated: 105 FML: MCP 9.42 Powered by Forge 14.23.4.2760 76 mods loaded, 76 mods active States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored | State | ID | Version | Source | Signature | |:--------- |:------------------------- |:------------------------ |:----------------------------------------------- |:---------------------------------------- | | UCHIJAAAA | minecraft | 1.12.2 | minecraft.jar | None | | UCHIJAAAA | mcp | 9.42 | minecraft.jar | None | | UCHIJAAAA | FML | 8.0.99.99 | forge-1.12.2-14.23.4.2760-universal.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 | | UCHIJAAAA | forge | 14.23.4.2760 | forge-1.12.2-14.23.4.2760-universal.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 | | UCHIJAAAA | itemblacklist | 1.4.0 | ItemBlacklist-1.4.0.jar | None | | UCHIJAAAA | antiqueatlas | 4.4.9 | antiqueatlas-1.12.2-4.4.9.jar | None | | UCHIJAAAA | examplemod | 1.0 | antiqueatlas-1.12.2-4.4.9.jar | None | | UCHIJAAAA | applecore | 3.1.4 | AppleCore-mc1.12.2-3.1.4.jar | None | | UCHIJAAAA | jei | 4.11.0.212 | jei_1.12.2-4.11.0.212.jar | None | | UCHIJAAAA | appleskin | 1.0.9 | AppleSkin-mc1.12-1.0.9.jar | None | | UCHIJAAAA | architecturecraft | @VERSION@ | architecturecraft-1.12-2.68.jar | None | | UCHIJAAAA | autoplant | 1.12-1.0.0 | autoplant-1.12-1.0.0.jar | None | | UCHIJAAAA | backpack | 3.0.2 | backpack-3.0.2-1.12.2.jar | None | | UCHIJAAAA | bibliocraft | 2.4.5 | BiblioCraft[v2.4.5][MC1.12.2].jar | None | | UCHIJAAAA | biomesoplenty | 7.0.1.2399 | BiomesOPlenty-1.12.2-7.0.1.2399-universal.jar | None | | UCHIJAAAA | bookshelf | 2.3.559 | Bookshelf-1.12.2-2.3.559.jar | d476d1b22b218a10d845928d1665d45fce301b27 | | UCHIJAAAA | chameleon | 1.12-4.1.3 | Chameleon-1.12-4.1.3.jar | None | | UCHIJAAAA | codechickenlib | 3.2.2.353 | CodeChickenLib-1.12.2-3.2.2.353-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | chickenchunks | 2.4.1.73 | ChickenChunks-1.12.2-2.4.1.73-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | chisel | MC1.12.2-0.2.1.34 | Chisel-MC1.12.2-0.2.1.34.jar | None | | UCHIJAAAA | redstoneflux | 2.0.2 | RedstoneFlux-1.12-2.0.2.3-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | cofhcore | 4.5.3 | CoFHCore-1.12.2-4.5.3.20-universal.jar | None | | UCHIJAAAA | cofhworld | 1.2.0 | CoFHWorld-1.12.2-1.2.0.5-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | ic2 | 2.8.99-ex112 | industrialcraft-2-2.8.99-ex112.jar | de041f9f6187debbc77034a344134053277aa3b0 | | UCHIJAAAA | compactsolars | 1.12.2-5.0.17.340 | CompactSolars-1.12.2-5.0.17.340-universal.jar | None | | UCHIJAAAA | customspawner | 3.11.4 | CustomMobSpawner-3.11.4.jar | None | | UCHIJAAAA | customnpcs | 1.12 | CustomNPCs_1.12.2(26aug18).jar | None | | UCHIJAAAA | ptrmodellib | 1.0.2 | PTRLib-1.0.2.jar | None | | UCHIJAAAA | props | 2.6.1 | Decocraft-2.6.1_1.12.2.jar | None | | UCHIJAAAA | mocreatures | 12.0.5 | DrZharks MoCreatures Mod-12.0.5.jar | None | | UCHIJAAAA | dynmap | 2.6-beta-1-23 | dynmap.jar | None | | UCHIJAAAA | dynmapblockscan | 3.0-alpha-1-8 | DynmapBlockScan-3.0-alpha-1-forge-1.12.2.jar | None | | UCHIJAAAA | eplus | 5.0.164 | EnchantingPlus-1.12.2-5.0.164.jar | d476d1b22b218a10d845928d1665d45fce301b27 | | UCHIJAAAA | enderstorage | 2.4.5.135 | EnderStorage-1.12.2-2.4.5.135-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | farseek | 2.3.1 | Farseek-1.12-2.3.1.jar | None | | UCHIJAAAA | forgemultipartcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | microblockcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | None | | UCHIJAAAA | minecraftmultipartcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | None | | UCHIJAAAA | gardenstuff | 1.12-2.1.1 | GardenStuff-1.12-2.1.1.jar | None | | UCHIJAAAA | gravestone | 1.0.18 | GraveStone-1.12.2-Graves-1.0.18.jar | None | | UCHIJAAAA | gravestone-extended | 1.2.6 | GraveStone-1.12.2-Extended-1.2.6.jar | None | | UCHIJAAAA | ichunutil | 7.1.4 | iChunUtil-1.12.2-7.1.4.jar | None | | UCHIJAAAA | hats | 7.0.0 | Hats-1.12.2-7.0.2.jar | None | | UCHIJAAAA | hatstand | 7.0.0 | HatStand-1.12.2-7.0.1.jar | None | | UCHIJAAAA | hopperducts | 1.5 | hopperducts-mc1.12-1.5.jar | None | | UCHIJAAAA | inventorytweaks | 1.63+release.109.220f184 | InventoryTweaks-1.63.jar | 55d2cd4f5f0961410bf7b91ef6c6bf00a766dcbe | | UCHIJAAAA | mantle | 1.12-1.3.2.24 | Mantle-1.12-1.3.2.24.jar | None | | UCHIJAAAA | mcmultipart | 2.5.3 | MCMultiPart-2.5.3.jar | None | | UCHIJAAAA | mrtjpcore | 2.1.3.35 | MrTJPCore-1.12.2-2.1.3.35-universal.jar | None | | UCHIJAAAA | nei | 2.4.1 | NotEnoughItems-1.12.2-2.4.1.238-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | harvestcraft | 1.12.2z | Pam's+HarvestCraft+1.12.2z.jar | None | | UCHIJAAAA | sonarcore | 5.0.16 | sonarcore-1.12.2-5.0.16-16.jar | None | | UCHIJAAAA | practicallogistics2 | 3.0.4 | practicallogistics2-1.12.2-3.0.4.jar | None | | UCHIJAAAA | projectred-core | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-Base.jar | None | | UCHIJAAAA | projectred-compat | 1.0 | ProjectRed-1.12.2-4.9.1.92-compat.jar | None | | UCHIJAAAA | projectred-integration | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-integration.jar | None | | UCHIJAAAA | projectred-transmission | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-integration.jar | None | | UCHIJAAAA | projectred-fabrication | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-fabrication.jar | None | | UCHIJAAAA | projectred-illumination | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-lighting.jar | None | | UCHIJAAAA | projectred-expansion | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-relocation | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-transportation | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-exploration | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-world.jar | None | | UCHIJAAAA | rangedpumps | 0.5 | rangedpumps-0.5.jar | None | | UCHIJAAAA | ruins | 17.1 | Ruins-1.12.2.jar | None | | UCHIJAAAA | storagedrawers | 1.12-5.3.5 | StorageDrawers-1.12.2-5.3.7.jar | None | | UCHIJAAAA | storagedrawersextra | @VERSION@ | StorageDrawersExtras-1.12-3.1.0.jar | None | | UCHIJAAAA | tconstruct | 1.12.2-2.10.1.84 | TConstruct-1.12.2-2.10.1.84.jar | None | | UCHIJAAAA | thermalfoundation | 2.5.0 | ThermalFoundation-1.12.2-2.5.0.19-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thermalexpansion | 5.5.0 | ThermalExpansion-1.12.2-5.5.0.29-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thermaldynamics | 2.5.1 | ThermalDynamics-1.12.2-2.5.1.14-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thutcore | 5.19.8 | thutcore-1.12.2-5.19.8.jar | None | | UCHIJAAAA | thutcore_compat | 1.0 | thutcore-1.12.2-5.19.8.jar | None | | UCHIJAAAA | thuttech | 6.1.5 | thuttech-1.12.2-6.1.5.jar | None | | UCHIJAAAA | worldedit | 6.1.8 | worldedit-forge-mc1.12-6.1.8-dist.jar | None | | UCHIJAAAA | hungeroverhaul | 1.12.2-1.3.3.jenkins148 | HungerOverhaul-1.12.2-1.3.3.jenkins148.jar | None | Loaded coremods (and transformers): FarseekCoreMod (Farseek-1.12-2.3.1.jar) farseek.core.FarseekClassTransformer Inventory Tweaks Coremod (InventoryTweaks-1.63.jar) invtweaks.forge.asm.ContainerTransformer CTMCorePlugin (CTM-MC1.12.2-0.3.2.18.jar) team.chisel.ctm.client.asm.CTMTransformer AppleCore (AppleCore-mc1.12.2-3.1.4.jar) squeek.applecore.asm.TransformerModuleHandler BetterFoliageLoader (BetterFoliage-MC1.12-2.1.10.jar) mods.betterfoliage.loader.BetterFoliageTransformer Pulsar/tconstruct loaded Pulses: - TinkerCommons (Enabled/Forced) - TinkerWorld (Enabled/Not Forced) - TinkerTools (Enabled/Not Forced) - TinkerHarvestTools (Enabled/Forced) - TinkerMeleeWeapons (Enabled/Forced) - TinkerRangedWeapons (Enabled/Forced) - TinkerModifiers (Enabled/Forced) - TinkerSmeltery (Enabled/Not Forced) - TinkerGadgets (Enabled/Not Forced) - TinkerOredict (Enabled/Forced) - TinkerIntegration (Enabled/Forced) - TinkerFluids (Enabled/Forced) - TinkerMaterials (Enabled/Forced) - TinkerModelRegister (Enabled/Forced) - chiselIntegration (Enabled/Not Forced) Profiler Position: N/A (disabled) Player Count: 1 / 20; [EntityPlayerMP['PsyVirus'/2214, l='world', x=259.81, y=9.00, z=2053.50]] Is Modded: Definitely; Server brand changed to 'fml,forge' Type: Dedicated Server (map_server.txt) ```
True
World breaking Crash if putting ProjectRed Lamps into an elevator i.e. as top or bottom - #### What happens: If you have built your elevator and want to create it with the device linker and have Lamp - blocks from ProjectRed in them, the server crashes and can't be started any more until you delete the elevator via MC-Edit or something. I'll also open an issue on PR Repo. #### What you expected to happen: A perfectly fine and fluid gaming ex...perie.. Well it's Minecraft. #### Steps to reproduce: 1. Same 2. as 3. What happens ____ Crash Report, including Mod List ``` ---- Minecraft Crash Report ---- WARNING: coremods are present: FarseekCoreMod (Farseek-1.12-2.3.1.jar) Inventory Tweaks Coremod (InventoryTweaks-1.63.jar) CTMCorePlugin (CTM-MC1.12.2-0.3.2.18.jar) AppleCore (AppleCore-mc1.12.2-3.1.4.jar) BetterFoliageLoader (BetterFoliage-MC1.12-2.1.10.jar) Contact their authors BEFORE contacting forge // Everything's going to plan. No, really, that was supposed to happen. Time: 9/20/18 7:20 PM Description: Ticking entity java.lang.NoClassDefFoundError: codechicken/lib/model/bakery/ModelBakery at mrtjp.projectred.illumination.BlockLamp.getExtendedState(blocks.scala:75) at thut.api.entity.blockentity.BlockEntityUpdater.applyEntityCollision(BlockEntityUpdater.java:157) at thut.api.entity.blockentity.BlockEntityBase.func_70108_f(BlockEntityBase.java:153) at thut.api.entity.blockentity.BlockEntityBase.checkCollision(BlockEntityBase.java:199) at thut.api.entity.blockentity.BlockEntityBase.func_70071_h_(BlockEntityBase.java:314) at net.minecraft.world.World.func_72866_a(World.java:1990) at net.minecraft.world.WorldServer.func_72866_a(WorldServer.java:832) at net.minecraft.world.World.func_72870_g(World.java:1952) at net.minecraft.world.World.func_72939_s(World.java:1756) at net.minecraft.world.WorldServer.func_72939_s(WorldServer.java:613) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:767) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.ClassNotFoundException: codechicken.lib.model.bakery.ModelBakery at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:191) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 15 more Caused by: net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerException: Exception in class transformer net.minecraftforge.fml.common.asm.transformers.SideTransformer@d02f8d from coremod FMLCorePlugin at net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerWrapper.transform(ASMTransformerWrapper.java:260) at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:279) at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:176) ... 17 more Caused by: java.lang.RuntimeException: Attempted to load class codechicken/lib/model/bakery/ModelBakery for invalid side SERVER at net.minecraftforge.fml.common.asm.transformers.SideTransformer.transform(SideTransformer.java:62) at net.minecraftforge.fml.common.asm.ASMTransformerWrapper$TransformerWrapper.transform(ASMTransformerWrapper.java:256) ... 19 more A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Thread: Server thread Stacktrace: at mrtjp.projectred.illumination.BlockLamp.getExtendedState(blocks.scala:75) at thut.api.entity.blockentity.BlockEntityUpdater.applyEntityCollision(BlockEntityUpdater.java:157) at thut.api.entity.blockentity.BlockEntityBase.func_70108_f(BlockEntityBase.java:153) at thut.api.entity.blockentity.BlockEntityBase.checkCollision(BlockEntityBase.java:199) at thut.api.entity.blockentity.BlockEntityBase.func_70071_h_(BlockEntityBase.java:314) at net.minecraft.world.World.func_72866_a(World.java:1990) at net.minecraft.world.WorldServer.func_72866_a(WorldServer.java:832) at net.minecraft.world.World.func_72870_g(World.java:1952) -- Entity being ticked -- Details: Entity Type: thuttech:lift (thut.tech.common.entity.EntityLift) Entity ID: 2022 Entity Name: Elevator Entity's Exact location: 258.50, 4.00, 2053.50 Entity's Block location: World: (258,4,2053), Chunk: (at 2,0,5 in 16,128; contains blocks 256,0,2048 to 271,255,2063), Region: (0,4; contains chunks 0,128 to 31,159, blocks 0,0,2048 to 511,255,2559) Entity's Momentum: 0.00, 0.00, 0.00 Entity's Passengers: [] Entity's Vehicle: ~~ERROR~~ NullPointerException: null Stacktrace: at net.minecraft.world.World.func_72939_s(World.java:1756) at net.minecraft.world.WorldServer.func_72939_s(WorldServer.java:613) -- Affected level -- Details: Level name: world All players: 1 total; [EntityPlayerMP['PsyVirus'/2214, l='world', x=259.81, y=9.00, z=2053.50]] Chunk stats: ServerChunkCache: 817 Drop: 0 Level seed: 4209520796579142338 Level generator: ID 06 - BIOMESOP, ver 0. Features enabled: true Level generator options: Level spawn location: World: (87,73,2035), Chunk: (at 7,4,3 in 5,127; contains blocks 80,0,2032 to 95,255,2047), Region: (0,3; contains chunks 0,96 to 31,127, blocks 0,0,1536 to 511,255,2047) Level time: 13471423 game time, 14130359 day time Level dimension: 0 Level storage version: 0x04ABD - Anvil Level weather: Rain time: 6353 (now: true), thunder time: 53727 (now: false) Level game mode: Game mode: survival (ID 0). Hardcore: false. Cheats: false Stacktrace: at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:767) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:397) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:668) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:526) at java.lang.Thread.run(Thread.java:748) -- System Details -- Details: Minecraft Version: 1.12.2 Operating System: Linux (amd64) version 4.4.0-135-generic Java Version: 1.8.0_181, Oracle Corporation Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 580625768 bytes (553 MB) / 2359259136 bytes (2249 MB) up to 10468982784 bytes (9984 MB) JVM Flags: 8 total; -Xms512M -Xmx10G -XX:NewSize=512M -XX:MaxNewSize=1G -XX:SurvivorRatio=2 -XX:+DisableExplicitGC -XX:+UseConcMarkSweepGC -XX:+AggressiveOpts IntCache: cache: 0, tcache: 0, allocated: 4, tallocated: 105 FML: MCP 9.42 Powered by Forge 14.23.4.2760 76 mods loaded, 76 mods active States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored | State | ID | Version | Source | Signature | |:--------- |:------------------------- |:------------------------ |:----------------------------------------------- |:---------------------------------------- | | UCHIJAAAA | minecraft | 1.12.2 | minecraft.jar | None | | UCHIJAAAA | mcp | 9.42 | minecraft.jar | None | | UCHIJAAAA | FML | 8.0.99.99 | forge-1.12.2-14.23.4.2760-universal.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 | | UCHIJAAAA | forge | 14.23.4.2760 | forge-1.12.2-14.23.4.2760-universal.jar | e3c3d50c7c986df74c645c0ac54639741c90a557 | | UCHIJAAAA | itemblacklist | 1.4.0 | ItemBlacklist-1.4.0.jar | None | | UCHIJAAAA | antiqueatlas | 4.4.9 | antiqueatlas-1.12.2-4.4.9.jar | None | | UCHIJAAAA | examplemod | 1.0 | antiqueatlas-1.12.2-4.4.9.jar | None | | UCHIJAAAA | applecore | 3.1.4 | AppleCore-mc1.12.2-3.1.4.jar | None | | UCHIJAAAA | jei | 4.11.0.212 | jei_1.12.2-4.11.0.212.jar | None | | UCHIJAAAA | appleskin | 1.0.9 | AppleSkin-mc1.12-1.0.9.jar | None | | UCHIJAAAA | architecturecraft | @VERSION@ | architecturecraft-1.12-2.68.jar | None | | UCHIJAAAA | autoplant | 1.12-1.0.0 | autoplant-1.12-1.0.0.jar | None | | UCHIJAAAA | backpack | 3.0.2 | backpack-3.0.2-1.12.2.jar | None | | UCHIJAAAA | bibliocraft | 2.4.5 | BiblioCraft[v2.4.5][MC1.12.2].jar | None | | UCHIJAAAA | biomesoplenty | 7.0.1.2399 | BiomesOPlenty-1.12.2-7.0.1.2399-universal.jar | None | | UCHIJAAAA | bookshelf | 2.3.559 | Bookshelf-1.12.2-2.3.559.jar | d476d1b22b218a10d845928d1665d45fce301b27 | | UCHIJAAAA | chameleon | 1.12-4.1.3 | Chameleon-1.12-4.1.3.jar | None | | UCHIJAAAA | codechickenlib | 3.2.2.353 | CodeChickenLib-1.12.2-3.2.2.353-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | chickenchunks | 2.4.1.73 | ChickenChunks-1.12.2-2.4.1.73-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | chisel | MC1.12.2-0.2.1.34 | Chisel-MC1.12.2-0.2.1.34.jar | None | | UCHIJAAAA | redstoneflux | 2.0.2 | RedstoneFlux-1.12-2.0.2.3-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | cofhcore | 4.5.3 | CoFHCore-1.12.2-4.5.3.20-universal.jar | None | | UCHIJAAAA | cofhworld | 1.2.0 | CoFHWorld-1.12.2-1.2.0.5-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | ic2 | 2.8.99-ex112 | industrialcraft-2-2.8.99-ex112.jar | de041f9f6187debbc77034a344134053277aa3b0 | | UCHIJAAAA | compactsolars | 1.12.2-5.0.17.340 | CompactSolars-1.12.2-5.0.17.340-universal.jar | None | | UCHIJAAAA | customspawner | 3.11.4 | CustomMobSpawner-3.11.4.jar | None | | UCHIJAAAA | customnpcs | 1.12 | CustomNPCs_1.12.2(26aug18).jar | None | | UCHIJAAAA | ptrmodellib | 1.0.2 | PTRLib-1.0.2.jar | None | | UCHIJAAAA | props | 2.6.1 | Decocraft-2.6.1_1.12.2.jar | None | | UCHIJAAAA | mocreatures | 12.0.5 | DrZharks MoCreatures Mod-12.0.5.jar | None | | UCHIJAAAA | dynmap | 2.6-beta-1-23 | dynmap.jar | None | | UCHIJAAAA | dynmapblockscan | 3.0-alpha-1-8 | DynmapBlockScan-3.0-alpha-1-forge-1.12.2.jar | None | | UCHIJAAAA | eplus | 5.0.164 | EnchantingPlus-1.12.2-5.0.164.jar | d476d1b22b218a10d845928d1665d45fce301b27 | | UCHIJAAAA | enderstorage | 2.4.5.135 | EnderStorage-1.12.2-2.4.5.135-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | farseek | 2.3.1 | Farseek-1.12-2.3.1.jar | None | | UCHIJAAAA | forgemultipartcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | microblockcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | None | | UCHIJAAAA | minecraftmultipartcbe | 2.5.0.71 | ForgeMultipart-1.12.2-2.5.0.71-universal.jar | None | | UCHIJAAAA | gardenstuff | 1.12-2.1.1 | GardenStuff-1.12-2.1.1.jar | None | | UCHIJAAAA | gravestone | 1.0.18 | GraveStone-1.12.2-Graves-1.0.18.jar | None | | UCHIJAAAA | gravestone-extended | 1.2.6 | GraveStone-1.12.2-Extended-1.2.6.jar | None | | UCHIJAAAA | ichunutil | 7.1.4 | iChunUtil-1.12.2-7.1.4.jar | None | | UCHIJAAAA | hats | 7.0.0 | Hats-1.12.2-7.0.2.jar | None | | UCHIJAAAA | hatstand | 7.0.0 | HatStand-1.12.2-7.0.1.jar | None | | UCHIJAAAA | hopperducts | 1.5 | hopperducts-mc1.12-1.5.jar | None | | UCHIJAAAA | inventorytweaks | 1.63+release.109.220f184 | InventoryTweaks-1.63.jar | 55d2cd4f5f0961410bf7b91ef6c6bf00a766dcbe | | UCHIJAAAA | mantle | 1.12-1.3.2.24 | Mantle-1.12-1.3.2.24.jar | None | | UCHIJAAAA | mcmultipart | 2.5.3 | MCMultiPart-2.5.3.jar | None | | UCHIJAAAA | mrtjpcore | 2.1.3.35 | MrTJPCore-1.12.2-2.1.3.35-universal.jar | None | | UCHIJAAAA | nei | 2.4.1 | NotEnoughItems-1.12.2-2.4.1.238-universal.jar | f1850c39b2516232a2108a7bd84d1cb5df93b261 | | UCHIJAAAA | harvestcraft | 1.12.2z | Pam's+HarvestCraft+1.12.2z.jar | None | | UCHIJAAAA | sonarcore | 5.0.16 | sonarcore-1.12.2-5.0.16-16.jar | None | | UCHIJAAAA | practicallogistics2 | 3.0.4 | practicallogistics2-1.12.2-3.0.4.jar | None | | UCHIJAAAA | projectred-core | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-Base.jar | None | | UCHIJAAAA | projectred-compat | 1.0 | ProjectRed-1.12.2-4.9.1.92-compat.jar | None | | UCHIJAAAA | projectred-integration | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-integration.jar | None | | UCHIJAAAA | projectred-transmission | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-integration.jar | None | | UCHIJAAAA | projectred-fabrication | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-fabrication.jar | None | | UCHIJAAAA | projectred-illumination | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-lighting.jar | None | | UCHIJAAAA | projectred-expansion | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-relocation | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-transportation | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-mechanical.jar | None | | UCHIJAAAA | projectred-exploration | 4.9.1.92 | ProjectRed-1.12.2-4.9.1.92-world.jar | None | | UCHIJAAAA | rangedpumps | 0.5 | rangedpumps-0.5.jar | None | | UCHIJAAAA | ruins | 17.1 | Ruins-1.12.2.jar | None | | UCHIJAAAA | storagedrawers | 1.12-5.3.5 | StorageDrawers-1.12.2-5.3.7.jar | None | | UCHIJAAAA | storagedrawersextra | @VERSION@ | StorageDrawersExtras-1.12-3.1.0.jar | None | | UCHIJAAAA | tconstruct | 1.12.2-2.10.1.84 | TConstruct-1.12.2-2.10.1.84.jar | None | | UCHIJAAAA | thermalfoundation | 2.5.0 | ThermalFoundation-1.12.2-2.5.0.19-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thermalexpansion | 5.5.0 | ThermalExpansion-1.12.2-5.5.0.29-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thermaldynamics | 2.5.1 | ThermalDynamics-1.12.2-2.5.1.14-universal.jar | 8a6abf2cb9e141b866580d369ba6548732eff25f | | UCHIJAAAA | thutcore | 5.19.8 | thutcore-1.12.2-5.19.8.jar | None | | UCHIJAAAA | thutcore_compat | 1.0 | thutcore-1.12.2-5.19.8.jar | None | | UCHIJAAAA | thuttech | 6.1.5 | thuttech-1.12.2-6.1.5.jar | None | | UCHIJAAAA | worldedit | 6.1.8 | worldedit-forge-mc1.12-6.1.8-dist.jar | None | | UCHIJAAAA | hungeroverhaul | 1.12.2-1.3.3.jenkins148 | HungerOverhaul-1.12.2-1.3.3.jenkins148.jar | None | Loaded coremods (and transformers): FarseekCoreMod (Farseek-1.12-2.3.1.jar) farseek.core.FarseekClassTransformer Inventory Tweaks Coremod (InventoryTweaks-1.63.jar) invtweaks.forge.asm.ContainerTransformer CTMCorePlugin (CTM-MC1.12.2-0.3.2.18.jar) team.chisel.ctm.client.asm.CTMTransformer AppleCore (AppleCore-mc1.12.2-3.1.4.jar) squeek.applecore.asm.TransformerModuleHandler BetterFoliageLoader (BetterFoliage-MC1.12-2.1.10.jar) mods.betterfoliage.loader.BetterFoliageTransformer Pulsar/tconstruct loaded Pulses: - TinkerCommons (Enabled/Forced) - TinkerWorld (Enabled/Not Forced) - TinkerTools (Enabled/Not Forced) - TinkerHarvestTools (Enabled/Forced) - TinkerMeleeWeapons (Enabled/Forced) - TinkerRangedWeapons (Enabled/Forced) - TinkerModifiers (Enabled/Forced) - TinkerSmeltery (Enabled/Not Forced) - TinkerGadgets (Enabled/Not Forced) - TinkerOredict (Enabled/Forced) - TinkerIntegration (Enabled/Forced) - TinkerFluids (Enabled/Forced) - TinkerMaterials (Enabled/Forced) - TinkerModelRegister (Enabled/Forced) - chiselIntegration (Enabled/Not Forced) Profiler Position: N/A (disabled) Player Count: 1 / 20; [EntityPlayerMP['PsyVirus'/2214, l='world', x=259.81, y=9.00, z=2053.50]] Is Modded: Definitely; Server brand changed to 'fml,forge' Type: Dedicated Server (map_server.txt) ```
non_process
world breaking crash if putting projectred lamps into an elevator i e as top or bottom what happens if you have built your elevator and want to create it with the device linker and have lamp blocks from projectred in them the server crashes and can t be started any more until you delete the elevator via mc edit or something i ll also open an issue on pr repo what you expected to happen a perfectly fine and fluid gaming ex perie well it s minecraft steps to reproduce same as what happens crash report including mod list minecraft crash report warning coremods are present farseekcoremod farseek jar inventory tweaks coremod inventorytweaks jar ctmcoreplugin ctm jar applecore applecore jar betterfoliageloader betterfoliage jar contact their authors before contacting forge everything s going to plan no really that was supposed to happen time pm description ticking entity java lang noclassdeffounderror codechicken lib model bakery modelbakery at mrtjp projectred illumination blocklamp getextendedstate blocks scala at thut api entity blockentity blockentityupdater applyentitycollision blockentityupdater java at thut api entity blockentity blockentitybase func f blockentitybase java at thut api entity blockentity blockentitybase checkcollision blockentitybase java at thut api entity blockentity blockentitybase func h blockentitybase java at net minecraft world world func a world java at net minecraft world worldserver func a worldserver java at net minecraft world world func g world java at net minecraft world world func s world java at net minecraft world worldserver func s worldserver java at net minecraft server minecraftserver func q minecraftserver java at net minecraft server dedicated dedicatedserver func q dedicatedserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java caused by java lang classnotfoundexception codechicken lib model bakery modelbakery at net minecraft launchwrapper launchclassloader findclass launchclassloader java at java lang classloader loadclass classloader java at java lang classloader loadclass classloader java more caused by net minecraftforge fml common asm asmtransformerwrapper transformerexception exception in class transformer net minecraftforge fml common asm transformers sidetransformer from coremod fmlcoreplugin at net minecraftforge fml common asm asmtransformerwrapper transformerwrapper transform asmtransformerwrapper java at net minecraft launchwrapper launchclassloader runtransformers launchclassloader java at net minecraft launchwrapper launchclassloader findclass launchclassloader java more caused by java lang runtimeexception attempted to load class codechicken lib model bakery modelbakery for invalid side server at net minecraftforge fml common asm transformers sidetransformer transform sidetransformer java at net minecraftforge fml common asm asmtransformerwrapper transformerwrapper transform asmtransformerwrapper java more a detailed walkthrough of the error its code path and all known details is as follows head thread server thread stacktrace at mrtjp projectred illumination blocklamp getextendedstate blocks scala at thut api entity blockentity blockentityupdater applyentitycollision blockentityupdater java at thut api entity blockentity blockentitybase func f blockentitybase java at thut api entity blockentity blockentitybase checkcollision blockentitybase java at thut api entity blockentity blockentitybase func h blockentitybase java at net minecraft world world func a world java at net minecraft world worldserver func a worldserver java at net minecraft world world func g world java entity being ticked details entity type thuttech lift thut tech common entity entitylift entity id entity name elevator entity s exact location entity s block location world chunk at in contains blocks to region contains chunks to blocks to entity s momentum entity s passengers entity s vehicle error nullpointerexception null stacktrace at net minecraft world world func s world java at net minecraft world worldserver func s worldserver java affected level details level name world all players total chunk stats serverchunkcache drop level seed level generator id biomesop ver features enabled true level generator options level spawn location world chunk at in contains blocks to region contains chunks to blocks to level time game time day time level dimension level storage version anvil level weather rain time now true thunder time now false level game mode game mode survival id hardcore false cheats false stacktrace at net minecraft server minecraftserver func q minecraftserver java at net minecraft server dedicated dedicatedserver func q dedicatedserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java system details details minecraft version operating system linux version generic java version oracle corporation java vm version openjdk bit server vm mixed mode oracle corporation memory bytes mb bytes mb up to bytes mb jvm flags total xx newsize xx maxnewsize xx survivorratio xx disableexplicitgc xx useconcmarksweepgc xx aggressiveopts intcache cache tcache allocated tallocated fml mcp powered by forge mods loaded mods active states u unloaded l loaded c constructed h pre initialized i initialized j post initialized a available d disabled e errored state id version source signature uchijaaaa minecraft minecraft jar none uchijaaaa mcp minecraft jar none uchijaaaa fml forge universal jar uchijaaaa forge forge universal jar uchijaaaa itemblacklist itemblacklist jar none uchijaaaa antiqueatlas antiqueatlas jar none uchijaaaa examplemod antiqueatlas jar none uchijaaaa applecore applecore jar none uchijaaaa jei jei jar none uchijaaaa appleskin appleskin jar none uchijaaaa architecturecraft version architecturecraft jar none uchijaaaa autoplant autoplant jar none uchijaaaa backpack backpack jar none uchijaaaa bibliocraft bibliocraft jar none uchijaaaa biomesoplenty biomesoplenty universal jar none uchijaaaa bookshelf bookshelf jar uchijaaaa chameleon chameleon jar none uchijaaaa codechickenlib codechickenlib universal jar uchijaaaa chickenchunks chickenchunks universal jar uchijaaaa chisel chisel jar none uchijaaaa redstoneflux redstoneflux universal jar uchijaaaa cofhcore cofhcore universal jar none uchijaaaa cofhworld cofhworld universal jar uchijaaaa industrialcraft jar uchijaaaa compactsolars compactsolars universal jar none uchijaaaa customspawner custommobspawner jar none uchijaaaa customnpcs customnpcs jar none uchijaaaa ptrmodellib ptrlib jar none uchijaaaa props decocraft jar none uchijaaaa mocreatures drzharks mocreatures mod jar none uchijaaaa dynmap beta dynmap jar none uchijaaaa dynmapblockscan alpha dynmapblockscan alpha forge jar none uchijaaaa eplus enchantingplus jar uchijaaaa enderstorage enderstorage universal jar uchijaaaa farseek farseek jar none uchijaaaa forgemultipartcbe forgemultipart universal jar uchijaaaa microblockcbe forgemultipart universal jar none uchijaaaa minecraftmultipartcbe forgemultipart universal jar none uchijaaaa gardenstuff gardenstuff jar none uchijaaaa gravestone gravestone graves jar none uchijaaaa gravestone extended gravestone extended jar none uchijaaaa ichunutil ichunutil jar none uchijaaaa hats hats jar none uchijaaaa hatstand hatstand jar none uchijaaaa hopperducts hopperducts jar none uchijaaaa inventorytweaks release inventorytweaks jar uchijaaaa mantle mantle jar none uchijaaaa mcmultipart mcmultipart jar none uchijaaaa mrtjpcore mrtjpcore universal jar none uchijaaaa nei notenoughitems universal jar uchijaaaa harvestcraft pam s harvestcraft jar none uchijaaaa sonarcore sonarcore jar none uchijaaaa jar none uchijaaaa projectred core projectred base jar none uchijaaaa projectred compat projectred compat jar none uchijaaaa projectred integration projectred integration jar none uchijaaaa projectred transmission projectred integration jar none uchijaaaa projectred fabrication projectred fabrication jar none uchijaaaa projectred illumination projectred lighting jar none uchijaaaa projectred expansion projectred mechanical jar none uchijaaaa projectred relocation projectred mechanical jar none uchijaaaa projectred transportation projectred mechanical jar none uchijaaaa projectred exploration projectred world jar none uchijaaaa rangedpumps rangedpumps jar none uchijaaaa ruins ruins jar none uchijaaaa storagedrawers storagedrawers jar none uchijaaaa storagedrawersextra version storagedrawersextras jar none uchijaaaa tconstruct tconstruct jar none uchijaaaa thermalfoundation thermalfoundation universal jar uchijaaaa thermalexpansion thermalexpansion universal jar uchijaaaa thermaldynamics thermaldynamics universal jar uchijaaaa thutcore thutcore jar none uchijaaaa thutcore compat thutcore jar none uchijaaaa thuttech thuttech jar none uchijaaaa worldedit worldedit forge dist jar none uchijaaaa hungeroverhaul hungeroverhaul jar none loaded coremods and transformers farseekcoremod farseek jar farseek core farseekclasstransformer inventory tweaks coremod inventorytweaks jar invtweaks forge asm containertransformer ctmcoreplugin ctm jar team chisel ctm client asm ctmtransformer applecore applecore jar squeek applecore asm transformermodulehandler betterfoliageloader betterfoliage jar mods betterfoliage loader betterfoliagetransformer pulsar tconstruct loaded pulses tinkercommons enabled forced tinkerworld enabled not forced tinkertools enabled not forced tinkerharvesttools enabled forced tinkermeleeweapons enabled forced tinkerrangedweapons enabled forced tinkermodifiers enabled forced tinkersmeltery enabled not forced tinkergadgets enabled not forced tinkeroredict enabled forced tinkerintegration enabled forced tinkerfluids enabled forced tinkermaterials enabled forced tinkermodelregister enabled forced chiselintegration enabled not forced profiler position n a disabled player count is modded definitely server brand changed to fml forge type dedicated server map server txt
0
472,250
13,621,529,148
IssuesEvent
2020-09-24 00:57:55
AtlasOfLivingAustralia/biocache-hubs
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
closed
Missing i18n labels for leaflet draw tool icons
IAD bug priority-low
The Leaflet draw tools plugin is missing the labels for the icons/tools. See screenshot: ![image](https://user-images.githubusercontent.com/532845/60867383-0aa1f800-a26e-11e9-840e-23e7b0afd6d5.png) I've seen this before and it was a JS loading order problem, so look there first.
1.0
Missing i18n labels for leaflet draw tool icons - The Leaflet draw tools plugin is missing the labels for the icons/tools. See screenshot: ![image](https://user-images.githubusercontent.com/532845/60867383-0aa1f800-a26e-11e9-840e-23e7b0afd6d5.png) I've seen this before and it was a JS loading order problem, so look there first.
non_process
missing labels for leaflet draw tool icons the leaflet draw tools plugin is missing the labels for the icons tools see screenshot i ve seen this before and it was a js loading order problem so look there first
0
76,605
14,652,364,488
IssuesEvent
2020-12-28 01:25:43
wolf-leo/wolfcode-comments
https://api.github.com/repos/wolf-leo/wolfcode-comments
opened
还记得你当年玩过的红色警戒吗?他开源了~~~
Gitalk http://www.wolfcode.com.cn/info/194/
http://www.wolfcode.com.cn/info/194/ 直接上开源地址:https://github.com/electronicarts/CnC_Remastered_Collection 特别说明:开源的《红色警戒》是《红色警戒1》红警1 的启动程序名为 RA95.exe,因此在国内也有很多叫「红警95」。此次开源,并不涉及游戏素材和游戏引擎,只包括
1.0
还记得你当年玩过的红色警戒吗?他开源了~~~ - http://www.wolfcode.com.cn/info/194/ 直接上开源地址:https://github.com/electronicarts/CnC_Remastered_Collection 特别说明:开源的《红色警戒》是《红色警戒1》红警1 的启动程序名为 RA95.exe,因此在国内也有很多叫「红警95」。此次开源,并不涉及游戏素材和游戏引擎,只包括
non_process
还记得你当年玩过的红色警戒吗?他开源了 直接上开源地址: 特别说明:开源的《红色警戒》是《 》 的启动程序名为 exe,因此在国内也有很多叫「 」。此次开源,并不涉及游戏素材和游戏引擎,只包括
0
21,376
29,202,228,326
IssuesEvent
2023-05-21 00:36:52
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / São Paulo, São Paulo, Brazil] Consultant/Java Developer - Híbrida (São Paulo) na Coodesh
SALVADOR BACK-END FRONT-END JAVA FULL-STACK ANGULAR REACT REQUISITOS PROCESSOS BACKEND GITHUB UMA QUALIDADE R NEGÓCIOS ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/consultant-java-developer-sr-194018626?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>KLB Group</strong> busca <strong><ins>Consultant /Java Developer</ins></strong> para compor seu time!</p> <p></p> <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> <p></p> <p><strong>Atividades e Atribuições: </strong></p> <ul> <li>&nbsp;Realizar atendimentos das demandas dos Portais de Sinistros.</li> <li>&nbsp;Analisar as demandas visando atender às necessidades dos clientes;</li> <li>&nbsp;Levantamento de requisitos;</li> <li>Prever impactos das alteraçõs, garantindo a integridade do sistema.</li> </ul> ## KLB GROUP BRASIL : <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> </p> ## Habilidades: - Java - Front-end System Design - Back-end System Design ## Local: São Paulo, São Paulo, Brazil ## Requisitos: - Ensino Superior Completo; - Conhecimento de tecnologias para desenvolvimento JAVA no Backend e Angular, React, entre outras no Frontend; - Estudo da Arquitetura das aplicações para propor soluções; - Entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios. ## Diferenciais: - Desejável conhecimento de Seguros. ## Benefícios: - Vale-refeição R$35,43 dia/ média R$779,70; - Vale-alimentação R$ 602,40; - Vale-transporte (6% de desconto); - Assistência médica/medicina em grupo; - Assistência odontológica R$10,87; - Auxílio Babá - R$ 855,16 até 6 meses; - Auxílio Creche - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - Auxílio Babá - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - PLR; - Seguro de vida; - Empréstimo; - Plano de previdência privada; - Programa de bolsa de estudo; - Total pass. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Consultant/Java Developer - Híbrida (São Paulo) na KLB GROUP BRASIL ](https://coodesh.com/vagas/consultant-java-developer-sr-194018626?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime CLT #### Categoria Full-Stack
1.0
[Hibrido / São Paulo, São Paulo, Brazil] Consultant/Java Developer - Híbrida (São Paulo) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/consultant-java-developer-sr-194018626?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>KLB Group</strong> busca <strong><ins>Consultant /Java Developer</ins></strong> para compor seu time!</p> <p></p> <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> <p></p> <p><strong>Atividades e Atribuições: </strong></p> <ul> <li>&nbsp;Realizar atendimentos das demandas dos Portais de Sinistros.</li> <li>&nbsp;Analisar as demandas visando atender às necessidades dos clientes;</li> <li>&nbsp;Levantamento de requisitos;</li> <li>Prever impactos das alteraçõs, garantindo a integridade do sistema.</li> </ul> ## KLB GROUP BRASIL : <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> </p> ## Habilidades: - Java - Front-end System Design - Back-end System Design ## Local: São Paulo, São Paulo, Brazil ## Requisitos: - Ensino Superior Completo; - Conhecimento de tecnologias para desenvolvimento JAVA no Backend e Angular, React, entre outras no Frontend; - Estudo da Arquitetura das aplicações para propor soluções; - Entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios. ## Diferenciais: - Desejável conhecimento de Seguros. ## Benefícios: - Vale-refeição R$35,43 dia/ média R$779,70; - Vale-alimentação R$ 602,40; - Vale-transporte (6% de desconto); - Assistência médica/medicina em grupo; - Assistência odontológica R$10,87; - Auxílio Babá - R$ 855,16 até 6 meses; - Auxílio Creche - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - Auxílio Babá - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - PLR; - Seguro de vida; - Empréstimo; - Plano de previdência privada; - Programa de bolsa de estudo; - Total pass. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Consultant/Java Developer - Híbrida (São Paulo) na KLB GROUP BRASIL ](https://coodesh.com/vagas/consultant-java-developer-sr-194018626?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime CLT #### Categoria Full-Stack
process
consultant java developer híbrida são paulo na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a klb group busca consultant java developer para compor seu time a klb group é especializada na implementação de projetos em empresas públicas e privadas seja em projetos de desenvolvimento produção ou transformação a klb group assegura a implementação eficaz mobilizando rapidamente uma equipe de especialistas de várias funções compras cadeia de suprimentos qualidade engenharia ti finanças etc com uma combinação única de experiência em design implementação e operação a klb group tem mais de funcionários na europa américa e ásia atividades e atribuições nbsp realizar atendimentos das demandas dos portais de sinistros nbsp analisar as demandas visando atender às necessidades dos clientes nbsp levantamento de requisitos prever impactos das alteraçõs garantindo a integridade do sistema klb group brasil a klb group é especializada na implementação de projetos em empresas públicas e privadas seja em projetos de desenvolvimento produção ou transformação a klb group assegura a implementação eficaz mobilizando rapidamente uma equipe de especialistas de várias funções compras cadeia de suprimentos qualidade engenharia ti finanças etc com uma combinação única de experiência em design implementação e operação a klb group tem mais de funcionários na europa américa e ásia habilidades java front end system design back end system design local são paulo são paulo brazil requisitos ensino superior completo conhecimento de tecnologias para desenvolvimento java no backend e angular react entre outras no frontend estudo da arquitetura das aplicações para propor soluções entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios diferenciais desejável conhecimento de seguros benefícios vale refeição r dia média r vale alimentação r vale transporte de desconto assistência médica medicina em grupo assistência odontológica r auxílio babá r até meses auxílio creche r a partir de meses até anos e meses auxílio babá r a partir de meses até anos e meses plr seguro de vida empréstimo plano de previdência privada programa de bolsa de estudo total pass como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime clt categoria full stack
1
5,053
7,860,886,036
IssuesEvent
2018-06-21 21:33:49
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Link Trac site to internal group page (Trac #3)
Migrated from Trac post_processing senkbeil@uwm.edu task
In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page. Also there is no component for internal site issues. Perhaps this means I am using this trac site wrong. Attachments: [plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff) [plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff) [plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff) [plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff) [plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff) [plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff) [plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff) [plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff) [plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/3 ```json { "status": "closed", "changetime": "2009-05-13T18:09:47", "description": "In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page.\n\nAlso there is no component for internal site issues. Perhaps this means I am using this trac site wrong.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238187000000", "component": "post_processing", "summary": "Link Trac site to internal group page", "priority": "minor", "keywords": "", "time": "2009-05-01T21:17:44", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "task" } ```
1.0
Link Trac site to internal group page (Trac #3) - In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page. Also there is no component for internal site issues. Perhaps this means I am using this trac site wrong. Attachments: [plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff) [plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff) [plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff) [plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff) [plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff) [plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff) [plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff) [plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff) [plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/3 ```json { "status": "closed", "changetime": "2009-05-13T18:09:47", "description": "In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page.\n\nAlso there is no component for internal site issues. Perhaps this means I am using this trac site wrong.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238187000000", "component": "post_processing", "summary": "Link Trac site to internal group page", "priority": "minor", "keywords": "", "time": "2009-05-01T21:17:44", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "task" } ```
process
link trac site to internal group page trac in order for this clubb trac site to be useful to the group it must be readily accessible a link should be placed on the internal group page also there is no component for internal site issues perhaps this means i am using this trac site wrong attachments migrated from json status closed changetime description in order for this clubb trac site to be useful to the group it must be readily accessible a link should be placed on the internal group page n nalso there is no component for internal site issues perhaps this means i am using this trac site wrong reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary link trac site to internal group page priority minor keywords time milestone owner senkbeil uwm edu type task
1
287,874
31,856,459,909
IssuesEvent
2023-09-15 07:49:52
Trinadh465/linux-4.1.15_CVE-2023-26607
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-26607
opened
CVE-2017-16536 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2017-16536 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/cx231xx/cx231xx-cards.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/cx231xx/cx231xx-cards.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The cx231xx_usb_probe function in drivers/media/usb/cx231xx/cx231xx-cards.c in the Linux kernel through 4.13.11 allows local users to cause a denial of service (NULL pointer dereference and system crash) or possibly have unspecified other impact via a crafted USB device. <p>Publish Date: 2017-11-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16536>CVE-2017-16536</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16536</a></p> <p>Release Date: 2017-11-04</p> <p>Fix Resolution: v4.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16536 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2017-16536 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/cx231xx/cx231xx-cards.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/cx231xx/cx231xx-cards.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The cx231xx_usb_probe function in drivers/media/usb/cx231xx/cx231xx-cards.c in the Linux kernel through 4.13.11 allows local users to cause a denial of service (NULL pointer dereference and system crash) or possibly have unspecified other impact via a crafted USB device. <p>Publish Date: 2017-11-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16536>CVE-2017-16536</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16536</a></p> <p>Release Date: 2017-11-04</p> <p>Fix Resolution: v4.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch main vulnerable source files drivers media usb cards c drivers media usb cards c vulnerability details the usb probe function in drivers media usb cards c in the linux kernel through allows local users to cause a denial of service null pointer dereference and system crash or possibly have unspecified other impact via a crafted usb device publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
7,059
10,218,229,223
IssuesEvent
2019-08-15 15:26:52
Project-Cartographer/H2PC_TagExtraction
https://api.github.com/repos/Project-Cartographer/H2PC_TagExtraction
closed
Undo processing on model_animation_graph/jmad
bug post-processing
Animations that make use of the inheritance reference still have inheritance related tag blocks in their modes tag block. Need to clear modes of invalid indexes.
1.0
Undo processing on model_animation_graph/jmad - Animations that make use of the inheritance reference still have inheritance related tag blocks in their modes tag block. Need to clear modes of invalid indexes.
process
undo processing on model animation graph jmad animations that make use of the inheritance reference still have inheritance related tag blocks in their modes tag block need to clear modes of invalid indexes
1
27,128
21,190,819,449
IssuesEvent
2022-04-08 17:11:57
WordPress/performance
https://api.github.com/repos/WordPress/performance
opened
Prepare 1.0.0 release
[Type] Enhancement Infrastructure
This issue is to track preparation of the upcoming `1.0.0` release up until publishing, which is due **April 18, 2022**. * [x] Define scope and tentative release date based on the scope * [ ] Create `release/1.0.0` branch closer to the release date * [ ] Prepare the release (**Monday, April 18, 2022**)
1.0
Prepare 1.0.0 release - This issue is to track preparation of the upcoming `1.0.0` release up until publishing, which is due **April 18, 2022**. * [x] Define scope and tentative release date based on the scope * [ ] Create `release/1.0.0` branch closer to the release date * [ ] Prepare the release (**Monday, April 18, 2022**)
non_process
prepare release this issue is to track preparation of the upcoming release up until publishing which is due april define scope and tentative release date based on the scope create release branch closer to the release date prepare the release monday april
0
6,555
9,647,745,105
IssuesEvent
2019-05-17 14:37:50
yosifov/tech-fundamentals
https://api.github.com/repos/yosifov/tech-fundamentals
closed
Replace Repeating Chars
Text Processing
Write a program that reads a string from the console and replaces any sequence of the same letters with a single corresponding letter. **Examples** Input | Output -- | -- aaaaabbbbbcdddeeeedssaa | abcdedsa qqqwerqwecccwd | qwerqwecwd
1.0
Replace Repeating Chars - Write a program that reads a string from the console and replaces any sequence of the same letters with a single corresponding letter. **Examples** Input | Output -- | -- aaaaabbbbbcdddeeeedssaa | abcdedsa qqqwerqwecccwd | qwerqwecwd
process
replace repeating chars write a program that reads a string from the console and replaces any sequence of the same letters with a single corresponding letter examples input output aaaaabbbbbcdddeeeedssaa abcdedsa qqqwerqwecccwd qwerqwecwd
1
505
2,956,959,452
IssuesEvent
2015-07-08 14:18:17
kangarko/ChatControl
https://api.github.com/repos/kangarko/ChatControl
closed
little bug with clearchat, and suggestions for next Update 5.0.9
to be processed
Current found Bugs: - The clear commands clears the whole chat and ignore the argument Ideas for next version: - placeholder %group, %group_prefix, %group_suffix and %displayedname (for /nick commands for exemple) - a Chatformat for each Permissiongroup (not needed if the placeholders would be there) - maybe the spy can turn into a commandwatcher with bypass permission and the abilities to ignore own commands. (whispering private messagesetc, these all are commands too, so why not?) - extand the clear command so u can give users permissions to clear just their own chat. (When many players are very chatty this is a very very useful feature) - a list in the setting of whitelisted urls that the users can post (ip of the server or www.google.com, etc..) -a better featue for the name notification: Currently u have to write the Full name. but when Names get longer then mostly players just use Shorts like just "cloud" for my name. An information sound when 4 letters equal of a playername would be cool and maybe the name can get colored in chat? (color chooseable in the settings) - maybe a bit better extantion of the filters. Currenly lets say i have the word Spammer on my list an someone write "Sppaaammmeeerrr" then the word cant scaned as Spammer from the plugin. -a globalmute command that completly mutes/unmutes the chat - negation of permission works a bit not really usefull.... example: my admins have some neagtions: - -chatcontrol.bypass.cap - -chatcontrol.bypass.replace - -chatcontrol.bypass.capitalize - -chatcontrol.bypass.punctuate normal this now should activate the chat replacer, the capfilter and the capitalizefilter but without the neagtion of - -chatcontrol.bypass.rules nothing happens. problem: if they got the negation of - -chatcontrol.bypass.rules then all filter work on them :3 also the commands, swear, etc... Edit: - take a look over the basic advertice filter :D i added a few thing in it and added some new. now the filter blocks urls, and ips :) before that players was able write in chat example "hallo.de" and other URLs. (u forgot some \b in the settings of it) I am using: Spigot 1.8.3 latest ProtocolLib latest Version of ChatControl 5.0.8b
1.0
little bug with clearchat, and suggestions for next Update 5.0.9 - Current found Bugs: - The clear commands clears the whole chat and ignore the argument Ideas for next version: - placeholder %group, %group_prefix, %group_suffix and %displayedname (for /nick commands for exemple) - a Chatformat for each Permissiongroup (not needed if the placeholders would be there) - maybe the spy can turn into a commandwatcher with bypass permission and the abilities to ignore own commands. (whispering private messagesetc, these all are commands too, so why not?) - extand the clear command so u can give users permissions to clear just their own chat. (When many players are very chatty this is a very very useful feature) - a list in the setting of whitelisted urls that the users can post (ip of the server or www.google.com, etc..) -a better featue for the name notification: Currently u have to write the Full name. but when Names get longer then mostly players just use Shorts like just "cloud" for my name. An information sound when 4 letters equal of a playername would be cool and maybe the name can get colored in chat? (color chooseable in the settings) - maybe a bit better extantion of the filters. Currenly lets say i have the word Spammer on my list an someone write "Sppaaammmeeerrr" then the word cant scaned as Spammer from the plugin. -a globalmute command that completly mutes/unmutes the chat - negation of permission works a bit not really usefull.... example: my admins have some neagtions: - -chatcontrol.bypass.cap - -chatcontrol.bypass.replace - -chatcontrol.bypass.capitalize - -chatcontrol.bypass.punctuate normal this now should activate the chat replacer, the capfilter and the capitalizefilter but without the neagtion of - -chatcontrol.bypass.rules nothing happens. problem: if they got the negation of - -chatcontrol.bypass.rules then all filter work on them :3 also the commands, swear, etc... Edit: - take a look over the basic advertice filter :D i added a few thing in it and added some new. now the filter blocks urls, and ips :) before that players was able write in chat example "hallo.de" and other URLs. (u forgot some \b in the settings of it) I am using: Spigot 1.8.3 latest ProtocolLib latest Version of ChatControl 5.0.8b
process
little bug with clearchat and suggestions for next update current found bugs the clear commands clears the whole chat and ignore the argument ideas for next version placeholder group group prefix group suffix and displayedname for nick commands for exemple a chatformat for each permissiongroup not needed if the placeholders would be there maybe the spy can turn into a commandwatcher with bypass permission and the abilities to ignore own commands whispering private messagesetc these all are commands too so why not extand the clear command so u can give users permissions to clear just their own chat when many players are very chatty this is a very very useful feature a list in the setting of whitelisted urls that the users can post ip of the server or etc a better featue for the name notification currently u have to write the full name but when names get longer then mostly players just use shorts like just cloud for my name an information sound when letters equal of a playername would be cool and maybe the name can get colored in chat color chooseable in the settings maybe a bit better extantion of the filters currenly lets say i have the word spammer on my list an someone write sppaaammmeeerrr then the word cant scaned as spammer from the plugin a globalmute command that completly mutes unmutes the chat negation of permission works a bit not really usefull example my admins have some neagtions chatcontrol bypass cap chatcontrol bypass replace chatcontrol bypass capitalize chatcontrol bypass punctuate normal this now should activate the chat replacer the capfilter and the capitalizefilter but without the neagtion of chatcontrol bypass rules nothing happens problem if they got the negation of chatcontrol bypass rules then all filter work on them also the commands swear etc edit take a look over the basic advertice filter d i added a few thing in it and added some new now the filter blocks urls and ips before that players was able write in chat example hallo de and other urls u forgot some b in the settings of it i am using spigot latest protocollib latest version of chatcontrol
1